Search results for: operator approach
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 13745

Search results for: operator approach

11135 Land Suitability Approach as an Effort to Design a Sustainable Tourism Area in Pacet Mojokerto

Authors: Erina Wulansari, Bambang Soemardiono, Ispurwono Soemarno

Abstract:

Designing sustainable tourism area is defined as an attempt to design an area, that brings the natural environmental conditions as components are available with a wealth of social conditions and the conservation of natural and cultural heritage. To understanding tourism area in this study is not only focus on the location of the tourist object, but rather to a tourist attraction around the area, tourism objects such as the existence of residential area (settlement), a commercial area, public service area, and the natural environmental area. The principle of success in designing a sustainable tourism area is able to integrate and balance between the limited space and the variety of activities that’s always continuously to growth up. The limited space in this area of tourism needs to be managed properly to minimize the damage of environmental as a result of tourism activities hue. This research aims to identify space in this area of tourism through land suitability approach as an effort to create a sustainable design, especially in terms of ecological. This study will be used several analytical techniques to achieve the research objectives as superimposing analysis with GIS 9.3 software and Analysis Hierarchy Process. Expected outcomes are in the form of classification and criteria of usable space in designing embodiment tourism area. In addition, this study can provide input to the order of settlement patterns as part of the environment in the area of sustainable tourism.

Keywords: sustainable tourism area, land suitability, limited space, environment, criteria

Procedia PDF Downloads 480
11134 Analysis of Complex Business Negotiations: Contributions from Agency-Theory

Authors: Jan Van Uden

Abstract:

The paper reviews classical agency-theory and its contributions to the analysis of complex business negotiations and gives an approach for the modification of the basic agency-model in order to examine the negotiation specific dimensions of agency-problems. By illustrating fundamental potentials for the modification of agency-theory in context of business negotiations the paper highlights recent empirical research that investigates agent-based negotiations and inter-team constellations. A general theoretical analysis of complex negotiation would be based on a two-level approach. First, the modification of the basic agency-model in order to illustrate the organizational context of business negotiations (i.e., multi-agent issues, common-agencies, multi-period models and the concept of bounded rationality). Second, the application of the modified agency-model on complex business negotiations to identify agency-problems and relating areas of risk in the negotiation process. The paper is placed on the first level of analysis – the modification. The method builds on the one hand on insights from behavior decision research (BRD) and on the other hand on findings from agency-theory as normative directives to the modification of the basic model. Through neoclassical assumptions concerning the fundamental aspects of agency-relationships in business negotiations (i.e., asymmetric information, self-interest, risk preferences and conflict of interests), agency-theory helps to draw solutions on stated worst-case-scenarios taken from the daily negotiation routine. As agency-theory is the only universal approach able to identify trade-offs between certain aspects of economic cooperation, insights obtained provide a deeper understanding of the forces that shape business negotiation complexity. The need for a modification of the basic model is illustrated by highlighting selected issues of business negotiations from agency-theory perspective: Negotiation Teams require a multi-agent approach under the condition that often decision-makers as superior-agents are part of the team. The diversity of competences and decision-making authority is a phenomenon that overrides the assumptions of classical agency-theory and varies greatly in context of certain forms of business negotiations. Further, the basic model is bound to dyadic relationships preceded by the delegation of decision-making authority and builds on a contractual created (vertical) hierarchy. As a result, horizontal dynamics within the negotiation team playing an important role for negotiation success are therefore not considered in the investigation of agency-problems. Also, the trade-off between short-term relationships within the negotiation sphere and the long-term relationships of the corporate sphere calls for a multi-period perspective taking into account the sphere-specific governance-mechanisms already established (i.e., reward and monitoring systems). Within the analysis, the implementation of bounded rationality is closely related to findings from BRD to assess the impact of negotiation behavior on underlying principal-agent-relationships. As empirical findings show, the disclosure and reservation of information to the agent affect his negotiation behavior as well as final negotiation outcomes. Last, in context of business negotiations, asymmetric information is often intended by decision-makers acting as superior-agents or principals which calls for a bilateral risk-approach to agency-relations.

Keywords: business negotiations, agency-theory, negotiation analysis, interteam negotiations

Procedia PDF Downloads 125
11133 Machine Learning for Targeting of Conditional Cash Transfers: Improving the Effectiveness of Proxy Means Tests to Identify Future School Dropouts and the Poor

Authors: Cristian Crespo

Abstract:

Conditional cash transfers (CCTs) have been targeted towards the poor. Thus, their targeting assessments check whether these schemes have been allocated to low-income households or individuals. However, CCTs have more than one goal and target group. An additional goal of CCTs is to increase school enrolment. Hence, students at risk of dropping out of school also are a target group. This paper analyses whether one of the most common targeting mechanisms of CCTs, a proxy means test (PMT), is suitable to identify the poor and future school dropouts. The PMT is compared with alternative approaches that use the outputs of a predictive model of school dropout. This model was built using machine learning algorithms and rich administrative datasets from Chile. The paper shows that using machine learning outputs in conjunction with the PMT increases targeting effectiveness by identifying more students who are either poor or future dropouts. This joint targeting approach increases effectiveness in different scenarios except when the social valuation of the two target groups largely differs. In these cases, the most likely optimal approach is to solely adopt the targeting mechanism designed to find the highly valued group.

Keywords: conditional cash transfers, machine learning, poverty, proxy means tests, school dropout prediction, targeting

Procedia PDF Downloads 187
11132 Embodied Spiritualities and Emerging Search for Social Transformation: An Embodied Ethnographic Study of Yoga Practices in Medellin, Colombia

Authors: Lina M. Vidal

Abstract:

This paper discusses yoga practices involvement in both self-transformation and social transformations by means of an embodied ethnographic approach to different initiatives for social change in Medellín. In the context of gradual popularization of embodied spiritualities, yoga practices have opened their way in calls for social change in a performative perspective which involves collective experiences, reflections and production of embodied knowledge. Through the reflection on bodily dimension and corporal experience, this ethnographic approach acknowledges inter-corporality and somatic modes of attention during observations and personal experiences. In social change initiatives that include yoga practices were identified transformations of common understanding on social issues such as it is produced by institutionalized education, health system and other fields of knowledge. This is clearly visible in yoga projects for children in vulnerable conditions, homeless people, prisoners, and young people recovering from drug addiction. These projects are often promoted by organizations and networks, which incorporate individual life stories into collective experiences. Dissemination of yoga is heading to a broad institutional and cultural legitimation of yoga and of spirituality that impact different fields of social work and everyday life in general. This way, yoga is becoming an embodied activist way of life and a legitimate field for social work.

Keywords: embodied ethnography, Medellin, social transformation, embodied spiritualities, yoga practices

Procedia PDF Downloads 168
11131 Academic Staff Recruitment in Islamic University: A Proposed Holistic Model

Authors: Syahruddin Sumardi Samindjaya, Indra Fajar Alamsyah, Junaidah Hashim

Abstract:

Purpose: This study attempts to explore and presents a proposed recruitment model in Islamic university which aligned with holistic role. Design/methodology/approach: It is a conceptual paper in nature. In turn, this study is designed to utilize exploratory approach. Literature and document review that related to this topic are used as the methods to analyse the content found. Findings: Recruitment for any organization is fundamental to achieve its goal effectively. Staffing in universities is vital due to the important role of lecturers. Currently, Islamic universities still adopt the common process of recruitment for their academic staffs. Whereas, they have own characteristics which are embedded in their institutions. Furthermore, the FCWC (Foundation, Capability, Worldview and Commitment) model of recruitment proposes to suit the holistic character of Islamic university. Research limitation/implications: Further studies are required to empirically validate the concept through systematic investigations. Additionally, measuring this model by a designed means is appreciated. Practical implications: The model provides the map and alternative tool of recruitment for Islamic universities to determine the process of recruitment which can appropriate their institutions. In addition, it also allows stakeholders and policy makers to consider regarding Islamic values that should inculcate in the Islamic higher learning institutions. Originality/value: This study initiates a foundational contribution for an early sequence of research.

Keywords: academic staff, Islamic values, recruitment model, university

Procedia PDF Downloads 169
11130 Joint Modeling of Longitudinal and Time-To-Event Data with Latent Variable

Authors: Xinyuan Y. Song, Kai Kang

Abstract:

Joint models for analyzing longitudinal and survival data are widely used to investigate the relationship between a failure time process and time-variant predictors. A common assumption in conventional joint models in the survival analysis literature is that all predictors are observable. However, this assumption may not always be supported because unobservable traits, namely, latent variables, which are indirectly observable and should be measured through multiple observed variables, are commonly encountered in the medical, behavioral, and financial research settings. In this study, a joint modeling approach to deal with this feature is proposed. The proposed model comprises three parts. The first part is a dynamic factor analysis model for characterizing latent variables through multiple observed indicators over time. The second part is a random coefficient trajectory model for describing the individual trajectories of latent variables. The third part is a proportional hazard model for examining the effects of time-invariant predictors and the longitudinal trajectories of time-variant latent risk factors on hazards of interest. A Bayesian approach coupled with a Markov chain Monte Carlo algorithm to perform statistical inference. An application of the proposed joint model to a study on the Alzheimer's disease neuroimaging Initiative is presented.

Keywords: Bayesian analysis, joint model, longitudinal data, time-to-event data

Procedia PDF Downloads 128
11129 Material Supply Mechanisms for Contemporary Assembly Systems

Authors: Rajiv Kumar Srivastava

Abstract:

Manufacturing of complex products such as automobiles and computers requires a very large number of parts and sub-assemblies. The design of mechanisms for delivery of these materials to the point of assembly is an important manufacturing system and supply chain challenge. Different approaches to this problem have been evolved for assembly lines designed to make large volumes of standardized products. However, contemporary assembly systems are required to concurrently produce a variety of products using approaches such as mixed model production, and at times even mass customization. In this paper we examine the material supply approaches for variety production in moderate to large volumes. The conventional approach for material delivery to high volume assembly lines is to supply and stock materials line-side. However for certain materials, especially when the same or similar items are used along the line, it is more convenient to supply materials in kits. Kitting becomes more preferable when lines concurrently produce multiple products in mixed model mode, since space requirements could increase as product/ part variety increases. At times such kits may travel along with the product, while in some situations it may be better to have delivery and station-specific kits rather than product-based kits. Further, in some mass customization situations it may even be better to have a single delivery and assembly station, to which an entire kit is delivered for fitment, rather than a normal assembly line. Finally, in low-moderate volume assembly such as in engineered machinery, it may be logistically more economical to gather materials in an order-specific kit prior to launching final assembly. We have studied material supply mechanisms to support assembly systems as observed in case studies of firms with different combinations of volume and variety/ customization. It is found that the appropriate approach tends to be a hybrid between direct line supply and different kitting modes, with the best mix being a function of the manufacturing and supply chain environment, as well as space and handling considerations. In our continuing work we are studying these scenarios further, through the use of descriptive models and progressing towards prescriptive models to help achieve the optimal approach, capturing the trade-offs between inventory, material handling, space, and efficient line supply.

Keywords: assembly systems, kitting, material supply, variety production

Procedia PDF Downloads 211
11128 A Simple Approach to Reliability Assessment of Structures via Anomaly Detection

Authors: Rims Janeliukstis, Deniss Mironovs, Andrejs Kovalovs

Abstract:

Operational Modal Analysis (OMA) is widely applied as a method for Structural Health Monitoring for structural damage identification and assessment by tracking the changes of the identified modal parameters over time. Unfortunately, modal parameters also depend on such external factors as temperature and loads. Any structural condition assessment using modal parameters should be done taking into consideration those external factors, otherwise there is a high chance of false positives. A method of structural reliability assessment based on anomaly detection technique called Machalanobis Squared Distance (MSD) is proposed. It requires a set of reference conditions to learn healthy state of a structure, which all future parameters are compared to. In this study, structural modal parameters (natural frequency and mode shape), as well as ambient temperature and loads acting on the structure are used as features. Numerical tests were performed on a finite element model of a carbon fibre reinforced polymer composite beam with delamination damage at various locations and of various severities. The advantages of the demonstrated approach include relatively few computational steps, ability to distinguish between healthy and damaged conditions and discriminate between different damage severities. It is anticipated to be promising in reliability assessment of massively produced structural parts.

Keywords: operational modal analysis, reliability assessment, anomaly detection, damage, mahalanobis squared distance

Procedia PDF Downloads 92
11127 An Integrated Approach to the Carbonate Reservoir Modeling: Case Study of the Eastern Siberia Field

Authors: Yana Snegireva

Abstract:

Carbonate reservoirs are known for their heterogeneity, resulting from various geological processes such as diagenesis and fracturing. These complexities may cause great challenges in understanding fluid flow behavior and predicting the production performance of naturally fractured reservoirs. The investigation of carbonate reservoirs is crucial, as many petroleum reservoirs are naturally fractured, which can be difficult due to the complexity of their fracture networks. This can lead to geological uncertainties, which are important for global petroleum reserves. The problem outlines the key challenges in carbonate reservoir modeling, including the accurate representation of fractures and their connectivity, as well as capturing the impact of fractures on fluid flow and production. Traditional reservoir modeling techniques often oversimplify fracture networks, leading to inaccurate predictions. Therefore, there is a need for a modern approach that can capture the complexities of carbonate reservoirs and provide reliable predictions for effective reservoir management and production optimization. The modern approach to carbonate reservoir modeling involves the utilization of the hybrid fracture modeling approach, including the discrete fracture network (DFN) method and implicit fracture network, which offer enhanced accuracy and reliability in characterizing complex fracture systems within these reservoirs. This study focuses on the application of the hybrid method in the Nepsko-Botuobinskaya anticline of the Eastern Siberia field, aiming to prove the appropriateness of this method in these geological conditions. The DFN method is adopted to model the fracture network within the carbonate reservoir. This method considers fractures as discrete entities, capturing their geometry, orientation, and connectivity. But the method has significant disadvantages since the number of fractures in the field can be very high. Due to limitations in the amount of main memory, it is very difficult to represent these fractures explicitly. By integrating data from image logs (formation micro imager), core data, and fracture density logs, a discrete fracture network (DFN) model can be constructed to represent fracture characteristics for hydraulically relevant fractures. The results obtained from the DFN modeling approaches provide valuable insights into the East Siberia field's carbonate reservoir behavior. The DFN model accurately captures the fracture system, allowing for a better understanding of fluid flow pathways, connectivity, and potential production zones. The analysis of simulation results enables the identification of zones of increased fracturing and optimization opportunities for reservoir development with the potential application of enhanced oil recovery techniques, which were considered in further simulations on the dual porosity and dual permeability models. This approach considers fractures as separate, interconnected flow paths within the reservoir matrix, allowing for the characterization of dual-porosity media. The case study of the East Siberia field demonstrates the effectiveness of the hybrid model method in accurately representing fracture systems and predicting reservoir behavior. The findings from this study contribute to improved reservoir management and production optimization in carbonate reservoirs with the use of enhanced and improved oil recovery methods.

Keywords: carbonate reservoir, discrete fracture network, fracture modeling, dual porosity, enhanced oil recovery, implicit fracture model, hybrid fracture model

Procedia PDF Downloads 61
11126 Understanding of Corporate Social Responsibility and Non-Governmental Organizations

Authors: Abdul Ghafar, Malini Nair

Abstract:

Non-governmental organizations have been seemed to struggle the battle of balancing many concerns with corporates which may impact on their financial solvency. Some of these concerns relates to uphold the relationship where weighing up the impacts of their involvement with corporates takes priority over the main purpose of creating valuable impacts for communities. To some extent, it can be argued that NGOs are influenced by corporates’ power to tackle contemporary issues rather than eradicating the root causes of such issues and transform the results into more sustainable manner. NGOs spend massive amount of energy, time and resources in order to move some corporates to embrace their social responsibilities. It has become a norm, where an active NGO that is becoming more successful on building partnerships with corporates is perceived to be more socially responsible. In contrast to this, as some researchers argue that the social responsibility for NGOs is not a voluntary act; they must exhibit the core values in all their practices require much attention to address. This article stresses the need of understanding ‘Social Responsibility’ of NGOs that stem from an argument that NGOs tend to act on narrow mandate rather than considering broader outcomes of their CSR initiatives. This paper argues that NGOs must focus on building capabilities of the recipients from CSR initiatives which should serve as a core value of partnerships mandate between NGOs, Corporates and Governments. We argue that SEN’s Capabilities Approach can further enhance the mainstream CSR agenda of NGOs which seems to incline more towards providing palliative solutions to social issues.

Keywords: non-profit organization, corporate social responsibility, partnerships, capabilities approach

Procedia PDF Downloads 217
11125 Reconstructed Phase Space Features for Estimating Post Traumatic Stress Disorder

Authors: Andre Wittenborn, Jarek Krajewski

Abstract:

Trauma-related sadness in speech can alter the voice in several ways. The generation of non-linear aerodynamic phenomena within the vocal tract is crucial when analyzing trauma-influenced speech production. They include non-laminar flow and formation of jets rather than well-behaved laminar flow aspects. Especially state-space reconstruction methods based on chaotic dynamics and fractal theory have been suggested to describe these aerodynamic turbulence-related phenomena of the speech production system. To extract the non-linear properties of the speech signal, we used the time delay embedding method to reconstruct from a scalar time series (reconstructed phase space, RPS). This approach results in the extraction of 7238 Features per .wav file (N= 47, 32 m, 15 f). The speech material was prompted by telling about autobiographical related sadness-inducing experiences (sampling rate 16 kHz, 8-bit resolution). After combining these features in a support vector machine based machine learning approach (leave-one-sample out validation), we achieved a correlation of r = .41 with the well-established, self-report ground truth measure (RATS) of post-traumatic stress disorder (PTSD).

Keywords: non-linear dynamics features, post traumatic stress disorder, reconstructed phase space, support vector machine

Procedia PDF Downloads 93
11124 Detection of Egg Proteins in Food Matrices (2011-2021)

Authors: Daniela Manila Bianchi, Samantha Lupi, Elisa Barcucci, Sandra Fragassi, Clara Tramuta, Lucia Decastelli

Abstract:

Introduction: The undeclared allergens detection in food products plays a fundamental role in the safety of the allergic consumer. The protection of allergic consumers is guaranteed, in Europe, by Regulation (EU) No 1169/2011 of the European Parliament, which governs the consumer's right to information and identifies 14 food allergens to be mandatorily indicated on food labels: among these, an egg is included. An egg can be present as an ingredient or as contamination in raw and cooked products. The main allergen egg proteins are ovomucoid, ovalbumin, lysozyme, and ovotransferrin. This study presents the results of a survey conducted in Northern Italy aimed at detecting the presence of undeclared egg proteins in food matrices in the latest ten years (2011-2021). Method: In the period January 2011 - October 2021, a total of 1205 different types of food matrices (ready-to-eat, meats, and meat products, bakery and pastry products, baby foods, food supplements, pasta, fish and fish products, preparations for soups and broths) were delivered to Food Control Laboratory of Istituto Zooprofilattico Sperimentale of Piemonte Liguria and Valle d’Aosta to be analyzed as official samples in the frame of Regional Monitoring Plan of Food Safety or in the contest of food poisoning. The laboratory is ISO 17025 accredited, and since 2019, it has represented the National Reference Centre for the detection in foods of substances causing food allergies or intolerances (CreNaRiA). All samples were stored in the laboratory according to food business operator instructions and analyzed within the expiry date for the detection of undeclared egg proteins. Analyses were performed with RIDASCREEN®FAST Ei/Egg (R-Biopharm ® Italia srl) kit: the method was internally validated and accredited with a Limit of Detection (LOD) equal to 2 ppm (mg/Kg). It is a sandwich enzyme immunoassay for the quantitative analysis of whole egg powder in foods. Results: The results obtained through this study showed that egg proteins were found in 2% (n. 28) of food matrices, including meats and meat products (n. 16), fish and fish products (n. 4), bakery and pastry products (n. 4), pasta (n. 2), preparations for soups and broths (n.1) and ready-to-eat (n. 1). In particular, in 2011 egg proteins were detected in 5% of samples, in 2012 in 4%, in 2013, 2016 and 2018 in 2%, in 2014, 2015 and 2019 in 3%. No egg protein traces were detected in 2017, 2020, and 2021. Discussion: Food allergies occur in the Western World in 2% of adults and up to 8% of children. Allergy to eggs is one of the most common food allergies in the pediatrics context. The percentage of positivity obtained from this study is, however, low. The trend over the ten years has been slightly variable, with comparable data.

Keywords: allergens, food, egg proteins, immunoassay

Procedia PDF Downloads 121
11123 A Framework for Designing Complex Product-Service Systems with a Multi-Domain Matrix

Authors: Yoonjung An, Yongtae Park

Abstract:

Offering a Product-Service System (PSS) is a well-accepted strategy that companies may adopt to provide a set of systemic solutions to customers. PSSs were initially provided in a simple form but now take diversified and complex forms involving multiple services, products and technologies. With the growing interest in the PSS, frameworks for the PSS development have been introduced by many researchers. However, most of the existing frameworks fail to examine various relations existing in a complex PSS. Since designing a complex PSS involves full integration of multiple products and services, it is essential to identify not only product-service relations but also product-product/ service-service relations. It is also equally important to specify how they are related for better understanding of the system. Moreover, as customers tend to view their purchase from a more holistic perspective, a PSS should be developed based on the whole system’s requirements, rather than focusing only on the product requirements or service requirements. Thus, we propose a framework to develop a complex PSS that is coordinated fully with the requirements of both worlds. Specifically, our approach adopts a multi-domain matrix (MDM). A MDM identifies not only inter-domain relations but also intra-domain relations so that it helps to design a PSS that includes highly desired and closely related core functions/ features. Also, various dependency types and rating schemes proposed in our approach would help the integration process.

Keywords: inter-domain relations, intra-domain relations, multi-domain matrix, product-service system design

Procedia PDF Downloads 629
11122 Evaluation of the Effect of Auriculotherapy on Pain Control and Sleep Quality in Chronic Patients

Authors: Fagner Luiz P. Salles, Janaina C. Oliveira, Ivair P. Cesar

Abstract:

Statement of the Problem: Auriculotherapy (AT) is a TCM technique, which uses seeds instead of needles, based physiologically on the mechanical stimulation of the cranial nerves. In the context of understanding the new concept of health of the WHO, the AT is an integrative approach for achieving Global Health Care so as to achieve the global health care concerns. This study aimed to evaluate the effect of auriculotherapy on pain and sleep quality in patients with chronic pain. Methodology and Theoretical Orientation: This study was performed between February and March 2017 at the Faculdade Estácio de Sá de Vitória, Brazil. The pain evaluation was through VAS in 4 periods: maximum, minimum, average and at the time of evaluation; the evaluation of sleep quality was used the Pittsburgh Sleep Quality Index. Socio-demographic data included: gender, age, use of medication and BMI. All data are presented as mean (standard deviation), Teste Mann-Whitney and T-student with P-values < 0.05 were regarded as significant. Findings: Participated in this study thirty-two individuals with age (M = 43.18, SD = 17.86), the time with pain in years (M = 3.67, SD = 3.68), 81.7% were female, 75% of the individuals used medication and BMI (M = 26.67; SD = 6.20). The pain presented improvement in the maximum level and the average of the pain and sleep quality before did not have statistically significant results. Conclusion and Significance: This study showed that TA is efficacy for reduction levels of pain. However, AT was not effective in improving sleep quality.

Keywords: auriculotherapy, chronic pain, sleep quality, integrative approach

Procedia PDF Downloads 194
11121 Knowledge Loss Risk Assessment for Departing Employees: An Exploratory Study

Authors: Muhammad Saleem Ullah Khan Sumbal, Eric Tsui, Ricky Cheong, Eric See To

Abstract:

Organizations are posed to a threat of valuable knowledge loss when employees leave either due to retirement, resignation, job change or because of disabilities e.g. death, etc. Due to changing economic conditions, globalization, and aging workforce, organizations are facing challenges regarding retention of valuable knowledge. On the one hand, large number of employees are going to retire in the organizations whereas on the other hand, younger generation does not want to work in a company for a long time and there is an increasing trend of frequent job change among the new generation. Because of these factors, organizations need to make sure that they capture the knowledge of employee before (s)he walks out of the door. The first step in this process is to know what type of knowledge employee possesses and whether this knowledge is important for the organization. Researchers reveal in the literature that despite the serious consequences of knowledge loss in terms of organizational productivity and competitive advantage, there has not been much work done in the area of knowledge loss assessment of departing employees. An important step in the knowledge retention process is to determine the critical ‘at risk’ knowledge. Thus, knowledge loss risk assessment is a process by which organizations can gauge the importance of knowledge of the departing employee. The purpose of this study is to explore this topic of knowledge loss risk assessment by conducting a qualitative study in oil and gas sector. By engaging in dialogues with managers and executives of the organizations through in-depth interviews and adopting a grounded methodology approach, the research will explore; i) Are there any measures adopted by organizations to assess the risk of knowledge loss from departing employees? ii) Which factors are crucial for knowledge loss assessment in the organizations? iii) How can we prioritize the employees for knowledge retention according to their criticality? Grounded theory approach is used when there is not much knowledge available in the area under research and thus new knowledge is generated about the topic through an in-depth exploration of the topic by using methods such as interviews and using a systematic approach to analyze the data. The outcome of the study will generate a model for the risk of knowledge loss through factors such as the likelihood of knowledge loss, the consequence/impact of knowledge loss and quality of the knowledge loss of departing employees. Initial results show that knowledge loss assessment is quite crucial for the organizations and it helps in determining what types of knowledge employees possess e.g. organizations knowledge, subject matter expertise or relationships knowledge. Based on that, it can be assessed which employee is more important for the organizations and how to prioritize the knowledge retention process for departing employees.

Keywords: knowledge loss, risk assessment, departing employees, Hong Kong organizations

Procedia PDF Downloads 390
11120 Segmentation of the Liver and Spleen From Abdominal CT Images Using Watershed Approach

Authors: Belgherbi Aicha, Hadjidj Ismahen, Bessaid Abdelhafid

Abstract:

The phase of segmentation is an important step in the processing and interpretation of medical images. In this paper, we focus on the segmentation of liver and spleen from the abdomen computed tomography (CT) images. The importance of our study comes from the fact that the segmentation of ROI from CT images is usually a difficult task. This difficulty is the gray’s level of which is similar to the other organ also the ROI are connected to the ribs, heart, kidneys, etc. Our proposed method is based on the anatomical information and mathematical morphology tools used in the image processing field. At first, we try to remove the surrounding and connected organs and tissues by applying morphological filters. This first step makes the extraction of interest regions easier. The second step consists of improving the quality of the image gradient. In this step, we propose a method for improving the image gradient to reduce these deficiencies by applying the spatial filters followed by the morphological filters. Thereafter we proceed to the segmentation of the liver, spleen. To validate the segmentation technique proposed, we have tested it on several images. Our segmentation approach is evaluated by comparing our results with the manual segmentation performed by an expert. The experimental results are described in the last part of this work.The system has been evaluated by computing the sensitivity and specificity between the semi-automatically segmented (liver and spleen) contour and the manually contour traced by radiological experts.

Keywords: CT images, liver and spleen segmentation, anisotropic diffusion filter, morphological filters, watershed algorithm

Procedia PDF Downloads 477
11119 Developing a Systemic Monoclonal Antibody Therapy for the Treatment of Large Burn Injuries

Authors: Alireza Hassanshahi, Xanthe Strudwick, Zlatko Kopecki, Allison J Cowin

Abstract:

Studies have shown that Flightless (Flii) is elevated in human wounds, including burns, and reducing the level of Flii is a promising approach for improving wound repair and reducing scar formation. The most effective approach has been to neutralise Flii activity using localized, intradermal application of function blocking monoclonal antibodies. However, large surface area burns are difficult to treat by intradermal injection of therapeutics, so the aim of this study was to investigate if a systemic injection of a monoclonal antibody against Flii could improve healing in mice following burn injury. Flii neutralizing antibodies (FnAbs) were labelled with Alxa-Fluor-680 for biodistribution studies and the healing effects of systemically administered FnAbs to mice with burn injuries. A partial thickness, 7% (70mm2) total body surface area scald burn injury was created on the dorsal surface of mice (n=10/group), and 100µL of Alexa-Flour-680-labeled FnAbs were injected into the intraperitoneal cavity (IP) at time of injury. The burns were imaged on days 0, 1, 2, 3, 4, and 7 using IVIS Lumina S5 Imaging System, and healing was assessed macroscopically, histologically, and using immunohistochemistry. Fluorescent radiance efficiency measurements showed that IP injected Alexa-Fluor-680-FnAbs localized at the site of burn injury from day 1, remaining there for the whole 7-day study. The burns treated with FnAbs showed a reduction in macroscopic wound area and an increased rate of epithelialization compared to controls. Immunohistochemistry for NIMP-R14 showed a reduction in the inflammatory infiltrate, while CD31/VEGF staining showed improved angiogenesis post-systemic FnAb treatment. These results suggest that systemically administered FnAbs are active within the burn site and can improve healing outcomes. The clinical application of systemically injected Flii monoclonal antibodies could therefore be a potential approach for promoting the healing of large surface area burns immediately after injury.

Keywords: biodistribution, burn, flightless, systemic, fnAbs

Procedia PDF Downloads 158
11118 Dynamic Modeling of Energy Systems Adapted to Low Energy Buildings in Lebanon

Authors: Nadine Yehya, Chantal Maatouk

Abstract:

Low energy buildings have been developed to achieve global climate commitments in reducing energy consumption. They comprise energy efficient buildings, zero energy buildings, positive buildings and passive house buildings. The reduced energy demands in Low Energy buildings call for advanced building energy modeling that focuses on studying active building systems such as heating, cooling and ventilation, improvement of systems performances, and development of control systems. Modeling and building simulation have expanded to cover different modeling approach i.e.: detailed physical model, dynamic empirical models, and hybrid approaches, which are adopted by various simulation tools. This paper uses DesignBuilder with EnergyPlus simulation engine in order to; First, study the impact of efficiency measures on building energy behavior by comparing Low energy residential model to a conventional one in Beirut-Lebanon. Second, choose the appropriate energy systems for the studied case characterized by an important cooling demand. Third, study dynamic modeling of Variable Refrigerant Flow (VRF) system in EnergyPlus that is chosen due to its advantages over other systems and its availability in the Lebanese market. Finally, simulation of different energy systems models with different modeling approaches is necessary to confront the different modeling approaches and to investigate the interaction between energy systems and building envelope that affects the total energy consumption of Low Energy buildings.

Keywords: physical model, variable refrigerant flow heat pump, dynamic modeling, EnergyPlus, the modeling approach

Procedia PDF Downloads 204
11117 Recognizing Customer Preferences Using Review Documents: A Hybrid Text and Data Mining Approach

Authors: Oshin Anand, Atanu Rakshit

Abstract:

The vast increment in the e-commerce ventures makes this area a prominent research stream. Besides several quantified parameters, the textual content of reviews is a storehouse of many information that can educate companies and help them earn profit. This study is an attempt in this direction. The article attempts to categorize data based on a computed metric that quantifies the influencing capacity of reviews rendering two categories of high and low influential reviews. Further, each of these document is studied to conclude several product feature categories. Each of these categories along with the computed metric is converted to linguistic identifiers and are used in an association mining model. The article makes a novel attempt to combine feature attraction with quantified metric to categorize review text and finally provide frequent patterns that depict customer preferences. Frequent mentions in a highly influential score depict customer likes or preferred features in the product whereas prominent pattern in low influencing reviews highlights what is not important for customers. This is achieved using a hybrid approach of text mining for feature and term extraction, sentiment analysis, multicriteria decision-making technique and association mining model.

Keywords: association mining, customer preference, frequent pattern, online reviews, text mining

Procedia PDF Downloads 376
11116 Investigation on Ultrahigh Heat Flux of Nanoporous Membrane Evaporation Using Dimensionless Lattice Boltzmann Method

Authors: W. H. Zheng, J. Li, F. J. Hong

Abstract:

Thin liquid film evaporation in ultrathin nanoporous membranes, which reduce the viscous resistance while still maintaining high capillary pressure and efficient liquid delivery, is a promising thermal management approach for high-power electronic devices cooling. Given the challenges and technical limitations of experimental studies for accurate interface temperature sensing, complex manufacturing process, and short duration of membranes, a dimensionless lattice Boltzmann method capable of restoring thermophysical properties of working fluid is particularly derived. The evaporation of R134a to its pure vapour ambient in nanoporous membranes with the pore diameter of 80nm, thickness of 472nm, and three porosities of 0.25, 0.33 and 0.5 are numerically simulated. The numerical results indicate that the highest heat transfer coefficient is about 1740kW/m²·K; the highest heat flux is about 1.49kW/cm² with only about the wall superheat of 8.59K in the case of porosity equals to 0.5. The dissipated heat flux scaled with porosity because of the increasing effective evaporative area. Additionally, the self-regulation of the shape and curvature of the meniscus under different operating conditions is also observed. This work shows a promising approach to forecast the membrane performance for different geometry and working fluids.

Keywords: high heat flux, ultrathin nanoporous membrane, thin film evaporation, lattice Boltzmann method

Procedia PDF Downloads 149
11115 Going beyond Stakeholder Participation

Authors: Florian Engel

Abstract:

Only with a radical change to an intrinsically motivated project team, through giving the employees the freedom for autonomy, mastery and purpose, it is then possible to develop excellent products. With these changes, combined with using a rapid application development approach, the group of users serves as an important indicator to test the market needs, rather than only as the stakeholders for requirements.

Keywords: intrinsic motivation, requirements elicitation, self-directed work, stakeholder participation

Procedia PDF Downloads 325
11114 Statistical Pattern Recognition for Biotechnological Process Characterization Based on High Resolution Mass Spectrometry

Authors: S. Fröhlich, M. Herold, M. Allmer

Abstract:

Early stage quantitative analysis of host cell protein (HCP) variations is challenging yet necessary for comprehensive bioprocess development. High resolution mass spectrometry (HRMS) provides a high-end technology for accurate identification alongside with quantitative information. Hereby we describe a flexible HRMS assay platform to quantify HCPs relevant in microbial expression systems such as E. Coli in both up and downstream development by means of MVDA tools. Cell pellets were lysed and proteins extracted, purified samples not further treated before applying the SMART tryptic digest kit. Peptides separation was optimized using an RP-UHPLC separation platform. HRMS-MSMS analysis was conducted on an Orbitrap Velos Elite applying CID. Quantification was performed label-free taking into account ionization properties and physicochemical peptide similarities. Results were analyzed using SIEVE 2.0 (Thermo Fisher Scientific) and SIMCA (Umetrics AG). The developed HRMS platform was applied to an E. Coli expression set with varying productivity and the corresponding downstream process. Selected HCPs were successfully quantified within the fmol range. Analysing HCP networks based on pattern analysis facilitated low level quantification and enhanced validity. This approach is of high relevance for high-throughput screening experiments during upstream development, e.g. for titer determination, dynamic HCP network analysis or product characterization. Considering the downstream purification process, physicochemical clustering of identified HCPs is of relevance to adjust buffer conditions accordingly. However, the technology provides an innovative approach for label-free MS based quantification relying on statistical pattern analysis and comparison. Absolute quantification based on physicochemical properties and peptide similarity score provides a technological approach without the need of sophisticated sample preparation strategies and is therefore proven to be straightforward, sensitive and highly reproducible in terms of product characterization.

Keywords: process analytical technology, mass spectrometry, process characterization, MVDA, pattern recognition

Procedia PDF Downloads 237
11113 Comparison of Number of Waves Surfed and Duration Using Global Positioning System and Inertial Sensors

Authors: João Madureira, Ricardo Lagido, Inês Sousa, Fraunhofer Portugal

Abstract:

Surf is an increasingly popular sport and its performance evaluation is often qualitative. This work aims at using a smartphone to collect and analyze the GPS and inertial sensors data in order to obtain quantitative metrics of the surfing performance. Two approaches are compared for detection of wave rides, computing the number of waves rode in a surfing session, the starting time of each wave and its duration. The first approach is based on computing the velocity from the Global Positioning System (GPS) signal and finding the velocity thresholds that allow identifying the start and end of each wave ride. The second approach adds information from the Inertial Measurement Unit (IMU) of the smartphone, to the velocity thresholds obtained from the GPS unit, to determine the start and end of each wave ride. The two methods were evaluated using GPS and IMU data from two surfing sessions and validated with similar metrics extracted from video data collected from the beach. The second method, combining GPS and IMU data, was found to be more accurate in determining the number of waves, start time and duration. This paper shows that it is feasible to use smartphones for quantification of performance metrics during surfing. In particular, detection of the waves rode and their duration can be accurately determined using the smartphone GPS and IMU.

Keywords: inertial measurement unit (IMU), global positioning system (GPS), smartphone, surfing performance

Procedia PDF Downloads 389
11112 Phytoremediation Waste Processing of Coffee in Various Concentration of Organic Materials Plant Using Kiambang

Authors: Siti Aminatu Zuhria

Abstract:

On wet coffee processing can improve the quality of coffee, but the coffee liquid waste that can pollute the environment. Liquid waste a lot of coffee resulting from the stripping and washing the coffee. This research will be carried out the process of handling liquid waste stripping coffee from the coffee skin with media phytoremediation using plants kiambang. The purpose of this study was to determine the characteristics of the coffee liquid waste and plant phytoremediation kiambang as agent in various concentrations of liquid waste coffee as well as determining the most optimal concentration in the improved quality of waste water quality standard approach. This research will be conducted through two stages, namely the preliminary study and the main study. In a preliminary study aims to determine the ability of the plant life kiambang as phytoremediation agent in the media well water, distilled water and liquid waste coffee. The main study will be conducted wastewater dilution and coffee will be obtained COD concentration variations. Results are expected at this research that can determine the ability of plants kiambang as an agent for phytoremediation in wastewater treatment with various concentrations of waste and the most optimal concentration in the improved quality of waste water quality standard approach.

Keywords: wet coffee processing, phytoremediation, Kiambang plant, variation concentration liquid waste

Procedia PDF Downloads 290
11111 A Fuzzy-Rough Feature Selection Based on Binary Shuffled Frog Leaping Algorithm

Authors: Javad Rahimipour Anaraki, Saeed Samet, Mahdi Eftekhari, Chang Wook Ahn

Abstract:

Feature selection and attribute reduction are crucial problems, and widely used techniques in the field of machine learning, data mining and pattern recognition to overcome the well-known phenomenon of the Curse of Dimensionality. This paper presents a feature selection method that efficiently carries out attribute reduction, thereby selecting the most informative features of a dataset. It consists of two components: 1) a measure for feature subset evaluation, and 2) a search strategy. For the evaluation measure, we have employed the fuzzy-rough dependency degree (FRFDD) of the lower approximation-based fuzzy-rough feature selection (L-FRFS) due to its effectiveness in feature selection. As for the search strategy, a modified version of a binary shuffled frog leaping algorithm is proposed (B-SFLA). The proposed feature selection method is obtained by hybridizing the B-SFLA with the FRDD. Nine classifiers have been employed to compare the proposed approach with several existing methods over twenty two datasets, including nine high dimensional and large ones, from the UCI repository. The experimental results demonstrate that the B-SFLA approach significantly outperforms other metaheuristic methods in terms of the number of selected features and the classification accuracy.

Keywords: binary shuffled frog leaping algorithm, feature selection, fuzzy-rough set, minimal reduct

Procedia PDF Downloads 201
11110 Aspiring to Achieve a Fairer Society

Authors: Bintou Jobe

Abstract:

Background: The research is focused on the concept of equality, diversity, and inclusion (EDI) and the need to achieve equity by treating individuals according to their circumstances and needs. The research is rooted in the UK Equality Act 2010, which emphasizes the importance of equal opportunities for all individuals regardless of their background and social life. However, inequality persists in society, particularly for those from minority backgrounds who face discrimination. Research Aim: The aim of this research is to promote equality, diversity, and inclusion by encouraging the regeneration of minds and the eradication of stereotypes. The focus is on promoting good Equality, Diversity and Inclusion practices in various settings, including schools, colleges, universities, and workplaces, to create environments where every individual feels a sense of belonging. Methodology: The research utilises a literature review approach to gather information on promoting inclusivity, diversity, and inclusion. Findings: The research highlights the significance of promoting equality, diversity, and inclusion practices to ensure that individuals receive the respect and dignity they deserve. It emphasises the importance of treating individuals based on their unique circumstances and needs rather than relying on stereotypes. The research also emphasises the benefits of diversity and inclusion in enhancing innovation, creativity, and productivity. The theoretical importance of this research is to raise awareness about the importance of regenerating minds, challenging stereotypes, and promoting equality, diversity, and inclusion. The emphasis is on treating individuals based on their circumstances and needs rather than relying on generalizations. Diversity and inclusion are beneficial in different settings, as highlighted by the research. By raising awareness about the importance of mind regeneration, eradicating stereotypes, and promoting equality, diversity, and inclusion, this research makes a significant contribution to the subject area. It emphasizes the necessity of treating individuals based on their unique circumstances instead of relying on generalizations. However, the methodology could be strengthened by incorporating primary research to complement the literature review approach. Data Collection and Analysis Procedures: The research utilised a literature review approach to gather relevant information on promoting inclusivity, diversity, and inclusion. NVivo software application was used to analysed and synthesize the findings to identify themes and support the research aim and objectives. Question Addressed: This research addresses the question of how to promote inclusivity, diversity, and inclusion and reduce the prevalence of stereotypes and prejudice. It explores the need to treat individuals based on their unique circumstances and needs rather than relying on generic assumptions. Encourage individuals to adopt a more inclusive approach. Provide managers with responsibility and training that helps them understand the importance of their roles in shaping the workplace culture. Have an equality, diversity, and inclusion manager from a majority background at the senior level who can speak up for underrepresented groups and flag any issues that need addressing. Conclusion: The research emphasizes the importance of promoting equality, diversity, and inclusion practices to create a fairer society. It highlights the need to challenge stereotypes, treat individuals according to their circumstances and needs, and promote a culture of respect and dignity.

Keywords: equality, fairer society, inclusion, diversity

Procedia PDF Downloads 37
11109 Automated Evaluation Approach for Time-Dependent Question Answering Pairs on Web Crawler Based Question Answering System

Authors: Shraddha Chaudhary, Raksha Agarwal, Niladri Chatterjee

Abstract:

This work demonstrates a web crawler-based generalized end-to-end open domain Question Answering (QA) system. An efficient QA system requires a significant amount of domain knowledge to answer any question with the aim to find an exact and correct answer in the form of a number, a noun, a short phrase, or a brief piece of text for the user's questions. Analysis of the question, searching the relevant document, and choosing an answer are three important steps in a QA system. This work uses a web scraper (Beautiful Soup) to extract K-documents from the web. The value of K can be calibrated on the basis of a trade-off between time and accuracy. This is followed by a passage ranking process using the MS-Marco dataset trained on 500K queries to extract the most relevant text passage, to shorten the lengthy documents. Further, a QA system is used to extract the answers from the shortened documents based on the query and return the top 3 answers. For evaluation of such systems, accuracy is judged by the exact match between predicted answers and gold answers. But automatic evaluation methods fail due to the linguistic ambiguities inherent in the questions. Moreover, reference answers are often not exhaustive or are out of date. Hence correct answers predicted by the system are often judged incorrect according to the automated metrics. One such scenario arises from the original Google Natural Question (GNQ) dataset which was collected and made available in the year 2016. Use of any such dataset proves to be inefficient with respect to any questions that have time-varying answers. For illustration, if the query is where will be the next Olympics? Gold Answer for the above query as given in the GNQ dataset is “Tokyo”. Since the dataset was collected in the year 2016, and the next Olympics after 2016 were in 2020 that was in Tokyo which is absolutely correct. But if the same question is asked in 2022 then the answer is “Paris, 2024”. Consequently, any evaluation based on the GNQ dataset will be incorrect. Such erroneous predictions are usually given to human evaluators for further validation which is quite expensive and time-consuming. To address this erroneous evaluation, the present work proposes an automated approach for evaluating time-dependent question-answer pairs. In particular, it proposes a metric using the current timestamp along with top-n predicted answers from a given QA system. To test the proposed approach GNQ dataset has been used and the system achieved an accuracy of 78% for a test dataset comprising 100 QA pairs. This test data was automatically extracted using an analysis-based approach from 10K QA pairs of the GNQ dataset. The results obtained are encouraging. The proposed technique appears to have the possibility of developing into a useful scheme for gathering precise, reliable, and specific information in a real-time and efficient manner. Our subsequent experiments will be guided towards establishing the efficacy of the above system for a larger set of time-dependent QA pairs.

Keywords: web-based information retrieval, open domain question answering system, time-varying QA, QA evaluation

Procedia PDF Downloads 88
11108 Consent and the Construction of Unlawfulness

Authors: Susanna Menis

Abstract:

The context of this study revolves around the theme of consent and the construction of unlawfulness in judicial decisions. It aims to explore the formation of societal perceptions of unlawfulness within the context of consensual sexual acts leading to harmful consequences. This study investigates how judges create legal rules that reflect social solidarity and protect against violence. Specifically, the research aims to understand the justification behind criminalising consensual sexual activity when categorised under different offences. The main question addressed in this study will evaluate the way judges create legal rules that they believe reflect social solidarity and protect against violence. The study employs a historical genealogy approach as its methodology. This approach allows for tracing back the original formation of societal perspectives on unlawfulness, thus highlighting the socially constructed nature of the present understanding. The data for this study will be collected through an extensive literature review, examining historical legal cases and documents that shape the understanding of unlawfulness. This will provide a comprehensive view of how social attitudes toward private sexual relations influenced the creation of legal rules. The theoretical importance of this research lies in its contribution to socio-legal scholarship. This study adds to the existing knowledge on the topic by exploring questions of unconscious bias and its origins. The findings shed light on how and why individuals possess unconscious biases, particularly within the judicial system. In conclusion, this study investigates judicial decisions concerning consensual sexual acts and the construction of unlawfulness. By employing a historical genealogy approach, the research sheds light on how judges create legal rules that reflect social solidarity and aim to protect against violence. The theoretical importance of this study lies in its contribution to understanding unconscious bias and its origins within the judicial system. Through data collection and analysis procedures, this study aims to provide valuable insights into the formation of social attitudes towards private sexual relations and its impact on legal rulings.

Keywords: consent, sexual offences, offences against the person, legal genealogy, social construct

Procedia PDF Downloads 44
11107 Decision Framework for Cross-Border Railway Infrastructure Projects

Authors: Dimitrios J. Dimitriou, Maria F. Sartzetaki

Abstract:

Transport infrastructure assets are key components of the national asset portfolio. The decision to invest in a new infrastructure in transports could take from a few years to some decades. This is mainly because of the need to reserve and spent many capitals, the long payback period, the number of the stakeholders involved in decision process and –many times- the investment and business risks are high. Therefore, the decision assessment framework is an essential challenge linked with the key decision factors meet the stakeholder expectations highlighting project trade-offs, financial risks, business uncertainties and market limitations. This paper examines the decision process for new transport infrastructure projects in cross border regions, where a wide range of stakeholders with different expectation is involved. According to a consequences analysis systemic approach, the relationship of transport infrastructure development, economic system development and stakeholder expectation is analyzed. Adopting the on system of system methodological approach, the decision making framework, variables, inputs and outputs are defined, highlighting the key shareholder’s role and expectations. The application provides the methodology outputs presenting the proposed decision framework for a strategic railway project in north Greece deals with the upgrade of the existing railway corridor connecting Greece, Turkey and Bulgaria.

Keywords: decision making, system of system, cross-border, infrastructure project

Procedia PDF Downloads 297
11106 From Type-I to Type-II Fuzzy System Modeling for Diagnosis of Hepatitis

Authors: Shahabeddin Sotudian, M. H. Fazel Zarandi, I. B. Turksen

Abstract:

Hepatitis is one of the most common and dangerous diseases that affects humankind, and exposes millions of people to serious health risks every year. Diagnosis of Hepatitis has always been a challenge for physicians. This paper presents an effective method for diagnosis of hepatitis based on interval Type-II fuzzy. This proposed system includes three steps: pre-processing (feature selection), Type-I and Type-II fuzzy classification, and system evaluation. KNN-FD feature selection is used as the preprocessing step in order to exclude irrelevant features and to improve classification performance and efficiency in generating the classification model. In the fuzzy classification step, an “indirect approach” is used for fuzzy system modeling by implementing the exponential compactness and separation index for determining the number of rules in the fuzzy clustering approach. Therefore, we first proposed a Type-I fuzzy system that had an accuracy of approximately 90.9%. In the proposed system, the process of diagnosis faces vagueness and uncertainty in the final decision. Thus, the imprecise knowledge was managed by using interval Type-II fuzzy logic. The results that were obtained show that interval Type-II fuzzy has the ability to diagnose hepatitis with an average accuracy of 93.94%. The classification accuracy obtained is the highest one reached thus far. The aforementioned rate of accuracy demonstrates that the Type-II fuzzy system has a better performance in comparison to Type-I and indicates a higher capability of Type-II fuzzy system for modeling uncertainty.

Keywords: hepatitis disease, medical diagnosis, type-I fuzzy logic, type-II fuzzy logic, feature selection

Procedia PDF Downloads 293