Search results for: hierarchical analytical process
16413 Social Business Process Management and Business Process Management Maturity
Authors: Dalia Suša Vugec, Vesna Bosilj Vukšić, Ljubica Milanović Glavan
Abstract:
Business process management (BPM) is a well-known holistic discipline focused on managing business processes with the intention of achieving higher level of BPM maturity and better organizational performance. In recent period, traditional BPM faced some of its limitations like model-reality divide and lost innovation. Following latest trends, as an attempt to overcome the issues of traditional BPM, there has been an introduction of applying the principles of social software in managing business processes which led to the development of social BPM. However, there are not many authors or studies dealing with this topic so this study aims to contribute to that literature gap and to examine the link between the level of BPM maturity and the usage of social BPM. To meet these objectives, a survey within the companies with more than 50 employees has been conducted. The results reveal that the usage of social BPM is higher within the companies which achieved higher level of BPM maturity. This paper provides an overview, analysis and discussion of collected data regarding BPM maturity and social BPM within the observed companies and identifies the main social BPM principles.Keywords: business process management, BPM maturity, process performance index, social BPM
Procedia PDF Downloads 32516412 Description of Decision Inconsistency in Intertemporal Choices and Representation of Impatience as a Reflection of Irrationality: Consequences in the Field of Personalized Behavioral Finance
Authors: Roberta Martino, Viviana Ventre
Abstract:
Empirical evidence has, over time, confirmed that the behavior of individuals is inconsistent with the descriptions provided by the Discounted Utility Model, an essential reference for calculating the utility of intertemporal prospects. The model assumes that individuals calculate the utility of intertemporal prospectuses by adding up the values of all outcomes obtained by multiplying the cardinal utility of the outcome by the discount function estimated at the time the outcome is received. The trend of the discount function is crucial for the preferences of the decision maker because it represents the perception of the future, and its trend causes temporally consistent or temporally inconsistent preferences. In particular, because different formulations of the discount function lead to various conclusions in predicting choice, the descriptive ability of models with a hyperbolic trend is greater than linear or exponential models. Suboptimal choices from any time point of view are the consequence of this mechanism, the psychological factors of which are encapsulated in the discount rate trend. In addition, analyzing the decision-making process from a psychological perspective, there is an equivalence between the selection of dominated prospects and a degree of impatience that decreases over time. The first part of the paper describes and investigates the anomalies of the discounted utility model by relating the cognitive distortions of the decision-maker to the emotional factors that are generated during the evaluation and selection of alternatives. Specifically, by studying the degree to which impatience decreases, it’s possible to quantify how the psychological and emotional mechanisms of the decision-maker result in a lack of decision persistence. In addition, this description presents inconsistency as the consequence of an inconsistent attitude towards time-delayed choices. The second part of the paper presents an experimental phase in which we show the relationship between inconsistency and impatience in different contexts. Analysis of the degree to which impatience decreases confirms the influence of the decision maker's emotional impulses for each anomaly in the utility model discussed in the first part of the paper. This work provides an application in the field of personalized behavioral finance. Indeed, the numerous behavioral diversities, evident even in the degrees of decrease in impatience in the experimental phase, support the idea that optimal strategies may not satisfy individuals in the same way. With the aim of homogenizing the categories of investors and to provide a personalized approach to advice, the results proven in the experimental phase are used in a complementary way with the information in the field of behavioral finance to implement the Analytical Hierarchy Process model in intertemporal choices, useful for strategic personalization. In the construction of the Analytic Hierarchy Process, the degree of decrease in impatience is understood as reflecting irrationality in decision-making and is therefore used for the construction of weights between anomalies and behavioral traits.Keywords: analytic hierarchy process, behavioral finance, financial anomalies, impatience, time inconsistency
Procedia PDF Downloads 6816411 An Experimental Analysis of Squeeze Casting Parameters for 2017 a Wrought Al Alloy
Authors: Mohamed Ben Amar, Najib Souissi, Chedly Bradai
Abstract:
A Taguchi design investigation has been made into the relationship between the ductility and process variables in a squeeze cast 2017A wrought aluminium alloy. The considered process parameters were: squeeze pressure, melt temperature and die preheating temperature. An orthogonal array (OA), main effect, signal-to-noise (S/N) ratio, and the analysis of variance (ANOVA) are employed to analyze the effect of casting parameters. The results have shown that the selected parameters significantly affect the ductility of 2017A wrought Al alloy castings. Optimal squeeze cast process parameters were provided to illustrate the proposed approach and the results were proven to be trustworthy through practical experiments.Keywords: Taguchi method, squeeze casting, process parameters, ductility, microstructure
Procedia PDF Downloads 40016410 Discrete Estimation of Spectral Density for Alpha Stable Signals Observed with an Additive Error
Authors: R. Sabre, W. Horrigue, J. C. Simon
Abstract:
This paper is interested in two difficulties encountered in practice when observing a continuous time process. The first is that we cannot observe a process over a time interval; we only take discrete observations. The second is the process frequently observed with a constant additive error. It is important to give an estimator of the spectral density of such a process taking into account the additive observation error and the choice of the discrete observation times. In this work, we propose an estimator based on the spectral smoothing of the periodogram by the polynomial Jackson kernel reducing the additive error. In order to solve the aliasing phenomenon, this estimator is constructed from observations taken at well-chosen times so as to reduce the estimator to the field where the spectral density is not zero. We show that the proposed estimator is asymptotically unbiased and consistent. Thus we obtain an estimate solving the two difficulties concerning the choice of the instants of observations of a continuous time process and the observations affected by a constant error.Keywords: spectral density, stable processes, aliasing, periodogram
Procedia PDF Downloads 13816409 Parametrical Simulation of Sheet Metal Forming Process to Control the Localized Thinning
Authors: Hatem Mrad, Alban Notin, Mohamed Bouazara
Abstract:
Sheet metal forming process has a multiple successive steps starting from sheets fixation to sheets evacuation. Often after forming operation, the sheet has defects requiring additional corrections steps. For example, in the drawing process, the formed sheet may have several defects such as springback, localized thinning and bends. All these defects are directly dependent on process, geometric and material parameters. The prediction and elimination of these defects requires the control of most sensitive parameters. The present study is concerned with a reliable parametric study of deep forming process in order to control the localized thinning. The proposed approach will be based on stochastic finite element method. Especially, the polynomial Chaos development will be used to establish a reliable relationship between input (process, geometric and material parameters) and output variables (sheet thickness). The commercial software Abaqus is used to conduct numerical finite elements simulations. The automatized parametrical modification is provided by coupling a FORTRAN routine, a PYTHON script and input Abaqus files.Keywords: sheet metal forming, reliability, localized thinning, parametric simulation
Procedia PDF Downloads 42316408 Optimization of Surface Roughness in Turning Process Utilizing Live Tooling via Taguchi Methodology
Authors: Weinian Wang, Joseph C. Chen
Abstract:
The objective of this research is to optimize the process of cutting cylindrical workpieces utilizing live tooling on a HAAS ST-20 lathe. Surface roughness (Ra) has been investigated as the indicator of quality characteristics for machining process. Aluminum alloy was used to conduct experiments due to its wide range usages in engineering structures and components where light weight or corrosion resistance is required. In this study, Taguchi methodology is utilized to determine the effects that each of the parameters has on surface roughness (Ra). A total of 18 experiments of each process were designed according to Taguchi’s L9 orthogonal array (OA) with four control factors at three levels of each and signal-to-noise ratios (S/N) were computed with Smaller the better equation for minimizing the system. The optimal parameters identified for the surface roughness of the turning operation utilizing live tooling were a feed rate of 3 inches/min(A3); a spindle speed of 1300 rpm(B3); a 2-flute titanium nitrite coated 3/8” endmill (C1); and a depth of cut of 0.025 inches (D2). The mean surface roughness of the confirmation runs in turning operation was 8.22 micro inches. The final results demonstrate that Taguchi methodology is a sufficient way of process improvement in turning process on surface roughness.Keywords: CNC milling operation, CNC turning operation, surface roughness, Taguchi parameter design
Procedia PDF Downloads 17716407 Analysis of Socio-Economics of Tuna Fisheries Management (Thunnus Albacares Marcellus Decapterus) in Makassar Waters Strait and Its Effect on Human Health and Policy Implications in Central Sulawesi-Indonesia
Authors: Siti Rahmawati
Abstract:
Indonesia has had long period of monetary economic crisis and it is followed by an upward trend in the price of fuel oil. This situation impacts all aspects of tuna fishermen community. For instance, the basic needs of fishing communities increase and the lower purchasing power then lead to economic and social instability as well as the health of fishermen household. To understand this AHP method is applied to acknowledge the model of tuna fisheries management priorities and cold chain marketing channel and the utilization levels that impact on human health. The study is designed as a development research with the number of 180 respondents. The data were analyzed by Analytical Hierarchy Process (AHP) method. The development of tuna fishery business can improve productivity of production with economic empowerment activities for coastal communities, improving the competitiveness of products, developing fish processing centers and provide internal capital for the development of optimal fishery business. From economic aspects, fishery business is more attracting because the benefit cost ratio of 2.86. This means that for 10 years, the economic life of this project can work well as B/C> 1 and therefore the rate of investment is economically viable. From the health aspects, tuna can reduce the risk of dying from heart disease by 50%, because tuna contain selenium in the human body. The consumption of 100 g of tuna meet 52.9% of the selenium in the body and activating the antioxidant enzyme glutathione peroxidaxe which can protect the body from free radicals and stimulate various cancers. The results of the analytic hierarchy process that the quality of tuna products is the top priority for export quality as well as quality control in order to compete in the global market. The implementation of the policy can increase the income of fishermen and reduce the poverty of fishermen households and have impact on the human health whose has high risk of disease.Keywords: management of tuna, social, economic, health
Procedia PDF Downloads 31716406 An Efficient Hardware/Software Workflow for Multi-Cores Simulink Applications
Authors: Asma Rebaya, Kaouther Gasmi, Imen Amari, Salem Hasnaoui
Abstract:
Over these last years, applications such as telecommunications, signal processing, digital communication with advanced features (Multi-antenna, equalization..) witness a rapid evaluation accompanied with an increase of user exigencies in terms of latency, the power of computation… To satisfy these requirements, the use of hardware/software systems is a common solution; where hardware is composed of multi-cores and software is represented by models of computation, synchronous data flow (SDF) graph for instance. Otherwise, the most of the embedded system designers utilize Simulink for modeling. The issue is how to simplify the c code generation, for a multi-cores platform, of an application modeled by Simulink. To overcome this problem, we propose a workflow allowing an automatic transformation from the Simulink model to the SDF graph and providing an efficient schedule permitting to optimize the number of cores and to minimize latency. This workflow goes from a Simulink application and a hardware architecture described by IP.XACT language. Based on the synchronous and hierarchical behavior of both models, the Simulink block diagram is automatically transformed into an SDF graph. Once this process is successfully achieved, the scheduler calculates the optimal cores’ number needful by minimizing the maximum density of the whole application. Then, a core is chosen to execute a specific graph task in a specific order and, subsequently, a compatible C code is generated. In order to perform this proposal, we extend Preesm, a rapid prototyping tool, to take the Simulink model as entry input and to support the optimal schedule. Afterward, we compared our results to this tool results, using a simple illustrative application. The comparison shows that our results strictly dominate the Preesm results in terms of number of cores and latency. In fact, if Preesm needs m processors and latency L, our workflow need processors and latency L'< L.Keywords: hardware/software system, latency, modeling, multi-cores platform, scheduler, SDF graph, Simulink model, workflow
Procedia PDF Downloads 27016405 Evaluation of Diagnosis Performance Based on Pairwise Model Construction and Filtered Data
Authors: Hyun-Woo Cho
Abstract:
It is quite important to utilize right time and intelligent production monitoring and diagnosis of industrial processes in terms of quality and safety issues. When compared with monitoring task, fault diagnosis represents the task of finding process variables responsible causing a specific fault in the process. It can be helpful to process operators who should investigate and eliminate root causes more effectively and efficiently. This work focused on the active use of combining a nonlinear statistical technique with a preprocessing method in order to implement practical real-time fault identification schemes for data-rich cases. To compare its performance to existing identification schemes, a case study on a benchmark process was performed in several scenarios. The results showed that the proposed fault identification scheme produced more reliable diagnosis results than linear methods. In addition, the use of the filtering step improved the identification results for the complicated processes with massive data sets.Keywords: diagnosis, filtering, nonlinear statistical techniques, process monitoring
Procedia PDF Downloads 24516404 Modeling the Demand for the Healthcare Services Using Data Analysis Techniques
Authors: Elizaveta S. Prokofyeva, Svetlana V. Maltseva, Roman D. Zaitsev
Abstract:
Rapidly evolving modern data analysis technologies in healthcare play a large role in understanding the operation of the system and its characteristics. Nowadays, one of the key tasks in urban healthcare is to optimize the resource allocation. Thus, the application of data analysis in medical institutions to solve optimization problems determines the significance of this study. The purpose of this research was to establish the dependence between the indicators of the effectiveness of the medical institution and its resources. Hospital discharges by diagnosis; hospital days of in-patients and in-patient average length of stay were selected as the performance indicators and the demand of the medical facility. The hospital beds by type of care, medical technology (magnetic resonance tomography, gamma cameras, angiographic complexes and lithotripters) and physicians characterized the resource provision of medical institutions for the developed models. The data source for the research was an open database of the statistical service Eurostat. The choice of the source is due to the fact that the databases contain complete and open information necessary for research tasks in the field of public health. In addition, the statistical database has a user-friendly interface that allows you to quickly build analytical reports. The study provides information on 28 European for the period from 2007 to 2016. For all countries included in the study, with the most accurate and complete data for the period under review, predictive models were developed based on historical panel data. An attempt to improve the quality and the interpretation of the models was made by cluster analysis of the investigated set of countries. The main idea was to assess the similarity of the joint behavior of the variables throughout the time period under consideration to identify groups of similar countries and to construct the separate regression models for them. Therefore, the original time series were used as the objects of clustering. The hierarchical agglomerate algorithm k-medoids was used. The sampled objects were used as the centers of the clusters obtained, since determining the centroid when working with time series involves additional difficulties. The number of clusters used the silhouette coefficient. After the cluster analysis it was possible to significantly improve the predictive power of the models: for example, in the one of the clusters, MAPE error was only 0,82%, which makes it possible to conclude that this forecast is highly reliable in the short term. The obtained predicted values of the developed models have a relatively low level of error and can be used to make decisions on the resource provision of the hospital by medical personnel. The research displays the strong dependencies between the demand for the medical services and the modern medical equipment variable, which highlights the importance of the technological component for the successful development of the medical facility. Currently, data analysis has a huge potential, which allows to significantly improving health services. Medical institutions that are the first to introduce these technologies will certainly have a competitive advantage.Keywords: data analysis, demand modeling, healthcare, medical facilities
Procedia PDF Downloads 14516403 Learning Materials of Atmospheric Pressure Plasma Process: Application in Wrinkle-Resistant Finishing of Cotton Fabric
Authors: C. W. Kan
Abstract:
Cotton fibre is a commonly-used natural fibre because of its good fibre strength, high moisture absorption behaviour and minimal static problems. However, one of the main drawbacks of cotton fibre is wrinkling after washing, which is recently overcome by wrinkle-resistant treatment. 1,2,3,4-butanetetracarboxylic acid (BTCA) could improve the wrinkle-resistant properties of cotton fibre. Although the BTCA process is an effective method for wrinkle resistant application of cotton fabrics, reduced fabric strength was observed after treatment. Therefore, this paper would explore the use of atmospheric pressure plasma treatment under different discharge powers as a pretreatment process to enhance the application of BTCA process on cotton fabric without generating adverse effect. The aim of this study is to provide learning information to the users to know how the atmospheric pressure plasma treatment can be incorporated in textile finishing process with positive impact.Keywords: learning materials, atmospheric pressure plasma treatment, cotton, wrinkle-resistant, BTCA
Procedia PDF Downloads 30616402 The Problem of the Use of Learning Analytics in Distance Higher Education: An Analytical Study of the Open and Distance University System in Mexico
Authors: Ismene Ithai Bras-Ruiz
Abstract:
Learning Analytics (LA) is employed by universities not only as a tool but as a specialized ground to enhance students and professors. However, not all the academic programs apply LA with the same goal and use the same tools. In fact, LA is formed by five main fields of study (academic analytics, action research, educational data mining, recommender systems, and personalized systems). These fields can help not just to inform academic authorities about the situation of the program, but also can detect risk students, professors with needs, or general problems. The highest level applies Artificial Intelligence techniques to support learning practices. LA has adopted different techniques: statistics, ethnography, data visualization, machine learning, natural language process, and data mining. Is expected that any academic program decided what field wants to utilize on the basis of his academic interest but also his capacities related to professors, administrators, systems, logistics, data analyst, and the academic goals. The Open and Distance University System (SUAYED in Spanish) of the University National Autonomous of Mexico (UNAM), has been working for forty years as an alternative to traditional programs; one of their main supports has been the employ of new information and communications technologies (ICT). Today, UNAM has one of the largest network higher education programs, twenty-six academic programs in different faculties. This situation means that every faculty works with heterogeneous populations and academic problems. In this sense, every program has developed its own Learning Analytic techniques to improve academic issues. In this context, an investigation was carried out to know the situation of the application of LA in all the academic programs in the different faculties. The premise of the study it was that not all the faculties have utilized advanced LA techniques and it is probable that they do not know what field of study is closer to their program goals. In consequence, not all the programs know about LA but, this does not mean they do not work with LA in a veiled or, less clear sense. It is very important to know the grade of knowledge about LA for two reasons: 1) This allows to appreciate the work of the administration to improve the quality of the teaching and, 2) if it is possible to improve others LA techniques. For this purpose, it was designed three instruments to determinate the experience and knowledge in LA. These were applied to ten faculty coordinators and his personnel; thirty members were consulted (academic secretary, systems manager, or data analyst, and coordinator of the program). The final report allowed to understand that almost all the programs work with basic statistics tools and techniques, this helps the administration only to know what is happening inside de academic program, but they are not ready to move up to the next level, this means applying Artificial Intelligence or Recommender Systems to reach a personalized learning system. This situation is not related to the knowledge of LA, but the clarity of the long-term goals.Keywords: academic improvements, analytical techniques, learning analytics, personnel expertise
Procedia PDF Downloads 12816401 Trend Detection Using Community Rank and Hawkes Process
Authors: Shashank Bhatnagar, W. Wilfred Godfrey
Abstract:
We develop in this paper, an approach to find the trendy topic, which not only considers the user-topic interaction but also considers the community, in which user belongs. This method modifies the previous approach of user-topic interaction to user-community-topic interaction with better speed-up in the range of [1.1-3]. We assume that trend detection in a social network is dependent on two things. The one is, broadcast of messages in social network governed by self-exciting point process, namely called Hawkes process and the second is, Community Rank. The influencer node links to others in the community and decides the community rank based on its PageRank and the number of users links to that community. The community rank decides the influence of one community over the other. Hence, the Hawkes process with the kernel of user-community-topic decides the trendy topic disseminated into the social network.Keywords: community detection, community rank, Hawkes process, influencer node, pagerank, trend detection
Procedia PDF Downloads 38516400 A CPS Based Design of Industrial Ecosystems
Authors: Maryam Shayan
Abstract:
Chemical Process Simulation (CPS) software has been generally utilized by chemical (process) designers to outline, test, advance, and coordinate process plants. It is relied upon that modern scientists to bring these same critical thinking advantages to the outline and operation of industrial ecosystems can utilize CPS. This paper gives modern environment researchers and experts with a prologue to CPS and a review of compound designing configuration standards. The paper highlights late research demonstrating that CPS can be utilized to model modern industrial ecosystems, and talks about the advantages of utilizing CPS to address a portion of the specialized difficulties confronting organizations partaking in an industrial ecosystem. CPS can be utilized to (i) quantitatively assess and analyze the potential ecological and monetary advantages of material and vitality linkages; (ii) unravel general plan, retrofit, or operational issues; (iii) help to distinguish complex and frequently irrational arrangements; and (iv) assess imagine a scenario in which situations. CPS ought to be a valuable expansion to the mechanical environment tool stash.Keywords: chemical process simulation (CPS), process plants, industrial ecosystems, compound designing
Procedia PDF Downloads 28116399 Effect of Injection Moulding Process Parameter on Tensile Strength of Using Taguchi Method
Authors: Gurjeet Singh, M. K. Pradhan, Ajay Verma
Abstract:
The plastic industry plays very important role in the economy of any country. It is generally among the leading share of the economy of the country. Since metals and their alloys are very rarely available on the earth. So to produce plastic products and components, which finds application in many industrial as well as household consumer products is beneficial. Since 50% plastic products are manufactured by injection moulding process. For production of better quality product, we have to control quality characteristics and performance of the product. The process parameters plays a significant role in production of plastic, hence the control of process parameter is essential. In this paper the effect of the parameters selection on injection moulding process has been described. It is to define suitable parameters in producing plastic product. Selecting the process parameter by trial and error is neither desirable nor acceptable, as it is often tends to increase the cost and time. Hence optimization of processing parameter of injection moulding process is essential. The experiments were designed with Taguchi’s orthogonal array to achieve the result with least number of experiments. Here Plastic material polypropylene is studied. Tensile strength test of material is done on universal testing machine, which is produced by injection moulding machine. By using Taguchi technique with the help of MiniTab-14 software the best value of injection pressure, melt temperature, packing pressure and packing time is obtained. We found that process parameter packing pressure contribute more in production of good tensile plastic product.Keywords: injection moulding, tensile strength, poly-propylene, Taguchi
Procedia PDF Downloads 28816398 Assessing the Adaptive Re-Use Potential of Buildings as Part of the Disaster Management Process
Authors: A. Esra İdemen, Sinan M. Şener, Emrah Acar
Abstract:
The technological paradigm of the disaster management field, especially in the case of governmental intervention strategies, is generally based on rapid and flexible accommodation solutions. From various technical solution patterns used to address the immediate housing needs of disaster victims, the adaptive re-use of existing buildings can be considered to be both low-cost and practical. However, there is a scarcity of analytical methods to screen, select and adapt buildings to help decision makers in cases of emergency. Following an extensive literature review, this paper aims to highlight key points and problem areas associated with the adaptive re-use of buildings within the disaster management context. In other disciplines such as real estate management, the adaptive re-use potential (ARP) of existing buildings is typically based on the prioritization of a set of technical and non-technical criteria which are then weighted to arrive at an economically viable investment decision. After a disaster, however, the assessment of the ARP of buildings requires consideration of different/additional layers of analysis which stem from general disaster management principles and the peculiarities of different types of disasters, as well as of their victims. In this paper, a discussion of the development of an adaptive re-use potential (ARP) assessment model is presented. It is thought that governmental and non-governmental decision makers who are required to take quick decisions to accommodate displaced masses following disasters are likely to benefit from the implementation of such a model.Keywords: adaptive re-use of buildings, disaster management, temporary housing, assessment model
Procedia PDF Downloads 33216397 Bayesian Prospective Detection of Small Area Health Anomalies Using Kullback Leibler Divergence
Authors: Chawarat Rotejanaprasert, Andrew Lawson
Abstract:
Early detection of unusual health events depends on the ability to detect rapidly any substantial changes in disease, thus facilitating timely public health interventions. To assist public health practitioners to make decisions, statistical methods are adopted to assess unusual events in real time. We introduce a surveillance Kullback-Leibler (SKL) measure for timely detection of disease outbreaks for small area health data. The detection methods are compared with the surveillance conditional predictive ordinate (SCPO) within the framework of Bayesian hierarchical Poisson modeling and applied to a case study of a group of respiratory system diseases observed weekly in South Carolina counties. Properties of the proposed surveillance techniques including timeliness and detection precision are investigated using a simulation study.Keywords: Bayesian, spatial, temporal, surveillance, prospective
Procedia PDF Downloads 31216396 Community-Based Assessment Approach to Empower Child with Disabilities: Institutional Study on Deaf Art Community in Yogyakarta, Indonesia
Authors: Mukhamad Fatkhullah, Arfan Fadli, Marini Kristina Situmeang, Siti Hazar Sitorus
Abstract:
The emergence of a community of people with disabilities along with the various works produced has made great progress to open the public eye to their existence in society. This study focuses attention on a community that is suspected to be one of the pioneers in pursuing the movement. It is Deaf Art Community (DAC), a community of persons with disabilities based in Yogyakarta, with deaf and speech-impaired members who use sign language in everyday communication. Knowing the movement of disabled communities is a good thing, the description of the things behind it then important to know as the basis for initiating similar movements. This research focuses on the question of how community of people with disabilities begin to take shape in different regions and interact with collaborative events. Qualitative method with in-depth interview as data collection techniques was used to describe the process of formation and the emergence of community. The analytical unit in the study initially focuses on the subject in the community, but in the process, it develops to institutional analysis. Therefore some informants were determined purposively and expanded using the snowball technique. The theory used in this research is Phenomenology of Alfred Schutz to be able to see reality from the subject and institutional point of view. The results of this study found that the community is formed because the existing educational institutions (both SLB and inclusion) are less able to empower and make children with disabilities become equal with the society. Through the SLB, the presence of children with disabilities becomes isolated from the society, especially in children of his or her age. Therefore, discrimination and labeling will never be separated from society's view. Meanwhile, facilities for the basic needs of children with disabilities can not be fully provided. Besides that, the guarantee of discrimination, glances, and unpleasant behavior from children without disability does not exist, which then indicates that the existing inclusion schools offer only symbolic acceptance. Thus, both in SLB and Inclusive Schools can not empower children with disabilities. Community-based assistance, in this case, has become an alternative to actually empowering children with disabilities. Not only giving them a place to interact, through the same community, children with disabilities will be guided to discover their talents and develop their potential to be self-reliant in the future.Keywords: children with disabilities, community-based assessment, community empowerment, social equity
Procedia PDF Downloads 26516395 Using Nonhomogeneous Poisson Process with Compound Distribution to Price Catastrophe Options
Authors: Rong-Tsorng Wang
Abstract:
In this paper, we derive a pricing formula for catastrophe equity put options (or CatEPut) with non-homogeneous loss and approximated compound distributions. We assume that the loss claims arrival process is a nonhomogeneous Poisson process (NHPP) representing the clustering occurrences of loss claims, the size of loss claims is a sequence of independent and identically distributed random variables, and the accumulated loss distribution forms a compound distribution and is approximated by a heavy-tailed distribution. A numerical example is given to calibrate parameters, and we discuss how the value of CatEPut is affected by the changes of parameters in the pricing model we provided.Keywords: catastrophe equity put options, compound distributions, nonhomogeneous Poisson process, pricing model
Procedia PDF Downloads 16716394 The Effect of Online Learning During the COVID-19 Pandemic on Student Mental
Authors: Adelia Desi Agnesita
Abstract:
The advent of a new disease called covid-19 made many major changes in the world, one of which is the process of learning and teaching. Learning formerly offline but now is done online, which makes students need adaptation to the learning process. The covid-19 pandemic that occurs almost worldwide causes activities that involve many people to be avoided, one of which is learning to teach. In Indonesia, since March 2020, the process of college learning is turning into online/ long-distance learning. It's to prevent the spread of the covid-19. Student online learning presents some of the obstacles to poor signals, many of the tasks, lack of focus, difficulty sleeping, and resulting stress.Keywords: learning, online, covid-19, pandemic
Procedia PDF Downloads 21716393 Experimental Investigations on the Mechanism of Stratified Liquid Mixing in a Cylinder
Authors: Chai Mingming, Li Lei, Lu Xiaoxia
Abstract:
In this paper, the mechanism of stratified liquids’ mixing in a cylinder is investigated. It is focused on the effects of Rayleigh-Taylor Instability (RTI) and rotation of the cylinder on liquid interface mixing. For miscible liquids, Planar Laser Induced Fluorescence (PLIF) technique is applied to record the concentration field for one liquid. Intensity of Segregation (IOS) is used to describe the mixing status. For immiscible liquids, High Speed Camera is adopted to record the development of the interface. The experiment of RTI indicates that it plays a great role in the mixing process, and meanwhile the large-scale mixing is triggered, and subsequently the span of the stripes decreases, showing that the mesoscale mixing is coming into being. The rotation experiments show that the spin-down process has a great role in liquid mixing, during which the upper liquid falls down rapidly along the wall and crashes into the lower liquid. During this process, a lot of interface instabilities are excited. Liquids mix rapidly in the spin-down process. It can be concluded that no matter what ways have been adopted to speed up liquid mixing, the fundamental reason is the interface instabilities which increase the area of the interface between liquids and increase the relative velocity of the two liquids.Keywords: interface instability, liquid mixing, Rayleigh-Taylor Instability, spin-down process, spin-up process
Procedia PDF Downloads 30116392 Aerogel Fabrication Via Modified Rapid Supercritical Extraction (RSCE) Process - Needle Valve Pressure Release
Authors: Haibo Zhao, Thomas Andre, Katherine Avery, Alper Kiziltas, Deborah Mielewski
Abstract:
Silica aerogels were fabricated through a modified rapid supercritical extraction (RSCE) process. The silica aerogels were made using a tetramethyl orthosilicate precursor and then placed in a hot press and brought to the supercritical point of the solvent, ethanol. In order to control the pressure release without a pressure controller, a needle valve was used. The resulting aerogels were then characterized for their physical and chemical properties and compared to silica aerogels created using similar methods. The aerogels fabricated using this modified RSCE method were found to have similar properties to those in other papers using the unmodified RSCE method. Silica aerogel infused glass blanket composite, graphene reinforced silica aerogel composite were also successfully fabricated by this new method. The modified RSCE process and system is a prototype for better gas outflow control with a lower cost of equipment setup. Potentially, this process could be evolved to a continuous low-cost high-volume production process to meet automotive requirements.Keywords: aerogel, automotive, rapid supercritical extraction process, low cost production
Procedia PDF Downloads 18416391 Positioning of Lesbian and Gay Workers within the Corporate Sector in Sri Lanka: The Case of Residents in the Colombo District
Authors: Pramoda Karunarathna, Hemamalie Gunatilaka
Abstract:
This study is based on experiences of Sri Lankan lesbian and gay workers’ career in the corporate sector, which include both manufacturing and service sectors. The study has started with the intention of shedding light on a grey area to observe the negative effects on lesbian and gay workers and their experiences while they are employed in the Sri Lankan corporate sector. In order to understand the experiences of lesbian and gay workers while they are at work within the corporate sector, the study seeks to address four questions. First research question is about the challenges faced by lesbian and gay workers while they are at work, and the second research question looks at their career patterns. Third research question seeks to address the behavior at work, and the fourth research question looks at the influence of class, religion, and cultural aspects on the career of lesbian and gay workers. Methodologically, the research was based on semi-structured interviews with nine participants (five gay men and four lesbian women) having work experience in the corporate sector and residing in Colombo, the capital city of Sri Lanka. The research found that the participants have gone through the process of developing sexual identity; gay men possess more feminine characteristics, while lesbian women possess more masculine characteristics. Further, their identity gets revealed in different ways, such as through the curriculum vitae, at the interviews, through the attire and behavior, and with the use of social media. The study also found that lesbian and gay workers experience discrimination due to violation of hierarchical power difference by other employees and marginalization, verbal and nonverbal abuse by other men at work are common experiences. Another finding is that lesbian and gay workers adopt strategies for survival at work, and they prefer the NGO sector to the corporate sector. In contrast, even within the corporate sector, advertising is preferred by lesbian and gay workers. Some of the Sri Lankan corporate sector organizations, especially multinational organizations, have initiated diversity training, and it might lead to making these organisations lesbian and gay-friendly workplaces in the future. It is also found that nearly 44 percent of the participants do not have a religion, and it is due to the rejection of deviant behaviours by most of the religions. In conclusion, lesbian and gay workers experience discrimination at work in the Sri Lankan corporate sector with an exception to the companies relating to advertising and non-governmental organisations is the sector that these workers prefer the most.Keywords: lesbian workers, gay workers, Sri Lankan corporate sector, discrimination
Procedia PDF Downloads 14216390 The Roles of Pay Satisfaction and Intent to Leave on Counterproductive Work Behavior among Non-Academic University Employees
Authors: Abiodun Musbau Lawal, Sunday Samson Babalola, Uzor Friday Ordu
Abstract:
Issue of employees counterproductive work behavior in government owned organization in emerging economies has continued to be a major concern. This study investigated the factors of pay satisfaction, intent to leave and age as predictors of counterproductive work behavior among non-academic employee in a Nigerian federal government owned university. A sample of 200 non-academic employees completed questionnaires. Hierarchical multiple regression was conducted to determine the contribution of each of the predictor variables on the criterion variable on counterproductive work behavior. Results indicate that age of participants (β = -.18; p < .05) significantly independently predicted CWB by accounting for 3% of the explained variance. Addition of pay satisfaction (β = -.14; p < .05) significantly accounted for 5% of the explained variance, while intent to leave (β = -.17; p < .05) further resulted in 8% of the explained variance in counterproductive work behavior. The importance of these findings with regards to reduction in counterproductive work behavior is highlighted.Keywords: counterproductive, work behaviour, pay satisfaction, intent to leave
Procedia PDF Downloads 38516389 A Survey of 2nd Year Students' Frequent Writing Error and the Effects of Participatory Error Correction Process
Authors: Chaiwat Tantarangsee
Abstract:
The purposes of this study are 1) to study the effects of participatory error correction process and 2) to find out the students’ satisfaction of such error correction process. This study is a Quasi Experimental Research with single group, in which data is collected 5 times preceding and following 4 experimental studies of participatory error correction process including providing coded indirect corrective feedback in the students’ texts with error treatment activities. Samples include 28 2nd year English Major students, Faculty of Humanities and Social Sciences, Suan Sunandha Rajabhat University. Tool for experimental study includes the lesson plan of the course; Reading and Writing English for Academic Purposes II, and tools for data collection include 5 writing tests of short texts and a questionnaire. Based on formative evaluation of the students’ writing ability prior to and after each of the 4 experiments, the research findings disclose the students’ higher scores with statistical difference at 0.05. Moreover, in terms of the effect size of such process, it is found that for mean of the students’ scores prior to and after the 4 experiments; d equals 1.0046, 1.1374, 1.297, and 1.0065 respectively. It can be concluded that participatory error correction process enables all of the students to learn equally well and there is improvement in their ability to write short texts. Finally, the students’ overall satisfaction of the participatory error correction process is in high level (Mean=4.32, S.D.=0.92).Keywords: coded indirect corrective feedback, participatory error correction process, error treatment, humanities and social sciences
Procedia PDF Downloads 52316388 Performance Evaluation of Production Schedules Based on Process Mining
Authors: Kwan Hee Han
Abstract:
External environment of enterprise is rapidly changing majorly by global competition, cost reduction pressures, and new technology. In these situations, production scheduling function plays a critical role to meet customer requirements and to attain the goal of operational efficiency. It deals with short-term decision making in the production process of the whole supply chain. The major task of production scheduling is to seek a balance between customer orders and limited resources. In manufacturing companies, this task is so difficult because it should efficiently utilize resource capacity under the careful consideration of many interacting constraints. At present, many computerized software solutions have been utilized in many enterprises to generate a realistic production schedule to overcome the complexity of schedule generation. However, most production scheduling systems do not provide sufficient information about the validity of the generated schedule except limited statistics. Process mining only recently emerged as a sub-discipline of both data mining and business process management. Process mining techniques enable the useful analysis of a wide variety of processes such as process discovery, conformance checking, and bottleneck analysis. In this study, the performance of generated production schedule is evaluated by mining event log data of production scheduling software system by using the process mining techniques since every software system generates event logs for the further use such as security investigation, auditing and error bugging. An application of process mining approach is proposed for the validation of the goodness of production schedule generated by scheduling software systems in this study. By using process mining techniques, major evaluation criteria such as utilization of workstation, existence of bottleneck workstations, critical process route patterns, and work load balance of each machine over time are measured, and finally, the goodness of production schedule is evaluated. By using the proposed process mining approach for evaluating the performance of generated production schedule, the quality of production schedule of manufacturing enterprises can be improved.Keywords: data mining, event log, process mining, production scheduling
Procedia PDF Downloads 28016387 Multinomial Dirichlet Gaussian Process Model for Classification of Multidimensional Data
Authors: Wanhyun Cho, Soonja Kang, Sanggoon Kim, Soonyoung Park
Abstract:
We present probabilistic multinomial Dirichlet classification model for multidimensional data and Gaussian process priors. Here, we have considered an efficient computational method that can be used to obtain the approximate posteriors for latent variables and parameters needed to define the multiclass Gaussian process classification model. We first investigated the process of inducing a posterior distribution for various parameters and latent function by using the variational Bayesian approximations and important sampling method, and next we derived a predictive distribution of latent function needed to classify new samples. The proposed model is applied to classify the synthetic multivariate dataset in order to verify the performance of our model. Experiment result shows that our model is more accurate than the other approximation methods.Keywords: multinomial dirichlet classification model, Gaussian process priors, variational Bayesian approximation, importance sampling, approximate posterior distribution, marginal likelihood evidence
Procedia PDF Downloads 44516386 The Development of Micro Patterns Using Benchtop Lithography for Marine Antifouling Applications
Authors: Felicia Wong Yen Myan, James Walker
Abstract:
Development of micro topographies usually begins with the fabrication of a master stamp. Fabrication of such small structures can be technically challenging and expensive. These techniques are often used for applications where patterns only cover a small surface area (e.g. semiconductors, microfluidic channels). This research investigated the use of benchtop lithography to fabricate patterns with average widths of 50 and 100 microns on silicon wafer substrates. Further development of this method will attempt to layer patterns to create hierarchical structures. Photomasks consisted of patterns printed onto transparency films with a high resolution printer and a fully patterned 10cm by 10cm area has been successfully developed. UV exposure was carried out with a self-made array of ultraviolet LEDs that was positioned a distance above a glass diffuser. Observations under a light microscope and SEM showed that developed patterns exhibit an adequate degree of fidelity with patterns from the master stamp.Keywords: lithography, antifouling, marine, microtopography
Procedia PDF Downloads 29016385 Bioinformatic Approaches in Population Genetics and Phylogenetic Studies
Authors: Masoud Sheidai
Abstract:
Biologists with a special field of population genetics and phylogeny have different research tasks such as populations’ genetic variability and divergence, species relatedness, the evolution of genetic and morphological characters, and identification of DNA SNPs with adaptive potential. To tackle these problems and reach a concise conclusion, they must use the proper and efficient statistical and bioinformatic methods as well as suitable genetic and morphological characteristics. In recent years application of different bioinformatic and statistical methods, which are based on various well-documented assumptions, are the proper analytical tools in the hands of researchers. The species delineation is usually carried out with the use of different clustering methods like K-means clustering based on proper distance measures according to the studied features of organisms. A well-defined species are assumed to be separated from the other taxa by molecular barcodes. The species relationships are studied by using molecular markers, which are analyzed by different analytical methods like multidimensional scaling (MDS) and principal coordinate analysis (PCoA). The species population structuring and genetic divergence are usually investigated by PCoA and PCA methods and a network diagram. These are based on bootstrapping of data. The Association of different genes and DNA sequences to ecological and geographical variables is determined by LFMM (Latent factor mixed model) and redundancy analysis (RDA), which are based on Bayesian and distance methods. Molecular and morphological differentiating characters in the studied species may be identified by linear discriminant analysis (DA) and discriminant analysis of principal components (DAPC). We shall illustrate these methods and related conclusions by giving examples from different edible and medicinal plant species.Keywords: GWAS analysis, K-Means clustering, LFMM, multidimensional scaling, redundancy analysis
Procedia PDF Downloads 12616384 Simulation Study of Asphaltene Deposition and Solubility of CO2 in the Brine during Cyclic CO2 Injection Process in Unconventional Tight Reservoirs
Authors: Rashid S. Mohammad, Shicheng Zhang, Sun Lu, Syed Jamal-Ud-Din, Xinzhe Zhao
Abstract:
A compositional reservoir simulation model (CMG-GEM) was used for cyclic CO2 injection process in unconventional tight reservoir. Cyclic CO2 injection is an enhanced oil recovery process consisting of injection, shut-in, and production. The study of cyclic CO2 injection and hydrocarbon recovery in ultra-low permeability reservoirs is mainly a function of rock, fluid, and operational parameters. CMG-GEM was used to study several design parameters of cyclic CO2 injection process to distinguish the parameters with maximum effect on the oil recovery and to comprehend the behavior of cyclic CO2 injection in tight reservoir. On the other hand, permeability reduction induced by asphaltene precipitation is one of the major issues in the oil industry due to its plugging onto the porous media which reduces the oil productivity. In addition to asphaltene deposition, solubility of CO2 in the aquifer is one of the safest and permanent trapping techniques when considering CO2 storage mechanisms in geological formations. However, the effects of the above uncertain parameters on the process of CO2 enhanced oil recovery have not been understood systematically. Hence, it is absolutely necessary to study the most significant parameters which dominate the process. The main objective of this study is to improve techniques for designing cyclic CO2 injection process while considering the effects of asphaltene deposition and solubility of CO2 in the brine in order to prevent asphaltene precipitation, minimize CO2 emission, optimize cyclic CO2 injection, and maximize oil production.Keywords: tight reservoirs, cyclic O₂ injection, asphaltene, solubility, reservoir simulation
Procedia PDF Downloads 387