Search results for: learner’s cognitive process
13623 Interpretation of Heritage Revitalization
Authors: Jarot Mahendra
Abstract:
The primary objective of this paper is to provide a view in the interpretation of the revitalization of heritage buildings. This objective is achieved by analyzing the concept of interpretation that is oriented in the perspective of law, urban spatial planning, and stakeholder perspective, and then develops the theoretical framework of interpretation in the cultural resources management through issues of identity, heritage as a process, and authenticity in heritage. The revitalization of heritage buildings with the interpretation of these three issues is that interpretation can be used as a communication process to express the meaning and relation of heritage to the community so as to avoid the conflict that will arise and develop as a result of different perspectives of stakeholders. Using case studies in Indonesia, this study focuses on the revitalization of heritage sites in the National Gallery of Indonesia (GNI). GNI is a cultural institution that uses several historical buildings that have been designated as heritage and have not been designated as a heritage according to the regulations applicable in Indonesia, in carrying out its function as the center of Indonesian art development and art museums. The revitalization of heritage buildings is taken as a step to meet space needs in running the current GNI function. In the revitalization master plan, there are physical interventions on the building of heritage and the removal of some historic buildings which will then be built new buildings at that location. The research matrix was used to map out the main elements of the study (the concept of GNI revitalization, heritage as identity, heritage as a process, and authenticity in the heritage). Expert interviews and document studies are the main tools used in collecting data. Qualitative data is then analyzed through content analysis and template analysis. This study identifies the significance of historic buildings (heritage buildings and buildings not defined as heritage) as an important value of history, architecture, education, and culture. The significance becomes the basis for revisiting the revitalization master plan which is then reviewed according to applicable regulations and the spatial layout of Jakarta. The interpretation that is built is (1) GNI is one of the elements of the embodiment of the National Cultural Center in the context of the region, where there are National Monument, National Museum and National Library in the same area, so the heritage not only gives identity to the past culture but the culture of current community; (2) The heritage should be seen as a dynamic cultural process towards the cultural change of community, where heritage must develop along with the urban development, so that the heritage buildings can remain alive and side by side with modern buildings but still observe the principles of preservation of heritage; (3) The authenticity of heritage should be able to balance the cultural heritage conservation approach with urban development, where authenticity can serve as a 'Value Transmitter' so that authenticity can be used to evaluate, preserve and manage heritage buildings by considering tangible and intangible aspects.Keywords: authenticity, culture process, identity, interpretation, revitalization
Procedia PDF Downloads 14813622 Monitoring the Drying and Grinding Process during Production of Celitement through a NIR-Spectroscopy Based Approach
Authors: Carolin Lutz, Jörg Matthes, Patrick Waibel, Ulrich Precht, Krassimir Garbev, Günter Beuchle, Uwe Schweike, Peter Stemmermann, Hubert B. Keller
Abstract:
Online measurement of the product quality is a challenging task in cement production, especially in the production of Celitement, a novel environmentally friendly hydraulic binder. The mineralogy and chemical composition of clinker in ordinary Portland cement production is measured by X-ray diffraction (XRD) and X ray fluorescence (XRF), where only crystalline constituents can be detected. But only a small part of the Celitement components can be measured via XRD, because most constituents have an amorphous structure. This paper describes the development of algorithms suitable for an on-line monitoring of the final processing step of Celitement based on NIR-data. For calibration intermediate products were dried at different temperatures and ground for variable durations. The products were analyzed using XRD and thermogravimetric analyses together with NIR-spectroscopy to investigate the dependency between the drying and the milling processes on one and the NIR-signal on the other side. As a result, different characteristic parameters have been defined. A short overview of the Celitement process and the challenging tasks of the online measurement and evaluation of the product quality will be presented. Subsequently, methods for systematic development of near-infrared calibration models and the determination of the final calibration model will be introduced. The application of the model on experimental data illustrates that NIR-spectroscopy allows for a quick and sufficiently exact determination of crucial process parameters.Keywords: calibration model, celitement, cementitious material, NIR spectroscopy
Procedia PDF Downloads 50013621 Innovation in Information Technology Services: Framework to Improve the Effectiveness and Efficiency of Information Technology Service Management Processes, Projects and Decision Support Management
Authors: Pablo Cardozo Herrera
Abstract:
In a dynamic market of Information Technology (IT) Service and with high quality demands and high performance requirements in decreasing costs, it is imperative that IT companies invest organizational effort in order to increase the effectiveness of their Information Technology Service Management (ITSM) processes through the improvement of ITSM project management and through solid support to the strategic decision-making process of IT directors. In this article, the author presents an analysis of common issues of IT companies around the world, with strategic needs of information unmet that provoke their ITSM processes and projects management that do not achieve the effectiveness and efficiency expected of their results. In response to the issues raised, the author proposes a framework consisting of an innovative theoretical framework model of ITSM management and a technological solution aligned to the Information Technology Infrastructure Library (ITIL) good practices guidance and ISO/IEC 20000-1 requirements. The article describes a research that proves the proposed framework is able to integrate, manage and coordinate in a holistic way, measurable and auditable, all ITSM processes and projects of IT organization and utilize the effectiveness assessment achieved for their strategic decision-making process increasing the process maturity level and improving the capacity of an efficient management.Keywords: innovation in IT services, ITSM processes, ITIL and ISO/IEC 20000-1, IT service management, IT service excellence
Procedia PDF Downloads 39713620 Modeling of Age Hardening Process Using Adaptive Neuro-Fuzzy Inference System: Results from Aluminum Alloy A356/Cow Horn Particulate Composite
Authors: Chidozie C. Nwobi-Okoye, Basil Q. Ochieze, Stanley Okiy
Abstract:
This research reports on the modeling of age hardening process using adaptive neuro-fuzzy inference system (ANFIS). The age hardening output (Hardness) was predicted using ANFIS. The input parameters were ageing time, temperature and percentage composition of cow horn particles (CHp%). The results show the correlation coefficient (R) of the predicted hardness values versus the measured values was of 0.9985. Subsequently, values outside the experimental data points were predicted. When the temperature was kept constant, and other input parameters were varied, the average relative error of the predicted values was 0.0931%. When the temperature was varied, and other input parameters kept constant, the average relative error of the hardness values predictions was 80%. The results show that ANFIS with coarse experimental data points for learning is not very effective in predicting process outputs in the age hardening operation of A356 alloy/CHp particulate composite. The fine experimental data requirements by ANFIS make it more expensive in modeling and optimization of age hardening operations of A356 alloy/CHp particulate composite.Keywords: adaptive neuro-fuzzy inference system (ANFIS), age hardening, aluminum alloy, metal matrix composite
Procedia PDF Downloads 15413619 An AK-Chart for the Non-Normal Data
Authors: Chia-Hau Liu, Tai-Yue Wang
Abstract:
Traditional multivariate control charts assume that measurement from manufacturing processes follows a multivariate normal distribution. However, this assumption may not hold or may be difficult to verify because not all the measurement from manufacturing processes are normal distributed in practice. This study develops a new multivariate control chart for monitoring the processes with non-normal data. We propose a mechanism based on integrating the one-class classification method and the adaptive technique. The adaptive technique is used to improve the sensitivity to small shift on one-class classification in statistical process control. In addition, this design provides an easy way to allocate the value of type I error so it is easier to be implemented. Finally, the simulation study and the real data from industry are used to demonstrate the effectiveness of the propose control charts.Keywords: multivariate control chart, statistical process control, one-class classification method, non-normal data
Procedia PDF Downloads 42313618 The Relationship between Confidence, Accuracy, and Decision Making in a Mobile Review Program
Authors: Carla Van De Sande, Jana Vandenberg
Abstract:
Just like physical skills, cognitive skills grow rusty over time unless they are regularly used and practiced, so academic breaks can have negative consequences on student learning and success. The Keeping in School Shape (KiSS) program is an engaging, accessible, and cost-effective intervention that harnesses the benefits of retrieval practice by using technology to help students maintain proficiency over breaks from school by delivering a daily review problem via text message or email. A growth mindset is promoted through feedback messages encouraging students to try again if they get a problem wrong and to take on a challenging problem if they get a problem correct. This paper reports on the relationship between confidence, accuracy, and decision-making during the implementation of the KiSS Program at a large university during winter break for students enrolled in an engineering introductory Calculus course sequence.Keywords: growth mindset, learning loss, on-the-go learning, retrieval practice
Procedia PDF Downloads 20513617 Cognitive Characteristics of Industrial Workers in Fuzzy Risk Assessment
Authors: Hyeon-Kyo Lim, Sang-Hun Byun
Abstract:
Risk assessment is carried out in most industrial plants for accident prevention, but there exists insufficient data for statistical decision making. It is commonly said that risk can be expressed as a product of consequence and likelihood of a corresponding hazard factor. Eventually, therefore, risk assessment involves human decision making which cannot be objective per se. This study was carried out to comprehend perceptive characteristics of human beings in industrial plants. Subjects were shown a set of illustrations describing scenes of industrial plants, and were asked to assess the risk of each scene with not only linguistic variables but also numeric scores in the aspect of consequence and likelihood. After that, their responses were formulated as fuzzy membership functions, and compared with those of university students who had no experience of industrial works. The results showed that risk level of industrial workers were lower than those of any other groups, which implied that the workers might generally have a tendency to neglect more hazard factors in their work fields.Keywords: fuzzy, hazard, linguistic variable, risk assessment
Procedia PDF Downloads 25513616 Efficiency of a Semantic Approach in Teaching Foreign Languages
Authors: Genady Shlomper
Abstract:
During the process of language teaching, each teacher faces some general and some specific problems. Some of these problems are mutual to all languages because they yield to the rules of cognition, conscience, perception, understanding and memory; to the physiological and psychological principles pertaining to the human race irrespective of origin and nationality. Still, every language is a distinctive system, possessing individual properties and an obvious identity, as a result of a development in specific natural, geographical, cultural and historical conditions. The individual properties emerge in the script, in the phonetics, morphology and syntax. All these problems can and should be a subject of a detailed research and scientific analysis, mainly from practical considerations and language teaching requirements. There are some formidable obstacles in the language acquisition process. Among the first to be mentioned is the existence of concepts and entire categories in foreign languages, which are absent in the language of the students. Such phenomena reflect specific ways of thinking and the world-outlook, which were shaped during the evolution. Hindi is the national language of India, which belongs to the group of Indo-Iranian languages from the Indo-European family of languages. The lecturer has gained experience in teaching Hindi language to native speakers of Uzbek, Russian and Hebrew languages. He will show the difficulties in the field of phonetics, morphology and syntax, which the students have to deal with during the acquisition of the language. In the proposed lecture the lecturer will share his experience in making the process of language teaching more efficient by using non-formal semantic approach.Keywords: applied linguistics, foreign language teaching, language teaching methodology, semantics
Procedia PDF Downloads 35613615 Multivariate Simulations of the Process of Forming the Automotive Connector Forging from ZK60 Alloy
Authors: Anna Dziubinska
Abstract:
The article presents the results of numerical simulations of the new forging process of the automotive connector forging from cast preform. The high-strength ZK60 alloy (belonging to the Mg-Zn-Zr group of Mg alloys) was selected for numerical tests. Currently, this part of the industry is produced by multi-stage forging consisting of operations: bending, preforming, and finishing. The use of the cast preform would enable forging this component in one operation. However, obtaining specific mechanical properties requires inducing a certain level of strain within the forged part. Therefore, the design of the preform, its shape, and volume are of paramount importance. In work presented in this article, preforms of different shapes were designed and assessed using Finite Element (FE) analysis. The research was funded by the Polish National Agency for Academic Exchange within the framework of the Bekker programme.Keywords: automotive connector, forging, magnesium alloy, numerical simulation, preform, ZK60
Procedia PDF Downloads 13213614 A General Framework for Measuring the Internal Fraud Risk of an Enterprise Resource Planning System
Authors: Imran Dayan, Ashiqul Khan
Abstract:
Internal corporate fraud, which is fraud carried out by internal stakeholders of a company, affects the well-being of the organisation just like its external counterpart. Even if such an act is carried out for the short-term benefit of a corporation, the act is ultimately harmful to the entity in the long run. Internal fraud is often carried out by relying upon aberrations from usual business processes. Business processes are the lifeblood of a company in modern managerial context. Such processes are developed and fine-tuned over time as a corporation grows through its life stages. Modern corporations have embraced technological innovations into their business processes, and Enterprise Resource Planning (ERP) systems being at the heart of such business processes is a testimony to that. Since ERP systems record a huge amount of data in their event logs, the logs are a treasure trove for anyone trying to detect any sort of fraudulent activities hidden within the day-to-day business operations and processes. This research utilises the ERP systems in place within corporations to assess the likelihood of prospective internal fraud through developing a framework for measuring the risks of fraud through Process Mining techniques and hence finds risky designs and loose ends within these business processes. This framework helps not only in identifying existing cases of fraud in the records of the event log, but also signals the overall riskiness of certain business processes, and hence draws attention for carrying out a redesign of such processes to reduce the chance of future internal fraud while improving internal control within the organisation. The research adds value by applying the concepts of Process Mining into the analysis of data from modern day applications of business process records, which is the ERP event logs, and develops a framework that should be useful to internal stakeholders for strengthening internal control as well as provide external auditors with a tool of use in case of suspicion. The research proves its usefulness through a few case studies conducted with respect to big corporations with complex business processes and an ERP in place.Keywords: enterprise resource planning, fraud risk framework, internal corporate fraud, process mining
Procedia PDF Downloads 33513613 Prediction of Compressive Strength of Concrete from Early Age Test Result Using Design of Experiments (Rsm)
Authors: Salem Alsanusi, Loubna Bentaher
Abstract:
Response Surface Methods (RSM) provide statistically validated predictive models that can then be manipulated for finding optimal process configurations. Variation transmitted to responses from poorly controlled process factors can be accounted for by the mathematical technique of propagation of error (POE), which facilitates ‘finding the flats’ on the surfaces generated by RSM. The dual response approach to RSM captures the standard deviation of the output as well as the average. It accounts for unknown sources of variation. Dual response plus propagation of error (POE) provides a more useful model of overall response variation. In our case, we implemented this technique in predicting compressive strength of concrete of 28 days in age. Since 28 days is quite time consuming, while it is important to ensure the quality control process. This paper investigates the potential of using design of experiments (DOE-RSM) to predict the compressive strength of concrete at 28th day. Data used for this study was carried out from experiment schemes at university of Benghazi, civil engineering department. A total of 114 sets of data were implemented. ACI mix design method was utilized for the mix design. No admixtures were used, only the main concrete mix constituents such as cement, coarse-aggregate, fine aggregate and water were utilized in all mixes. Different mix proportions of the ingredients and different water cement ratio were used. The proposed mathematical models are capable of predicting the required concrete compressive strength of concrete from early ages.Keywords: mix proportioning, response surface methodology, compressive strength, optimal design
Procedia PDF Downloads 26713612 Scaling-Down an Agricultural Waste Biogas Plant Fermenter
Authors: Matheus Pessoa, Matthias Kraume
Abstract:
Scale-Down rules in process engineering help us to improve and develop Industrial scale parameters into lab scale. Several scale-down rules available in the literature like Impeller Power Number, Agitation device Power Input, Substrate Tip Speed, Reynolds Number and Cavern Development were investigated in order to stipulate the rotational speed to operate an 11 L working volume lab-scale bioreactor within industrial process parameters. Herein, xanthan gum was used as a fluid with a representative viscosity of a hypothetical biogas plant, with H/D = 1 and central agitation, fermentation broth using sewage sludge and sugar beet pulp as substrate. The results showed that the cavern development strategy was the best method for establishing a rotational speed for the bioreactor operation, while the other rules presented values out of reality for this article proposes.Keywords: anaerobic digestion, cavern development, scale down rules, xanthan gum
Procedia PDF Downloads 49313611 A Multi-criteria Decision Method For The Recruitment Of Academic Personnel Based On The Analytical Hierarchy Process And The Delphi Method In A Neutrosophic Environment (Full Text)
Authors: Antonios Paraskevas, Michael Madas
Abstract:
For a university to maintain its international competitiveness in education, it is essential to recruit qualitative academic staff as it constitutes its most valuable asset. This selection demonstrates a significant role in achieving strategic objectives, particularly by emphasizing a firm commitment to exceptional student experience and innovative teaching and learning practices of high quality. In this vein, the appropriate selection of academic staff establishes a very important factor of competitiveness, efficiency and reputation of an academic institute. Within this framework, our work demonstrates a comprehensive methodological concept that emphasizes on the multi-criteria nature of the problem and on how decision makers could utilize our approach in order to proceed to the appropriate judgment. The conceptual framework introduced in this paper is built upon a hybrid neutrosophic method based on the Neutrosophic Analytical Hierarchy Process (N-AHP), which uses the theory of neutrosophy sets and is considered suitable in terms of significant degree of ambiguity and indeterminacy observed in decision-making process. To this end, our framework extends the N-AHP by incorporating the Neutrosophic Delphi Method (N-DM). By applying the N-DM, we can take into consideration the importance of each decision-maker and their preferences per evaluation criterion. To the best of our knowledge, the proposed model is the first which applies Neutrosophic Delphi Method in the selection of academic staff. As a case study, it was decided to use our method to a real problem of academic personnel selection, having as main goal to enhance the algorithm proposed in previous scholars’ work, and thus taking care of the inherit ineffectiveness which becomes apparent in traditional multi-criteria decision-making methods when dealing with situations alike. As a further result, we prove that our method demonstrates greater applicability and reliability when compared to other decision models.Keywords: analytical hierarchy process, delphi method, multi-criteria decision maiking method, neutrosophic set theory, personnel recruitment
Procedia PDF Downloads 20013610 Ideological Manipulations and Cultural-Norm Constraints
Authors: Masoud Hassanzade Novin, Bahloul Salmani
Abstract:
Translation cannot be considered as a simple linguistic act. Through the rise of descriptive approach in the late 1970s and 1980s, translation process managed to meet the requirements of social aspects as well as linguistic approaches. To have the translation considered as the cross-cultural communication through which various cultures communicate in ideological and cultural constraints, the contrastive analysis was conducted in this paper to reveal the distortions imposed in the translated texts. The corpus of the study involved the novel 1984 written by George Orwell and its Persian translated texts which were analyzed through the qualitative type of the research based on critical discourse analysis (CDA) and Toury's norms as well as Lefever's concepts of ideology. Results of the study revealed the point that ideology and the cultural constraints were considered as an important stimulus which can control the process of the translation.Keywords: critical discourse analysis, ideology, norms, translated texts
Procedia PDF Downloads 33613609 Problem Based Learning and Teaching by Example in Dimensioning of Mechanisms: Feedback
Authors: Nicolas Peyret, Sylvain Courtois, Gaël Chevallier
Abstract:
This article outlines the development of the Project Based Learning (PBL) at the level of a last year’s Bachelor’s Degree. This form of pedagogy has for objective to allow a better involving of the students from the beginning of the module. The theoretical contributions are introduced during the project to solving a technological problem. The module in question is the module of mechanical dimensioning method of Supméca a French engineering school. This school issues a Master’s Degree. While the teaching methods used in primary and secondary education are frequently renewed in France at the instigation of teachers and inspectors, higher education remains relatively traditional in its practices. Recently, some colleagues have felt the need to put the application back at the heart of their theoretical teaching. This need is induced by the difficulty of covering all the knowledge deductively before its application. It is therefore tempting to make the students 'learn by doing', even if it doesn’t cover some parts of the theoretical knowledge. The other argument that supports this type of learning is the lack of motivation the students have for the magisterial courses. The role-play allowed scenarios favoring interaction between students and teachers… However, this pedagogical form known as 'pedagogy by project' is difficult to apply in the first years of university studies because of the low level of autonomy and individual responsibility that the students have. The question of what the student actually learns from the initial program as well as the evaluation of the competences acquired by the students in this type of pedagogy also remains an open problem. Thus we propose to add to the pedagogy by project format a regressive part of interventionism by the teacher based on pedagogy by example. This pedagogical scenario is based on the cognitive load theory and Bruner's constructivist theory. It has been built by relying on the six points of the encouragement process defined by Bruner, with a concrete objective, to allow the students to go beyond the basic skills of dimensioning and allow them to acquire the more global skills of engineering. The implementation of project-based teaching coupled with pedagogy by example makes it possible to compensate for the lack of experience and autonomy of first-year students, while at the same time involving them strongly in the first few minutes of the module. In this project, students have been confronted with the real dimensioning problems and are able to understand the links and influences between parameter variations and dimensioning, an objective that we did not reach in classical teaching. It is this form of pedagogy which allows to accelerate the mastery of basic skills and so spend more time on the engineer skills namely the convergence of each dimensioning in order to obtain a validated mechanism. A self-evaluation of the project skills acquired by the students will also be presented.Keywords: Bruner's constructivist theory, mechanisms dimensioning, pedagogy by example, problem based learning
Procedia PDF Downloads 19013608 High Pressure Delignification Process for Nanocrystalline Cellulose Production from Agro-Waste Biomass
Authors: Sakinul Islam, Nhol Kao, Sati Bhattacharya, Rahul Gupta
Abstract:
Nanocrystalline cellulose (NCC) has been widely used for miscellaneous applications due to its superior properties over other nanomaterials. However, the major problems associated with the production of NCC are long reaction time, low production rate and inefficient process. The mass production of NCC within a short period of time is still a great challenge. The main objective of this study is to produce NCC from rice husk agro waste biomass from a high pressure delignification process (HPDP), followed by bleaching and hydrolysis processes. The HPDP has not been explored for NCC production from rice husk biomass (RHB) until now. In order to produce NCC, powder rice husk (PRH) was placed into a stainless steel reactor at 80 ˚C under 5 bars. Aqueous solution of NaOH (4M) was used for the dissolution of lignin and other amorphous impurities from PRH. After certain experimental times (1h, 3.5h and 6h), bleaching and hydrolysis were carried out on delignified samples. NaOCl (20%) and H2SO4 (4M) solutions were used for bleaching and hydrolysis processes, respectively. The NCC suspension from hydrolysis was sonicated and neutralized by buffer solution for various characterisations. Finally NCC suspension was dried and analyzed by FTIR, XRD, SEM, AFM and TEM. The chemical composition of NCC and PRH was estimated by TAPPI (Technical Association of Pulp and Paper Industry) standard methods to observe the product purity. It was found that, the 6h of the HPDP was more efficient to produce good quality NCC than that at 1h and 3.5h due to low separation of non-cellulosic components from RHB. The analyses indicated the crystallinity of NCC to be 71 %, particle size of 20-50 nm (diameter) and 100-200 nm in length.Keywords: nanocrystalline cellulose, NCC, high pressure delignification, bleaching, hydrolysis, agro-waste biomass
Procedia PDF Downloads 26413607 Improvement of Ground Truth Data for Eye Location on Infrared Driver Recordings
Authors: Sorin Valcan, Mihail Gaianu
Abstract:
Labeling is a very costly and time consuming process which aims to generate datasets for training neural networks in several functionalities and projects. For driver monitoring system projects, the need for labeled images has a significant impact on the budget and distribution of effort. This paper presents the modifications done to an algorithm used for the generation of ground truth data for 2D eyes location on infrared images with drivers in order to improve the quality of the data and performance of the trained neural networks. The algorithm restrictions become tougher, which makes it more accurate but also less constant. The resulting dataset becomes smaller and shall not be altered by any kind of manual label adjustment before being used in the neural networks training process. These changes resulted in a much better performance of the trained neural networks.Keywords: labeling automation, infrared camera, driver monitoring, eye detection, convolutional neural networks
Procedia PDF Downloads 11813606 Rehabilitation Team after Brain Damages as Complex System Integrating Consciousness
Authors: Olga Maksakova
Abstract:
A work with unconscious patients after acute brain damages besides special knowledge and practical skills of all the participants requires a very specific organization. A lot of said about team approach in neurorehabilitation, usually as for outpatient mode. Rehabilitologists deal with fixed patient problems or deficits (motion, speech, cognitive or emotional disorder). Team-building means superficial paradigm of management psychology. Linear mode of teamwork fits casual relationships there. Cases with deep altered states of consciousness (vegetative states, coma, and confusion) require non-linear mode of teamwork: recovery of consciousness might not be the goal due to phenomenon uncertainty. Rehabilitation team as Semi-open Complex System includes the patient as a part. Patient's response pattern becomes formed not only with brain deficits but questions-stimuli, context, and inquiring person. Teamwork is sourcing of phenomenology knowledge of patient's processes as Third-person approach is replaced with Second- and after First-person approaches. Here is a chance for real-time change. Patient’s contacts with his own body and outward things create a basement for restoration of consciousness. The most important condition is systematic feedbacks to any minimal movement or vegetative signal of the patient. Up to now, recovery work with the most severe contingent is carried out in the mode of passive physical interventions, while an effective rehabilitation team should include specially trained psychologists and psychotherapists. It is they who are able to create a network of feedbacks with the patient and inter-professional ones building up the team. Characteristics of ‘Team-Patient’ system (TPS) are energy, entropy, and complexity. Impairment of consciousness as the absence of linear contact appears together with a loss of essential functions (low energy), vegetative-visceral fits (excessive energy and low order), motor agitation (excessive energy and excessive order), etc. Techniques of teamwork are different in these cases for resulting optimization of the system condition. Directed regulation of the system complexity is one of the recovery tools. Different signs of awareness appear as a result of system self-organization. Joint meetings are an important part of teamwork. Regular or event-related discussions form the language of inter-professional communication, as well as the patient's shared mental model. Analysis of complex communication process in TPS may be useful for creation of the general theory of consciousness.Keywords: rehabilitation team, urgent rehabilitation, severe brain damage, consciousness disorders, complex system theory
Procedia PDF Downloads 14613605 Leaching of Copper from Copper Ore Using Sulphuric Acid in the Presence of Hydrogen Peroxide as an Oxidizing Agent: An Optimized Process
Authors: Hilary Rutto
Abstract:
Leaching with acids are the most commonly reagents used to remove copper ions from its copper ores. It is important that the process conditions are optimized to improve the leaching efficiency. In the present study the effects of pH, oxidizing agent (hydrogen peroxide), stirring speed, solid to liquid ratio and acid concentration on the leaching of copper ions from it ore were investigated using a pH Stat apparatus. Copper ions were analyzed at the end of each experiment using Atomic Absorption (AAS) machine. Results showed that leaching efficiency improved with an increase in acid concentration, stirring speed, oxidizing agent, pH and decreased with an increase in the solid to liquid ratio.Keywords: leaching, copper, oxidizing agent, pH stat apparatus
Procedia PDF Downloads 37713604 Design of the Fiber Lay-Up for the Composite Wind Turbine Blade in VARTM
Authors: Tzai-Shiung Li, Wen-Bin Young
Abstract:
The wind turbine blade sustains various kinds of loadings during the operating and parking state. Due to the increasing size of the wind turbine blade, it is important to arrange the composite materials in a sufficient way to reach the optimal utilization of the material strength. In the fabrication process of the vacuum assisted resin transfer molding, the fiber content of the turbine blade depends on the vacuum pressure. In this study, a design of the fiber layup for the vacuum assisted resin transfer molding is conducted to achieve the efficient utilization the material strength. This design is for the wind turbine blade consisting of shell skins with or without the spar structure.Keywords: resin film infiltration, vacuum assisted resin transfer molding process, wind turbine blade, composite materials
Procedia PDF Downloads 38413603 Statistically Significant Differences of Carbon Dioxide and Carbon Monoxide Emission in Photocopying Process
Authors: Kiurski S. Jelena, Kecić S. Vesna, Oros B. Ivana
Abstract:
Experimental results confirmed the temporal variation of carbon dioxide and carbon monoxide concentration during the working shift of the photocopying process in a small photocopying shop in Novi Sad, Serbia. The statistically significant differences of target gases were examined with two-way analysis of variance without replication followed by Scheffe's post hoc test. The existence of statistically significant differences was obtained for carbon monoxide emission which is pointed out with F-values (12.37 and 31.88) greater than Fcrit (6.94) in contrary to carbon dioxide emission (F-values of 1.23 and 3.12 were less than Fcrit). Scheffe's post hoc test indicated that sampling point A (near the photocopier machine) and second time interval contribute the most on carbon monoxide emission.Keywords: analysis of variance, carbon dioxide, carbon monoxide, photocopying indoor, Scheffe's test
Procedia PDF Downloads 32713602 Design of a Tool for Generating Test Cases from BPMN
Authors: Prat Yotyawilai, Taratip Suwannasart
Abstract:
Business Process Model and Notation (BPMN) is more important in the business process and creating functional models, and is a standard for OMG, which becomes popular in various organizations and in education. Researches related to software testing based on models are prominent. Although most researches use the UML model in software testing, not many researches use the BPMN Model in creating test cases. Therefore, this research proposes a design of a tool for generating test cases from the BPMN. The model is analyzed and the details of the various components are extracted before creating a flow graph. Both details of components and the flow graph are used in generating test cases.Keywords: software testing, test case, BPMN, flow graph
Procedia PDF Downloads 55513601 Efficient Position Based Operation Code Authentication
Authors: Hashim Ali, Sheheryar Khan
Abstract:
Security for applications is always been a keen issue of concern. In general, security is to allow access of grant to legal user or to deny non-authorized access to the system. Shoulder surfing is an observation technique to hack an account or to enter into a system. When a malicious observer is capturing or recording the fingers of a user while he is entering sensitive inputs (PIN, Passwords etc.) and may be able to observe user’s password credential. It is very rigorous for a novice user to prevent himself from shoulder surfing or unaided observer in a public place while accessing his account. In order to secure the user account, there are five factors of authentication; they are: “(i) something you have, (ii) something you are, (iii) something you know, (iv) somebody you know, (v) something you process”. A technique has been developed of fifth-factor authentication “something you process” to provide novel approach to the user. In this paper, we have applied position based operational code authentication in such a way to more easy and user friendly to the user.Keywords: shoulder surfing, malicious observer, sensitive inputs, authentication
Procedia PDF Downloads 27213600 The Determinants of Co-Production for Value Co-Creation: Quadratic Effects
Authors: Li-Wei Wu, Chung-Yu Wang
Abstract:
Recently, interest has been generated in the search for a new reference framework for value creation that is centered on the co-creation process. Co-creation implies cooperative value creation between service firms and customers and requires the building of experiences as well as the resolution of problems through the combined effort of the parties in the relationship. For customers, values are always co-created through their participation in services. Customers can ultimately determine the value of the service in use. This new approach emphasizes that a customer’s participation in the service process is considered indispensable to value co-creation. An important feature of service in the context of exchange is co-production, which implies that a certain amount of participation is needed from customers to co-produce a service and hence co-create value. Co-production no doubt helps customers better understand and take charge of their own roles in the service process. Thus, this proposal is to encourage co-production, thus facilitating value co-creation of that is reflected in both customers and service firms. Four determinants of co-production are identified in this study, namely, commitment, trust, asset specificity, and decision-making uncertainty. Commitment is an essential dimension that directly results in successful cooperative behaviors. Trust helps establish a relational environment that is fundamental to cross-border cooperation. Asset specificity motivates co-production because this determinant may enhance return on asset investment. Decision-making uncertainty prompts customers to collaborate with service firms in making decisions. In other words, customers adjust their roles and are increasingly engaged in co-production when commitment, trust, asset specificity, and decision-making uncertainty are enhanced. Although studies have examined the preceding effects, to our best knowledge, none has empirically examined the simultaneous effects of all the curvilinear relationships in a single study. When these determinants are excessive, however, customers will not engage in co-production process. In brief, we suggest that the relationships of commitment, trust, asset specificity, and decision-making uncertainty with co-production are curvilinear or are inverse U-shaped. These new forms of curvilinear relationships have not been identified in existing literature on co-production; therefore, they complement extant linear approaches. Most importantly, we aim to consider both the bright and the dark sides of the determinants of co-production.Keywords: co-production, commitment, trust, asset specificity, decision-making uncertainty
Procedia PDF Downloads 18813599 Optimizing Approach for Sifting Process to Solve a Common Type of Empirical Mode Decomposition Mode Mixing
Authors: Saad Al-Baddai, Karema Al-Subari, Elmar Lang, Bernd Ludwig
Abstract:
Empirical mode decomposition (EMD), a new data-driven of time-series decomposition, has the advantage of supposing that a time series is non-linear or non-stationary, as is implicitly achieved in Fourier decomposition. However, the EMD suffers of mode mixing problem in some cases. The aim of this paper is to present a solution for a common type of signals causing of EMD mode mixing problem, in case a signal suffers of an intermittency. By an artificial example, the solution shows superior performance in terms of cope EMD mode mixing problem comparing with the conventional EMD and Ensemble Empirical Mode decomposition (EEMD). Furthermore, the over-sifting problem is also completely avoided; and computation load is reduced roughly six times compared with EEMD, an ensemble number of 50.Keywords: empirical mode decomposition (EMD), mode mixing, sifting process, over-sifting
Procedia PDF Downloads 39513598 Performance, Scalability and Reliability Engineering: Shift Left and Shift Right Approach
Authors: Jyothirmayee Pola
Abstract:
Ideally, a test-driven development (TDD) or agile or any other process should be able to define and implement performance, scalability, and reliability (PSR) of the product with a higher quality of service (QOS) and should have the ability to fix any PSR issues with lesser cost before it hits the production. Most PSR test strategies for new product introduction (NPI) include assumptions about production load requirements but never accurate. NPE (New product Enhancement) include assumptions for new features that are being developed whilst workload distribution for older features can be derived by analyzing production transactions. This paper talks about how to shift left PSR towards design phase of release management process to get better QOS w.r.t PSR for any product under development. It also explains the ROI for future customer onboarding both for Service Oriented Architectures (SOA) and Microservices architectures and how to define PSR requirements.Keywords: component PSR, performance engineering, performance tuning, reliability, return on investment, scalability, system PSR
Procedia PDF Downloads 7613597 Transesterification of Jojoba Oil Wax Using Microwave Technique
Authors: Moataz Elsawy, Hala F. Naguib, Hilda A. Aziz, Eid A. Ismail, Labiba I. Hussein, Maher Z. Elsabee
Abstract:
Jojoba oil-wax is extracted from the seeds of the jojoba (Simmondsia chinensis Link Schneider), a perennial shrub that grows in semi-desert areas in Egypt and in some parts of the world. The main uses of jojoba oil wax are in the cosmetics and pharmaceutical industry, but new uses could arise related to the search of new energetic crops. This paper summarizes a process to convert the jojoba oil wax to biodiesel by transesterification with ethanol and a series of aliphatic alcohols using a more economic and energy saving method in a domestic microwave. The effect of time and power of the microwave on the extent of the transesterification using ethanol and other aliphatic alcohols has been studied. The separation of the alkyl esters from the fatty alcohols rich fraction has been done in a single crystallization step at low temperature (−18°C) from low boiling point petroleum ether. Gas chromatography has been used to follow up the transesterification process. All products have been characterized by spectral analysis.Keywords: jojoba oil, transesterification, microwave, gas chromatography jojoba esters, jojoba alcohol
Procedia PDF Downloads 46213596 Study of Biofuel Produced by Babassu Oil Fatty Acids Esterification
Authors: F. A. F. da Ponte, J. Q. Malveira, I. A. Maciel, M. C. G. Albuquerque
Abstract:
In this work aviation, biofuel production was studied by fatty acids (C6 to C16) esterification. The process variables in heterogeneous catalysis were evaluated using an experimental design. Temperature and reaction time were the studied parameters, and the methyl esters content was the response of the experimental design. An ion exchange resin was used as a heterogeneous catalyst. The process optimization was carried out using response surface methodology (RSM) and polynomial model of second order. Results show that the most influential variables on the linear coefficient of each effect studied were temperature and reaction time. The best result of methyl esters conversion in the experimental design was under the conditions: 10% wt of catalyst; 100 °C and 4 hours of reaction. The best-achieved conversion was 96.5% wt of biofuel.Keywords: esterification, ion-exchange resins, response surface methodology, biofuel
Procedia PDF Downloads 49613595 Development of a Triangular Evaluation Protocol in a Multidisciplinary Design Process of an Ergometric Step
Authors: M. B. Ricardo De Oliveira, A. Borghi-Silva, E. Paravizo, F. Lizarelli, L. Di Thomazzo, D. Braatz
Abstract:
Prototypes are a critical feature in the product development process, as they help the project team visualize early concept flaws, communicate ideas and introduce an initial product testing. Involving stakeholders, such as consumers and users, in prototype tests allows the gathering of valuable feedback, contributing for a better product and making the design process more participatory. Even though recent studies have shown that user evaluation of prototypes is valuable, few articles provide a method or protocol on how designers should conduct it. This multidisciplinary study (involving the areas of physiotherapy, engineering and computer science) aims to develop an evaluation protocol, using an ergometric step prototype as the product prototype to be assessed. The protocol consisted of performing two tests (the 2 Minute Step Test and the Portability Test) to allow users (patients) and consumers (physiotherapists) to have an experience with the prototype. Furthermore, the protocol contained four Likert-Scale questionnaires (one for users and three for consumers), that inquired participants about how they perceived the design characteristics of the product (performance, safety, materials, maintenance, portability, usability and ergonomics), in their use of the prototype. Additionally, the protocol indicated the need to conduct interviews with the product designers, in order to link their feedback to the ones from the consumers and users. Both tests and interviews were recorded for further analysis. The participation criteria for the study was gender and age for patients, gender and experience with 2 Minute Step Test for physiotherapists and involvement level in the product development project for designers. The questionnaire's reliability was validated using Cronbach's Alpha and the quantitative data of the questionnaires were analyzed using non-parametric hypothesis tests with a significance level of 0.05 (p <0.05) and descriptive statistics. As a result, this study provides a concise evaluation protocol which can assist designers in their development process, collecting quantitative feedback from consumer and users, and qualitative feedback from designers.Keywords: Product Design, Product Evaluation, Prototypes, Step
Procedia PDF Downloads 11813594 Performance Analysis of the Time-Based and Periodogram-Based Energy Detector for Spectrum Sensing
Authors: Sadaf Nawaz, Adnan Ahmed Khan, Asad Mahmood, Chaudhary Farrukh Javed
Abstract:
Classically, an energy detector is implemented in time domain (TD). However, frequency domain (FD) based energy detector has demonstrated an improved performance. This paper presents a comparison between the two approaches as to analyze their pros and cons. A detailed performance analysis of the classical TD energy-detector and the periodogram based detector is performed. Exact and approximate mathematical expressions for probability of false alarm (Pf) and probability of detection (Pd) are derived for both approaches. The derived expressions naturally lead to an analytical as well as intuitive reasoning for the improved performance of (Pf) and (Pd) in different scenarios. Our analysis suggests the dependence improvement on buffer sizes. Pf is improved in FD, whereas Pd is enhanced in TD based energy detectors. Finally, Monte Carlo simulations results demonstrate the analysis reached by the derived expressions.Keywords: cognitive radio, energy detector, periodogram, spectrum sensing
Procedia PDF Downloads 378