Search results for: subsidiary performance
7821 Improving Productivity in a Glass Production Line through Applying Principles of Total Productive Maintenance (TPM)
Authors: Omar Bataineh
Abstract:
Total productive maintenance (TPM) is a principle-based method that aims to get a high-level production with no breakdowns, no slow running and no defects. Key principles of TPM were applied in this work to improve the performance of the glass production line at United Beverage Company in Kuwait, which is producing bottles of soft drinks. Principles such as 5S as a foundation for TPM implementation, developing a program for equipment management, Cause and Effect Analysis (CEA), quality improvement, training and education of employees were employed. After the completion of TPM implementation, it was possible to increase the Overall Equipment Effectiveness (OEE) from 23% to 40%. Procedia PDF Downloads 3377820 Characterization of the Music Admission Requirements and Evaluation of the Relationship among Motivation and Performance Achievement
Authors: Antonio M. Oliveira, Patricia Oliveira-Silva, Jose Matias Alves, Gary McPherson
Abstract:
The music teaching is oriented towards offering formal music training. Due to its specificities, this vocational program starts at a very young age. Although provided by the State, the offer is limited to 6 schools throughout the country, which means that the vacancies for prospective students are very limited every year. It is therefore crucial that these vacancies be taken by especially motivated children grown within households that offer the ideal setting for success. Some of the instruments used to evaluate musical performance are highly sensitive to specific previous training, what represents a severe validity problem for testing children who have had restricted opportunities for formal training. Moreover, these practices may be unfair because, for instance, they may not reflect the candidates’ music aptitudes. Based on what constitutes a prerequisite for making an excellent music student, researchers in this field have long argued that motivation, task commitment, and parents’ support are as important as ability. Thus, the aim of this study is: (1) to prepare an inventory of admission requirements in Australia, Portugal and Ireland; (2) to examine whether the candidates to music conservatories and parents’ level of motivation, assessed at three evaluation points (i.e., admission, at the end of the first year, and at the end of the second year), correlates positively with the candidates’ progress in learning a musical instrument (i.e., whether motivation at the admission may predict student musicianship); (3) an adaptation of an existing instrument to assess the motivation (i.e., to adapt the items to the music setting, focusing on the motivation for playing a musical instrument). The inclusion criteria are: only children registered in the administrative services to be evaluated for entrance to the conservatory will be accepted for this study. The expected number of participants is fifty (5-6 years old) in all the three frequency schemes: integrated, articulated and supplementary. Revisiting musical admission procedures is of particular importance and relevance to musical education because this debate may bring guidance and assistance about the needed improvement to make the process of admission fairer and more transparent.Keywords: music learning, music admission requirements, student’s motivation, parent’s motivation
Procedia PDF Downloads 1667819 Unsupervised Part-of-Speech Tagging for Amharic Using K-Means Clustering
Authors: Zelalem Fantahun
Abstract:
Part-of-speech tagging is the process of assigning a part-of-speech or other lexical class marker to each word into naturally occurring text. Part-of-speech tagging is the most fundamental and basic task almost in all natural language processing. In natural language processing, the problem of providing large amount of manually annotated data is a knowledge acquisition bottleneck. Since, Amharic is one of under-resourced language, the availability of tagged corpus is the bottleneck problem for natural language processing especially for POS tagging. A promising direction to tackle this problem is to provide a system that does not require manually tagged data. In unsupervised learning, the learner is not provided with classifications. Unsupervised algorithms seek out similarity between pieces of data in order to determine whether they can be characterized as forming a group. This paper explicates the development of unsupervised part-of-speech tagger using K-Means clustering for Amharic language since large amount of data is produced in day-to-day activities. In the development of the tagger, the following procedures are followed. First, the unlabeled data (raw text) is divided into 10 folds and tokenization phase takes place; at this level, the raw text is chunked at sentence level and then into words. The second phase is feature extraction which includes word frequency, syntactic and morphological features of a word. The third phase is clustering. Among different clustering algorithms, K-means is selected and implemented in this study that brings group of similar words together. The fourth phase is mapping, which deals with looking at each cluster carefully and the most common tag is assigned to a group. This study finds out two features that are capable of distinguishing one part-of-speech from others these are morphological feature and positional information and show that it is possible to use unsupervised learning for Amharic POS tagging. In order to increase performance of the unsupervised part-of-speech tagger, there is a need to incorporate other features that are not included in this study, such as semantic related information. Finally, based on experimental result, the performance of the system achieves a maximum of 81% accuracy.Keywords: POS tagging, Amharic, unsupervised learning, k-means
Procedia PDF Downloads 4517818 Mechanical Properties of Poly(Propylene)-Based Graphene Nanocomposites
Authors: Luiza Melo De Lima, Tito Trindade, Jose M. Oliveira
Abstract:
The development of thermoplastic-based graphene nanocomposites has been of great interest not only to the scientific community but also to different industrial sectors. Due to the possible improvement of performance and weight reduction, thermoplastic nanocomposites are a great promise as a new class of materials. These nanocomposites are of relevance for the automotive industry, namely because the emission limits of CO2 emissions imposed by the European Commission (EC) regulations can be fulfilled without compromising the car’s performance but by reducing its weight. Thermoplastic polymers have some advantages over thermosetting polymers such as higher productivity, lower density, and recyclability. In the automotive industry, for example, poly(propylene) (PP) is a common thermoplastic polymer, which represents more than half of the polymeric raw material used in automotive parts. Graphene-based materials (GBM) are potential nanofillers that can improve the properties of polymer matrices at very low loading. In comparison to other composites, such as fiber-based composites, weight reduction can positively affect their processing and future applications. However, the properties and performance of GBM/polymer nanocomposites depend on the type of GBM and polymer matrix, the degree of dispersion, and especially the type of interactions between the fillers and the polymer matrix. In order to take advantage of the superior mechanical strength of GBM, strong interfacial strength between GBM and the polymer matrix is required for efficient stress transfer from GBM to the polymer. Thus, chemical compatibilizers and physicochemical modifications have been reported as important tools during the processing of these nanocomposites. In this study, PP-based nanocomposites were obtained by a simple melt blending technique, using a Brabender type mixer machine. Graphene nanoplatelets (GnPs) were applied as structural reinforcement. Two compatibilizers were used to improve the interaction between PP matrix and GnPs: PP graft maleic anhydride (PPgMA) and PPgMA modified with tertiary amine alcohol (PPgDM). The samples for tensile and Charpy impact tests were obtained by injection molding. The results suggested the GnPs presence can increase the mechanical strength of the polymer. However, it was verified that the GnPs presence can promote a decrease of impact resistance, turning the nanocomposites more fragile than neat PP. The compatibilizers’ incorporation increases the impact resistance, suggesting that the compatibilizers can enhance the adhesion between PP and GnPs. Compared to neat PP, Young’s modulus of non-compatibilized nanocomposite increase demonstrated that GnPs incorporation can promote a stiffness improvement of the polymer. This trend can be related to the several physical crosslinking points between the PP matrix and the GnPs. Furthermore, the decrease of strain at a yield of PP/GnPs, together with the enhancement of Young’s modulus, confirms that the GnPs incorporation led to an increase in stiffness but to a decrease in toughness. Moreover, the results demonstrated that incorporation of compatibilizers did not affect Young’s modulus and strain at yield results compared to non-compatibilized nanocomposite. The incorporation of these compatibilizers showed an improvement of nanocomposites’ mechanical properties compared both to those the non-compatibilized nanocomposite and to a PP sample used as reference.Keywords: graphene nanoplatelets, mechanical properties, melt blending processing, poly(propylene)-based nanocomposites
Procedia PDF Downloads 1877817 The Fabrication and Characterization of a Honeycomb Ceramic Electric Heater with a Conductive Coating
Authors: Siming Wang, Qing Ni, Yu Wu, Ruihai Xu, Hong Ye
Abstract:
Porous electric heaters, compared to conventional electric heaters, exhibit excellent heating performance due to their large specific surface area. Porous electric heaters employ porous metallic materials or conductive porous ceramics as the heating element. The former attains a low heating power with a fixed current due to the low electrical resistivity of metal. Although the latter can bypass the inherent challenges of porous metallic materials, the fabrication process of the conductive porous ceramics is complicated and high cost. This work proposed a porous ceramic electric heater with dielectric honeycomb ceramic as a substrate and surface conductive coating as a heating element. The conductive coating was prepared by the sol-gel method using silica sol and methyl trimethoxysilane as raw materials and graphite powder as conductive fillers. The conductive mechanism and degradation reason of the conductive coating was studied by electrical resistivity and thermal stability analysis. The heating performance of the proposed heater was experimentally investigated by heating air and deionized water. The results indicate that the electron transfer is achieved by forming the conductive network through the contact of the graphite flakes. With 30 wt% of graphite, the electrical resistivity of the conductive coating can be as low as 0.88 Ω∙cm. The conductive coating exhibits good electrical stability up to 500°C but degrades beyond 600°C due to the formation of many cracks in the coating caused by the weight loss and thermal expansion. The results also show that the working medium has a great influence on the volume power density of the heater. With air under natural convection as the working medium, the volume power density attains 640.85 kW/m3, which can be increased by 5 times when using deionized water as the working medium. The proposed honeycomb ceramic electric heater has the advantages of the simple fabrication method, low cost, and high volume power density, demonstrating great potential in the fluid heating field.Keywords: conductive coating, honeycomb ceramic electric heater, high specific surface area, high volume power density
Procedia PDF Downloads 1537816 High-Pressure Steam Turbine for Medium-Scale Concentrated Solar Power Plants
Authors: Ambra Giovannelli, Coriolano Salvini
Abstract:
Many efforts have been spent in the design and development of Concentrated Solar Power (CPS) Plants worldwide. Most of them are for on-grid electricity generation and they are large plants which can benefit from the economies of scale. Nevertheless, several potential applications for Small and Medium-Scale CSP plants can be relevant in the industrial sector as well as for off-grid purposes (i.e. in rural contexts). In a wide range of industrial processes, CSP technologies can be used for heat generation replacing conventional primary sources. For such market, proven technologies (usually hybrid solutions) already exist: more than 100 installations, especially in developing countries, are in operation and performance can be verified. On the other hand, concerning off-grid applications, solar technologies are not so mature. Even if the market offers a potential deployment of such systems, especially in countries where the access to grid is strongly limited, optimized solutions have not been developed yet. In this context, steam power plants can be taken into consideration for medium scale installations, due to the recent results achieved with direct steam generation systems based on paraboloidal dish or Fresnel lens solar concentrators. Steam at 4.0-4.5 MPa and 500°C can be produced directly by means of innovative solar receivers (some prototypes already exist). Although it could seem a promising technology, presently, steam turbines commercially available do not cover the required cycle specifications. In particular, while low-pressure turbines already exist on the market, high-pressure groups, necessary for the abovementioned applications, are not available. The present paper deals with the preliminary design of a high-pressure steam turbine group for a medium-scale CSP plant (200-1000 kWe). Such a group is arranged in a single geared package composed of four radial expander wheels. Such wheels have been chosen on the basis of automotive turbocharging technology and then modified to take the new requirements into account. Results related to the preliminary geometry selection and to the analysis of the high-pressure turbine group performance are reported and widely discussed.Keywords: concentrated solar power (CSP) plants, steam turbine, radial turbine, medium-scale power plants
Procedia PDF Downloads 2177815 Numerical Investigations of Unstable Pressure Fluctuations Behavior in a Side Channel Pump
Authors: Desmond Appiah, Fan Zhang, Shouqi Yuan, Wei Xueyuan, Stephen N. Asomani
Abstract:
The side channel pump has distinctive hydraulic performance characteristics over other vane pumps because of its generation of high pressure heads in only one impeller revolution. Hence, there is soaring utilization and application in the fields of petrochemical, food processing fields, automotive and aerospace fuel pumping where high heads are required at low flows. The side channel pump is characterized by unstable flow because after fluid flows into the impeller passage, it moves into the side channel and comes back to the impeller again and then moves to the next circulation. Consequently, the flow leaves the side channel pump following a helical path. However, the pressure fluctuation exhibited in the flow greatly contributes to the unwanted noise and vibration which is associated with the flow. In this paper, a side channel pump prototype was examined thoroughly through numerical calculations based on SST k-ω turbulence model to ascertain the pressure fluctuation behavior. The pressure fluctuation intensity of the 3D unstable flow dynamics were carefully investigated under different working conditions 0.8QBEP, 1.0 QBEP and 1.2QBEP. The results showed that the pressure fluctuation distribution around the pressure side of the blade is greater than the suction side at the impeller and side channel interface (z=0) for all three operating conditions. Part-load condition 0.8QBEP recorded the highest pressure fluctuation distribution because of the high circulation velocity thus causing an intense exchanged flow between the impeller and side channel. Time and frequency domains spectra of the pressure fluctuation patterns in the impeller and the side channel were also analyzed under the best efficiency point value, QBEP using the solution from the numerical calculations. It was observed from the time-domain analysis that the pressure fluctuation characteristics in the impeller flow passage increased steadily until the flow reached the interrupter which separates low-pressure at the inflow from high pressure at the outflow. The pressure fluctuation amplitudes in the frequency domain spectrum at the different monitoring points depicted a gentle decreasing trend of the pressure amplitudes which was common among the operating conditions. The frequency domain also revealed that the main excitation frequencies occurred at 600Hz, 1200Hz, and 1800Hz and continued in the integers of the rotating shaft frequency. Also, the mass flow exchange plots indicated that the side channel pump is characterized with many vortex flows. Operating conditions 0.8QBEP, 1.0 QBEP depicted less and similar vortex flow while 1.2Q recorded many vortex flows around the inflow, middle and outflow regions. The results of the numerical calculations were finally verified experimentally. The performance characteristics curves from the simulated results showed that 0.8QBEP working condition recorded a head increase of 43.03% and efficiency decrease of 6.73% compared to 1.0QBEP. It can be concluded that for industrial applications where the high heads are mostly required, the side channel pump can be designed to operate at part-load conditions. This paper can serve as a source of information in order to optimize a reliable performance and widen the applications of the side channel pumps.Keywords: exchanged flow, pressure fluctuation, numerical simulation, side channel pump
Procedia PDF Downloads 1367814 Effect of Sulphur Concentration on Microbial Population and Performance of a Methane Biofilter
Authors: Sonya Barzgar, J. Patrick, A. Hettiaratchi
Abstract:
Methane (CH4) is reputed as the second largest contributor to greenhouse effect with a global warming potential (GWP) of 34 related to carbon dioxide (CO2) over the 100-year horizon, so there is a growing interest in reducing the emissions of this gas. Methane biofiltration (MBF) is a cost effective technology for reducing low volume point source emissions of methane. In this technique, microbial oxidation of methane is carried out by methane-oxidizing bacteria (methanotrophs) which use methane as carbon and energy source. MBF uses a granular medium, such as soil or compost, to support the growth of methanotrophic bacteria responsible for converting methane to carbon dioxide (CO₂) and water (H₂O). Even though the biofiltration technique has been shown to be an efficient, practical and viable technology, the design and operational parameters, as well as the relevant microbial processes have not been investigated in depth. In particular, limited research has been done on the effects of sulphur on methane bio-oxidation. Since bacteria require a variety of nutrients for growth, to improve the performance of methane biofiltration, it is important to establish the input quantities of nutrients to be provided to the biofilter to ensure that nutrients are available to sustain the process. The study described in this paper was conducted with the aim of determining the influence of sulphur on methane elimination in a biofilter. In this study, a set of experimental measurements has been carried out to explore how the conversion of elemental sulphur could affect methane oxidation in terms of methanotrophs growth and system pH. Batch experiments with different concentrations of sulphur were performed while keeping the other parameters i.e. moisture content, methane concentration, oxygen level and also compost at their optimum level. The study revealed the tolerable limit of sulphur without any interference to the methane oxidation as well as the particular sulphur concentration leading to the greatest methane elimination capacity. Due to the sulphur oxidation, pH varies in a transient way which affects the microbial growth behavior. All methanotrophs are incapable of growth at pH values below 5.0 and thus apparently are unable to oxidize methane. Herein, the certain pH for the optimal growth of methanotrophic bacteria is obtained. Finally, monitoring methane concentration over time in the presence of sulphur is also presented for laboratory scale biofilters.Keywords: global warming, methane biofiltration (MBF), methane oxidation, methanotrophs, pH, sulphur
Procedia PDF Downloads 2367813 The Moderating Role of Payment Platform Applications’ Relationship with Increasing Purchase Intention Among Customers in Kuwait - Unified Theory of Acceptance and Sustainable Use of Technology Model
Authors: Ahmad Alsaber
Abstract:
This paper aims to understand the intermediary role of the payment platform applications by analyzing the various factors that can influence the desirability of utilizing said payment services in Kuwait, as well as to determine the effect of the presence of different types of payment platforms on the variables of the “Unified Theory of Acceptance and Use of Technology” (UTAUT) model. The UTAUT model's findings will provide an important understanding of the moderating role of payment platform mobile applications. This study will explore the influence of payment platform mobile applications on customer purchase intentions in Kuwait by employing a quantitative survey of 200 local customers. Questions will cover their usage of payment platforms, purchase intent, and overall satisfaction. The information gathered is then analyzed using descriptive statistics and correlation analysis in order to gain insights. The research hopes to provide greater insight into the effect of mobile payment platforms on customer purchase intentions in Kuwait. This research will provide important implications to marketers and customer service providers, informing their strategies and initiatives, as well as offer recommendations to payment platform providers on how to improve customer satisfaction and security. The study results suggest that the likelihood of a purchase is affected by performance expectancy, effort expectancy, social influence, risk, and trust. The purpose of this research is to understand the advancements in the different variables that Kuwaiti customers consider while dealing with mobile banking applications. With the implementation of stronger security measures, progressively more payment platform applications are being utilized in the Kuwaiti marketplace, making them more desirable with their accessibility and usability. With the development of the Kuwaiti digital economy, it is expected that mobile banking will have a greater impact on banking transactions and services in the future.Keywords: purchase intention, UTAUT, performance expectancy, social influence, risk, trust
Procedia PDF Downloads 1167812 Super-Hydrophilic TFC Membrane with High Stability in Oil
Authors: M. Obaid, Nasser A. M. Barakat, Fadali O.A
Abstract:
Low stability in oil media and the hydrophobicity problems of the ploysulfone electrospun membranes could be overcome in the present study. Synthesis of super-hydrophilic and highly stable in oil polysulfone electrospun nanofiber membrane was achieved by electrospinning of polysulfone solution containing NaOH salt followed by activation of the dried electrospun membrane by deposition of polyamide layer on the surface using m-phenylenediamine and 1,3,5-benzenetricarbonyl chloride. The introduced membrane has super-hydrophilicity characteristic (contact angle=3o), excellent stability in oil media and distinct performance in oil-water separation process.Keywords: electrospinning, oil-degradability, membrane, nanofibers
Procedia PDF Downloads 4817811 A Study on Reinforced Concrete Beams Enlarged with Polymer Mortar and UHPFRC
Authors: Ga Ye Kim, Hee Sun Kim, Yeong Soo Shin
Abstract:
Many studies have been done on the repair and strengthening method of concrete structure, so far. The traditional retrofit method was to attach fiber sheet such as CFRP (Carbon Fiber Reinforced Polymer), GFRP (Glass Fiber Reinforced Polymer) and AFRP (Aramid Fiber Reinforced Polymer) on the concrete structure. However, this method had many downsides in that there are a risk of debonding and an increase in displacement by a shortage of structure section. Therefore, it is effective way to enlarge the structural member with polymer mortar or Ultra-High Performance Fiber Reinforced Concrete (UHPFRC) as a means of strengthening concrete structure. This paper intends to investigate structural performance of reinforced concrete (RC) beams enlarged with polymer mortar and compare the experimental results with analytical results. Nonlinear finite element analyses were conducted to compare the experimental results and predict structural behavior of retrofitted RC beams accurately without cost consuming experimental process. In addition, this study aims at comparing differences of retrofit material between commonly used material (polymer mortar) and recently used material (UHPFRC) by conducting nonlinear finite element analyses. In the first part of this paper, the RC beams having different cover type were fabricated for the experiment and the size of RC beams was 250 millimeters in depth, 150 millimeters in width and 2800 millimeters in length. To verify the experiment, nonlinear finite element models were generated using commercial software ABAQUS 6.10-3. From this study, both experimental and analytical results demonstrated good strengthening effect on RC beam and showed similar tendency. For the future, the proposed analytical method can be used to predict the effect of strengthened RC beam. In the second part of the study, the main parameters were type of retrofit materials. The same nonlinear finite element models were generated to compare the polymer mortar with UHPFRCC. Two types of retrofit material were evaluated and retrofit effect was verified by analytical results.Keywords: retrofit material, polymer mortar, UHPFRC, nonlinear finite element analysis
Procedia PDF Downloads 4187810 Permeable Asphalt Pavement as a Measure of Urban Green Infrastructure in the Extreme Events Mitigation
Authors: Márcia Afonso, Cristina Fael, Marisa Dinis-Almeida
Abstract:
Population growth in cities has led to an increase in the infrastructures construction, including buildings and roadways. This aspect leads directly to the soils waterproofing. In turn, changes in precipitation patterns are developing into higher and more frequent intensities. Thus, these two conjugated aspects decrease the rainwater infiltration into soils and increase the volume of surface runoff. The practice of green and sustainable urban solutions has encouraged research in these areas. The porous asphalt pavement, as a green infrastructure, is part of practical solutions set to address urban challenges related to land use and adaptation to climate change. In this field, permeable pavements with porous asphalt mixtures (PA) have several advantages in terms of reducing the runoff generated by the floods. The porous structure of these pavements, compared to a conventional asphalt pavement, allows the rainwater infiltration in the subsoil, and consequently, the water quality improvement. This green infrastructure solution can be applied in cities, particularly in streets or parking lots to mitigate the floods effects. Over the years, the pores of these pavements can be filled by sediment, reducing their function in the rainwater infiltration. Thus, double layer porous asphalt (DLPA) was developed to mitigate the clogging effect and facilitate the water infiltration into the lower layers. This study intends to deepen the knowledge of the performance of DLPA when subjected to clogging. The experimental methodology consisted on four evaluation phases of the DLPA infiltration capacity submitted to three precipitation events (100, 200 and 300 mm/h) in each phase. The evaluation first phase determined the behavior after DLPA construction. In phases two and three, two 500 g/m2 clogging cycles were performed, totaling a 1000 g/m2 final simulation. Sand with gradation accented in fine particles was used as clogging material. In the last phase, the DLPA was subjected to simple sweeping and vacuuming maintenance. A precipitation simulator, type sprinkler, capable of simulating the real precipitation was developed for this purpose. The main conclusions show that the DLPA has the capacity to drain the water, even after two clogging cycles. The infiltration results of flows lead to an efficient performance of the DPLA in the surface runoff attenuation, since this was not observed in any of the evaluation phases, even at intensities of 200 and 300 mm/h, simulating intense precipitation events. The infiltration capacity under clogging conditions decreased about 7% on average in the three intensities relative to the initial performance that is after construction. However, this was restored when subjected to simple maintenance, recovering the DLPA hydraulic functionality. In summary, the study proved the efficacy of using a DLPA when it retains thicker surface sediments and limits the fine sediments entry to the remaining layers. At the same time, it is guaranteed the rainwater infiltration and the surface runoff reduction and is therefore a viable solution to put into practice in permeable pavements.Keywords: clogging, double layer porous asphalt, infiltration capacity, rainfall intensity
Procedia PDF Downloads 4917809 Early Diagnosis of Myocardial Ischemia Based on Support Vector Machine and Gaussian Mixture Model by Using Features of ECG Recordings
Authors: Merve Begum Terzi, Orhan Arikan, Adnan Abaci, Mustafa Candemir
Abstract:
Acute myocardial infarction is a major cause of death in the world. Therefore, its fast and reliable diagnosis is a major clinical need. ECG is the most important diagnostic methodology which is used to make decisions about the management of the cardiovascular diseases. In patients with acute myocardial ischemia, temporary chest pains together with changes in ST segment and T wave of ECG occur shortly before the start of myocardial infarction. In this study, a technique which detects changes in ST/T sections of ECG is developed for the early diagnosis of acute myocardial ischemia. For this purpose, a database of real ECG recordings that contains a set of records from 75 patients presenting symptoms of chest pain who underwent elective percutaneous coronary intervention (PCI) is constituted. 12-lead ECG’s of the patients were recorded before and during the PCI procedure. Two ECG epochs, which are the pre-inflation ECG which is acquired before any catheter insertion and the occlusion ECG which is acquired during balloon inflation, are analyzed for each patient. By using pre-inflation and occlusion recordings, ECG features that are critical in the detection of acute myocardial ischemia are identified and the most discriminative features for the detection of acute myocardial ischemia are extracted. A classification technique based on support vector machine (SVM) approach operating with linear and radial basis function (RBF) kernels to detect ischemic events by using ST-T derived joint features from non-ischemic and ischemic states of the patients is developed. The dataset is randomly divided into training and testing sets and the training set is used to optimize SVM hyperparameters by using grid-search method and 10fold cross-validation. SVMs are designed specifically for each patient by tuning the kernel parameters in order to obtain the optimal classification performance results. As a result of implementing the developed classification technique to real ECG recordings, it is shown that the proposed technique provides highly reliable detections of the anomalies in ECG signals. Furthermore, to develop a detection technique that can be used in the absence of ECG recording obtained during healthy stage, the detection of acute myocardial ischemia based on ECG recordings of the patients obtained during ischemia is also investigated. For this purpose, a Gaussian mixture model (GMM) is used to represent the joint pdf of the most discriminating ECG features of myocardial ischemia. Then, a Neyman-Pearson type of approach is developed to provide detection of outliers that would correspond to acute myocardial ischemia. Neyman – Pearson decision strategy is used by computing the average log likelihood values of ECG segments and comparing them with a range of different threshold values. For different discrimination threshold values and number of ECG segments, probability of detection and probability of false alarm values are computed, and the corresponding ROC curves are obtained. The results indicate that increasing number of ECG segments provide higher performance for GMM based classification. Moreover, the comparison between the performances of SVM and GMM based classification showed that SVM provides higher classification performance results over ECG recordings of considerable number of patients.Keywords: ECG classification, Gaussian mixture model, Neyman–Pearson approach, support vector machine
Procedia PDF Downloads 1627808 The Development of Assessment Criteria Framework for Sustainable Healthcare Buildings in China
Authors: Chenyao Shen, Jie Shen
Abstract:
The rating system provides an effective framework for assessing building environmental performance and integrating sustainable development into building and construction processes; as it can be used as a design tool by developing appropriate sustainable design strategies and determining performance measures to guide the sustainable design and decision-making processes. Healthcare buildings are resource (water, energy, etc.) intensive. To maintain high-cost operations and complex medical facilities, they require a great deal of hazardous and non-hazardous materials, stringent control of environmental parameters, and are responsible for producing polluting emission. Compared with other types of buildings, the impact of healthcare buildings on the full cycle of the environment is particularly large. With broad recognition among designers and operators that energy use can be reduced substantially, many countries have set up their own green rating systems for healthcare buildings. There are four main green healthcare building evaluation systems widely acknowledged in the world - Green Guide for Health Care (GGHC), which was jointly organized by the United States HCWH and CMPBS in 2003; BREEAM Healthcare, issued by the British Academy of Building Research (BRE) in 2008; the Green Star-Healthcare v1 tool, released by the Green Building Council of Australia (GBCA) in 2009; and LEED Healthcare 2009, released by the United States Green Building Council (USGBC) in 2011. In addition, the German Association of Sustainable Building (DGNB) has also been developing the German Sustainable Building Evaluation Criteria (DGNB HC). In China, more and more scholars and policy makers have recognized the importance of assessment of sustainable development, and have adapted some tools and frameworks. China’s first comprehensive assessment standard for green building (the GBTs) was issued in 2006 (lately updated in 2014), promoting sustainability in the built-environment and raise awareness of environmental issues among architects, engineers, contractors as well as the public. However, healthcare building was not involved in the evaluation system of GBTs because of its complex medical procedures, strict requirements of indoor/outdoor environment and energy consumption of various functional rooms. Learn from advanced experience of GGHC, BREEAM, and LEED HC above, China’s first assessment criteria for green hospital/healthcare buildings was finally released in December 2015. Combined with both quantitative and qualitative assessment criteria, the standard highlight the differences between healthcare and other public buildings in meeting the functional needs for medical facilities and special groups. This paper has focused on the assessment criteria framework for sustainable healthcare buildings, for which the comparison of different rating systems is rather essential. Descriptive analysis is conducted together with the cross-matrix analysis to reveal rich information on green assessment criteria in a coherent manner. The research intends to know whether the green elements for healthcare buildings in China are different from those conducted in other countries, and how to improve its assessment criteria framework.Keywords: assessment criteria framework, green building design, healthcare building, building performance rating tool
Procedia PDF Downloads 1467807 Confidence Envelopes for Parametric Model Selection Inference and Post-Model Selection Inference
Authors: I. M. L. Nadeesha Jayaweera, Adao Alex Trindade
Abstract:
In choosing a candidate model in likelihood-based modeling via an information criterion, the practitioner is often faced with the difficult task of deciding just how far up the ranked list to look. Motivated by this pragmatic necessity, we construct an uncertainty band for a generalized (model selection) information criterion (GIC), defined as a criterion for which the limit in probability is identical to that of the normalized log-likelihood. This includes common special cases such as AIC & BIC. The method starts from the asymptotic normality of the GIC for the joint distribution of the candidate models in an independent and identically distributed (IID) data framework and proceeds by deriving the (asymptotically) exact distribution of the minimum. The calculation of an upper quantile for its distribution then involves the computation of multivariate Gaussian integrals, which is amenable to efficient implementation via the R package "mvtnorm". The performance of the methodology is tested on simulated data by checking the coverage probability of nominal upper quantiles and compared to the bootstrap. Both methods give coverages close to nominal for large samples, but the bootstrap is two orders of magnitude slower. The methodology is subsequently extended to two other commonly used model structures: regression and time series. In the regression case, we derive the corresponding asymptotically exact distribution of the minimum GIC invoking Lindeberg-Feller type conditions for triangular arrays and are thus able to similarly calculate upper quantiles for its distribution via multivariate Gaussian integration. The bootstrap once again provides a default competing procedure, and we find that similar comparison performance metrics hold as for the IID case. The time series case is complicated by far more intricate asymptotic regime for the joint distribution of the model GIC statistics. Under a Gaussian likelihood, the default in most packages, one needs to derive the limiting distribution of a normalized quadratic form for a realization from a stationary series. Under conditions on the process satisfied by ARMA models, a multivariate normal limit is once again achieved. The bootstrap can, however, be employed for its computation, whence we are once again in the multivariate Gaussian integration paradigm for upper quantile evaluation. Comparisons of this bootstrap-aided semi-exact method with the full-blown bootstrap once again reveal a similar performance but faster computation speeds. One of the most difficult problems in contemporary statistical methodological research is to be able to account for the extra variability introduced by model selection uncertainty, the so-called post-model selection inference (PMSI). We explore ways in which the GIC uncertainty band can be inverted to make inferences on the parameters. This is being attempted in the IID case by pivoting the CDF of the asymptotically exact distribution of the minimum GIC. For inference one parameter at a time and a small number of candidate models, this works well, whence the attained PMSI confidence intervals are wider than the MLE-based Wald, as expected.Keywords: model selection inference, generalized information criteria, post model selection, Asymptotic Theory
Procedia PDF Downloads 867806 Perfomance of PAPR Reduction in OFDM System for Wireless Communications
Authors: Alcardo Alex Barakabitze, Saddam Aziz, Muhammad Zubair
Abstract:
The Orthogonal Frequency Division Multiplexing (OFDM) is a special form of multicarrier transmission that splits the total transmission bandwidth into a number of orthogonal and non-overlapping subcarriers and transmit the collection of bits called symbols in parallel using these subcarriers. In this paper, we explore the Peak to Average Power Reduction (PAPR) problem in OFDM systems. We provide the performance analysis of CCDF and BER through MATLAB simulations.Keywords: bit error ratio (BER), OFDM, peak to average power reduction (PAPR), sub-carriers
Procedia PDF Downloads 5427805 Investigation of the Working Processes in Thermocompressor Operating on Cryogenic Working Fluid
Authors: Evgeny V. Blagin, Aleksandr I. Dovgjallo, Dmitry A. Uglanov
Abstract:
This article deals with research of the working process in the thermocompressor which operates on cryogenic working fluid. Thermocompressor is device suited for the conversation of heat energy directly to the potential energy of pressure. Suggested thermocompressor is suited for operation during liquid natural gas (LNG) re-gasification and is placed after evaporator. Such application of thermocompressor allows using of the LNG cold energy for rising of working fluid pressure, which then can be used for electricity generation or another purpose. Thermocompressor consists of two chambers divided by the regenerative heat exchanger. Calculation algorithm for unsteady calculation of thermocompressor working process was suggested. The results of this investigation are to change of thermocompressor’s chambers temperature and pressure during the working cycle. These distributions help to find out the parameters, which significantly influence thermocompressor efficiency. These parameters include regenerative heat exchanger coefficient of the performance (COP) dead volume of the chambers, working frequency of the thermocompressor etc. Exergy analysis was performed to estimate thermocompressor efficiency. Cryogenic thermocompressor operated on nitrogen working fluid was chosen as a prototype. Calculation of the temperature and pressure change was performed with taking into account heat fluxes through regenerator and thermocompressor walls. Temperature of the cold chamber significantly differs from the results of steady calculation, which is caused by friction of the working fluid in regenerator and heat fluxes from the hot chamber. The rise of the cold chamber temperature leads to decreasing of thermocompressor delivery volume. Temperature of hot chamber differs negligibly because losses due to heat fluxes to a cold chamber are compensated by the friction of the working fluid in the regenerator. Optimal working frequency was selected. Main results of the investigation: -theoretical confirmation of thermocompressor operation capability on the cryogenic working fluid; -optimal working frequency was found; -value of the cold chamber temperature differs from the starting value much more than the temperature of the hot chamber; -main parameters which influence thermocompressor performance are regenerative heat exchanger COP and heat fluxes through regenerator and thermocompressor walls.Keywords: cold energy, liquid natural gas, thermocompressor, regenerative heat exchanger
Procedia PDF Downloads 5827804 Modern Methods of Construction (MMC): The Potentials and Challenges of Using Prefabrication Technology for Building Modern Houses in Afghanistan
Authors: Latif Karimi, Yasuhide Mochida
Abstract:
The purpose of this paper is to study Modern Methods of Construction (MMC); specifically, the prefabrication technology and check the applicability, suitability, and benefits of this construction technique over conventional methods for building new houses in Afghanistan. Construction industry and house building sector are a key contributor to Afghanistan’s economy. However, this sector is challenged with lack of innovation and severe impacts that it has on the environment due to huge amount of construction waste from building, demolition and or renovation activities. This paper studies the prefabrication technology, a popular MMC that is becoming more common, improving in quality and being available in a variety of budgets. Several feasibility studies worldwide have revealed that this method is the way forward in improving construction industry performance as it has been proven to reduce construction time, construction wastes and improve the environmental performance of the construction processes. In addition, this study emphasizes on 'sustainability' in-house building, since it is a common challenge in housing construction projects on a global scale. This challenge becomes more severe in the case of under-developed countries, like Afghanistan. Because, most of the houses are being built in the absence of a serious quality control mechanism and dismissive to basic requirements of sustainable houses; well-being, cost-effectiveness, minimization - prevention of wastes production during construction and use, and severe environmental impacts in view of a life cycle assessment. Methodology: A literature review and study of the conventional practices of building houses in urban areas of Afghanistan. A survey is also being completed to study the potentials and challenges of using prefabrication technology for building modern houses in the cities across the country. A residential housing project is selected for case study to determine the drawbacks of current construction methods vs. prefabrication technique for building a new house. Originality: There are little previous research available about MMC considering its specific impacts on sustainability related to house building practices. This study will be specifically of interest to a broad range of people, including planners, construction managers, builders, and house owners.Keywords: modern methods of construction (MMC), prefabrication, prefab houses, sustainable construction, modern houses
Procedia PDF Downloads 2437803 A Shift-Share Analysis: Manufacturing Employment Specialisation at uMhlathuze Local Municipality, South Africa
Authors: Mlondi Ndovela
Abstract:
Globally, the manufacturing employment has been declining and the South African manufacturing sector experiences the very same trend. Despite the commonality between the global and South African manufacturing trend, there is an understanding that local areas provide distinct contributions to the provincial/national economy. Therefore, the growth/decline of a particular manufacturing division in one local area may not be evident in another area since economic performances vary from region to region. In view of the above, the study employed the Esteban-Marquillas model of shift-share analysis (SSA) to conduct an empirical analysis of manufacturing employment performance at uMhlathuze Local Municipality in the KwaZulu-Natal province. The study set out two objectives; those are, to quantify uMhlathuze manufacturing jobs that are attributed to the provincial manufacturing employment trends and identify manufacturing divisions are growing/declining in terms of employment. To achieve these objectives, the study sampled manufacturing employment data from 2010 to 2017 and this data was categorised into ten manufacturing divisions. Furthermore, the Esteban-Marquillas model calculated manufacturing employment in terms of two effects, namely; provincial growth effect (PGE) and industrial mix effect (IME). The results show that even though uMhlathuze manufacturing sector has a positive PGE (+230), the municipality performed poorly in terms of IME (-291). A further analysis included other economic sectors of the municipality to draw employment performance comparison and the study found that agriculture; construction; trade, catering and accommodation; and transport, storage and communication, performed well above manufacturing sector in terms of PGE (+826) and IME (+532). This suggests that uMhlathuze manufacturing sector is not necessarily declining; however, other economic sectors are growing faster and bigger than it is, therefore, reducing the employment share of the manufacturing sector. To promote manufacturing growth from a policy standpoint, the government could create favourable macroeconomic policies such as import substitution policies and support labour-intensive manufacturing divisions. As a result, these macroeconomic policies can help to protect local manufacturing firms and stimulate the growth of manufacturing employment.Keywords: allocation effect, Esteban-Marquillas model, manufacturing employment, regional competitive effect, shift-share analysis
Procedia PDF Downloads 1417802 A Safety-Door for Earthquake Disaster Prevention - Part II
Authors: Daniel Y. Abebe, Jaehyouk Choi
Abstract:
The safety of door has not given much attention. The main problem of doors during and after earthquake is that they are unable to be opened because deviation from its original position by the lateral load. The aim of this research is to develop and evaluate a safety door that keeps the door frame in its original position or keeps its edge angles perpendicular during and post-earthquake. Nonlinear finite element analysis was conducted in order to evaluate the structural performance and behavior of the proposed door under both monotonic and cyclic loading.Keywords: safety-door, earthquake disaster, low yield point steel, passive energy dissipating device, FE analysis
Procedia PDF Downloads 4737801 Experimental Studies on the Effect of Premixing Methods in Anaerobic Digestor with Corn Stover
Authors: M. Sagarika, M. Chandra Sekhar
Abstract:
Agricultural residues are producing in large quantities in India and account for abundant but underutilized source of renewable biomass in agriculture. In India, the amount of crop residues available is estimated to be approximately 686 million tons. Anaerobic digestion is a promising option to utilize the surplus agricultural residues and can produce biogas and digestate. Biogas is mainly methane (CH4), which can be utilized as an energy source in replacement for fossil fuels such as natural gas, oil, in other hand, digestate contains high amounts of nutrients, can be employed as fertilizer. Solid state anaerobic digestion (total solids ≥ 15%) is suitable for agricultural residues, as it reduces the problems like stratification and floating issues that occur in liquid anaerobic digestion (total solids < 15%). The major concern in solid-state anaerobic digestion is the low mass transfer of feedstock and inoculum that resulting in low performance. To resolve this low mass transfer issue, effective mixing of feedstock and inoculum is required. Mechanical mixing using stirrer at the time of digestion process can be done, but it is difficult to operate the stirring of feedstock with high solids percentage and high viscosity. Complete premixing of feedstock and inoculum is an alternative method, which is usual in lab scale studies but may not be affordable due to high energy demand in large-scale digesters. Developing partial premixing methods may reduce this problem. Current study is to improve the performance of solid-state anaerobic digestion of corn stover at feedstock to inoculum ratios 3 and 5, by applying partial premixing methods and to compare the complete premixing method with two partial premixing methods which are two alternative layers of feedstock and inoculum and three alternative layers of feedstock and inoculum where higher inoculum ratios in the top layers. From experimental studies it is observed that, partial premixing method with three alternative layers of feedstock and inoculum yielded good methane.Keywords: anaerobic digestion, premixing methods, methane yield, corn stover, volatile solids
Procedia PDF Downloads 2347800 Multi-Objective Discrete Optimization of External Thermal Insulation Composite Systems in Terms of Thermal and Embodied Energy Performance
Authors: Berfin Yildiz
Abstract:
These days, increasing global warming effects, limited amount of energy resources, etc., necessitates the awareness that must be present in every profession group. The architecture and construction sectors are responsible for both the embodied and operational energy of the materials. This responsibility has led designers to seek alternative solutions for energy-efficient material selection. The choice of energy-efficient material requires consideration of the entire life cycle, including the building's production, use, and disposal energy. The aim of this study is to investigate the method of material selection of external thermal insulation composite systems (ETICS). Embodied and in-use energy values of material alternatives were used for the evaluation in this study. The operational energy is calculated according to the u-value calculation method defined in the TS 825 (Thermal Insulation Requirements) standard for Turkey, and the embodied energy is calculated based on the manufacturer's Energy Performance Declaration (EPD). ETICS consists of a wall, adhesive, insulation, lining, mechanical, mesh, and exterior finishing materials. In this study, lining, mechanical, and mesh materials were ignored because EPD documents could not be obtained. The material selection problem is designed as a hypothetical volume area (5x5x3m) and defined as a multi-objective discrete optimization problem for external thermal insulation composite systems. Defining the problem as a discrete optimization problem is important in order to choose between materials of various thicknesses and sizes. Since production and use energy values, which are determined as optimization objectives in the study, are often conflicting values, material selection is defined as a multi-objective optimization problem, and it is aimed to obtain many solution alternatives by using Hypervolume (HypE) algorithm. The enrollment process started with 100 individuals and continued for 50 generations. According to the obtained results, it was observed that autoclaved aerated concrete and Ponce block as wall material, glass wool, as insulation material gave better results.Keywords: embodied energy, multi-objective discrete optimization, performative design, thermal insulation
Procedia PDF Downloads 1417799 Simulation of Channel Models for Device-to-Device Application of 5G Urban Microcell Scenario
Authors: H. Zormati, J. Chebil, J. Bel Hadj Tahar
Abstract:
Next generation wireless transmission technology (5G) is expected to support the development of channel models for higher frequency bands, so clarification of high frequency bands is the most important issue in radio propagation research for 5G, multiple urban microcellular measurements have been carried out at 60 GHz. In this paper, the collected data is uniformly analyzed with focus on the path loss (PL), the objective is to compare simulation results of some studied channel models with the purpose of testing the performance of each one.Keywords: 5G, channel model, 60GHz channel, millimeter-wave, urban microcell
Procedia PDF Downloads 3197798 Loading Methodology for a Capacity Constrained Job-Shop
Authors: Viraj Tyagi, Ajai Jain, P. K. Jain, Aarushi Jain
Abstract:
This paper presents a genetic algorithm based loading methodology for a capacity constrained job-shop with the consideration of alternative process plans for each part to be produced. Performance analysis of the proposed methodology is carried out for two case studies by considering two different manufacturing scenarios. Results obtained indicate that the methodology is quite effective in improving the shop load balance, and hence, it can be included in the frameworks of manufacturing planning systems of job-shop oriented industries.Keywords: manufacturing planning, loading, genetic algorithm, job shop
Procedia PDF Downloads 3007797 Digital Twin for University Campus: Workflow, Applications and Benefits
Authors: Frederico Fialho Teixeira, Islam Mashaly, Maryam Shafiei, Jurij Karlovsek
Abstract:
The ubiquity of data gathering and smart technologies, advancements in virtual technologies, and the development of the internet of things (IoT) have created urgent demands for the development of frameworks and efficient workflows for data collection, visualisation, and analysis. Digital twin, in different scales of the city into the building, allows for bringing together data from different sources to generate fundamental and illuminating insights for the management of current facilities and the lifecycle of amenities as well as improvement of the performance of current and future designs. Over the past two decades, there has been growing interest in the topic of digital twin and their applications in city and building scales. Most such studies look at the urban environment through a homogeneous or generalist lens and lack specificity in particular characteristics or identities, which define an urban university campus. Bridging this knowledge gap, this paper offers a framework for developing a digital twin for a university campus that, with some modifications, could provide insights for any large-scale digital twin settings like towns and cities. It showcases how currently unused data could be purposefully combined, interpolated and visualised for producing analysis-ready data (such as flood or energy simulations or functional and occupancy maps), highlighting the potential applications of such a framework for campus planning and policymaking. The research integrates campus-level data layers into one spatial information repository and casts light on critical data clusters for the digital twin at the campus level. The paper also seeks to raise insightful and directive questions on how digital twin for campus can be extrapolated to city-scale digital twin. The outcomes of the paper, thus, inform future projects for the development of large-scale digital twin as well as urban and architectural researchers on potential applications of digital twin in future design, management, and sustainable planning, to predict problems, calculate risks, decrease management costs, and improve performance.Keywords: digital twin, smart campus, framework, data collection, point cloud
Procedia PDF Downloads 687796 Dissolution of South African Limestone for Wet Flue Gas Desulphurization
Authors: Lawrence Koech, Ray Everson, Hein Neomagus, Hilary Rutto
Abstract:
Wet Flue gas desulphurization (FGD) systems are commonly used to remove sulphur dioxide from flue gas by contacting it with limestone in aqueous phase which is obtained by dissolution. Dissolution is important as it affects the overall performance of a wet FGD system. In the present study, effects of pH, stirring speed, solid to liquid ratio and acid concentration on the dissolution of limestone using an organic acid (adipic acid) were investigated. This was investigated using the pH stat apparatus. Calcium ions were analyzed at the end of each experiment using Atomic Absorption (AAS) machine.Keywords: desulphurization, limestone, dissolution, pH stat apparatus
Procedia PDF Downloads 4617795 An Efficient Process Analysis and Control Method for Tire Mixing Operation
Authors: Hwang Ho Kim, Do Gyun Kim, Jin Young Choi, Sang Chul Park
Abstract:
Since tire production process is very complicated, company-wide management of it is very difficult, necessitating considerable amounts of capital and labors. Thus, productivity should be enhanced and maintained competitive by developing and applying effective production plans. Among major processes for tire manufacturing, consisting of mixing component preparation, building and curing, the mixing process is an essential and important step because the main component of tire, called compound, is formed at this step. Compound as a rubber synthesis with various characteristics plays its own role required for a tire as a finished product. Meanwhile, scheduling tire mixing process is similar to flexible job shop scheduling problem (FJSSP) because various kinds of compounds have their unique orders of operations, and a set of alternative machines can be used to process each operation. In addition, setup time required for different operations may differ due to alteration of additives. In other words, each operation of mixing processes requires different setup time depending on the previous one, and this kind of feature, called sequence dependent setup time (SDST), is a very important issue in traditional scheduling problems such as flexible job shop scheduling problems. However, despite of its importance, there exist few research works dealing with the tire mixing process. Thus, in this paper, we consider the scheduling problem for tire mixing process and suggest an efficient particle swarm optimization (PSO) algorithm to minimize the makespan for completing all the required jobs belonging to the process. Specifically, we design a particle encoding scheme for the considered scheduling problem, including a processing sequence for compounds and machine allocation information for each job operation, and a method for generating a tire mixing schedule from a given particle. At each iteration, the coordination and velocity of particles are updated, and the current solution is compared with new solution. This procedure is repeated until a stopping condition is satisfied. The performance of the proposed algorithm is validated through a numerical experiment by using some small-sized problem instances expressing the tire mixing process. Furthermore, we compare the solution of the proposed algorithm with it obtained by solving a mixed integer linear programming (MILP) model developed in previous research work. As for performance measure, we define an error rate which can evaluate the difference between two solutions. As a result, we show that PSO algorithm proposed in this paper outperforms MILP model with respect to the effectiveness and efficiency. As the direction for future work, we plan to consider scheduling problems in other processes such as building, curing. We can also extend our current work by considering other performance measures such as weighted makespan or processing times affected by aging or learning effects.Keywords: compound, error rate, flexible job shop scheduling problem, makespan, particle encoding scheme, particle swarm optimization, sequence dependent setup time, tire mixing process
Procedia PDF Downloads 2657794 Effects of Robot-Assisted Hand Training on Upper Extremity Performance in Patients with Stroke: A Randomized Crossover Controlled, Assessor-Blinded Study
Authors: Hsin-Chieh Lee, Fen-Ling Kuo, Jui-Chi Lin
Abstract:
Background: Upper extremity functional impairment that occurs after stroke includes hemiplegia, synergy movement, muscle hypertonicity, and somatosensory impairment, which result in inefficient and inaccurate movement. Robot-assisted rehabilitation is an intensive training approach that is effective in sensorimotor and hand function recovery. However, these systems mostly focused on the proximal part of the upper limb rather than the distal part. The device used in our study was Gloreha Sinfonia, which focuses on the distal part of the upper limb and uses a dynamic support system to facilitate the whole limb function. The objective of this study was to investigate the effects of robot-assisted therapy (RT) with Gloreha device on sensorimotor, and ADLs in patients with stroke. Method: Patients with stroke (N=25) participated AB or BA (A = 12 RT sessions and B = 12 conventional therapy (CT) sessions) for 6 weeks (60 min at each session, twice a week), with 1-month break for washout period. The performance of the patients was assessed by a blinded assessor at 4 time points (pretest 1, posttest 1, pretest 2, posttest 2) which including the Fugl–Meyer Assessment-upper extremity (FMA-UE), box and block test, electromyography of the extensor digitorum communis (EDC) and brachioradialis, a grip dynamometer for motor evaluation; Semmes–Weinstein hand monofilament and Revision of the Nottingham Sensory Assessment for sensory evaluation; and the Modified Barthel Index (MBI) for assessing the ADL ability. Result: RT group significantly improved FMA-UE proximal scores (p = 0.038), FMA-UE total scores (p = 0.046), and MBI (p = 0.030). The EDC exhibited higher efficiency during the small block grasping task in the RT group than in the CT group (p = 0.050). Conclusions: RT with the Gloreha device might lead to beneficial effects on arm motor function, ADL ability, and EDC muscle recruitment efficacy in patients with subacute to chronic stroke.Keywords: activities of daily living, hand function, robotic rehabilitation, stroke
Procedia PDF Downloads 1187793 The Collaboration between Resident and Non-resident Patent Applicants as a Strategy to Accelerate Technological Advance in Developing Nations
Authors: Hugo Rodríguez
Abstract:
Migrations of researchers, scientists, and inventors are a widespread phenomenon in modern times. In some cases, migrants stay linked to research groups in their countries of origin, either out of their own conviction or because of government policies. We examine different linear models of technological development (using the Ordinary Least Squares (OLS) technique) in eight selected countries and find that the collaborations between resident and nonresident patent applicants correlate with different levels of performance of the technological policies in three different scenarios. Therefore, the reinforcement of that link must be considered a powerful tool for technological development.Keywords: development, collaboration, patents, technology
Procedia PDF Downloads 1277792 A Dataset of Program Educational Objectives Mapped to ABET Outcomes: Data Cleansing, Exploratory Data Analysis and Modeling
Authors: Addin Osman, Anwar Ali Yahya, Mohammed Basit Kamal
Abstract:
Datasets or collections are becoming important assets by themselves and now they can be accepted as a primary intellectual output of a research. The quality and usage of the datasets depend mainly on the context under which they have been collected, processed, analyzed, validated, and interpreted. This paper aims to present a collection of program educational objectives mapped to student’s outcomes collected from self-study reports prepared by 32 engineering programs accredited by ABET. The manual mapping (classification) of this data is a notoriously tedious, time consuming process. In addition, it requires experts in the area, which are mostly not available. It has been shown the operational settings under which the collection has been produced. The collection has been cleansed, preprocessed, some features have been selected and preliminary exploratory data analysis has been performed so as to illustrate the properties and usefulness of the collection. At the end, the collection has been benchmarked using nine of the most widely used supervised multiclass classification techniques (Binary Relevance, Label Powerset, Classifier Chains, Pruned Sets, Random k-label sets, Ensemble of Classifier Chains, Ensemble of Pruned Sets, Multi-Label k-Nearest Neighbors and Back-Propagation Multi-Label Learning). The techniques have been compared to each other using five well-known measurements (Accuracy, Hamming Loss, Micro-F, Macro-F, and Macro-F). The Ensemble of Classifier Chains and Ensemble of Pruned Sets have achieved encouraging performance compared to other experimented multi-label classification methods. The Classifier Chains method has shown the worst performance. To recap, the benchmark has achieved promising results by utilizing preliminary exploratory data analysis performed on the collection, proposing new trends for research and providing a baseline for future studies.Keywords: ABET, accreditation, benchmark collection, machine learning, program educational objectives, student outcomes, supervised multi-class classification, text mining
Procedia PDF Downloads 172