Search results for: actual exam time usage
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 20404

Search results for: actual exam time usage

19654 Real-Time Visualization Using GPU-Accelerated Filtering of LiDAR Data

Authors: Sašo Pečnik, Borut Žalik

Abstract:

This paper presents a real-time visualization technique and filtering of classified LiDAR point clouds. The visualization is capable of displaying filtered information organized in layers by the classification attribute saved within LiDAR data sets. We explain the used data structure and data management, which enables real-time presentation of layered LiDAR data. Real-time visualization is achieved with LOD optimization based on the distance from the observer without loss of quality. The filtering process is done in two steps and is entirely executed on the GPU and implemented using programmable shaders.

Keywords: filtering, graphics, level-of-details, LiDAR, real-time visualization

Procedia PDF Downloads 285
19653 Construction Time - Cost Trade-Off Analysis Using Fuzzy Set Theory

Authors: V. S. S. Kumar, B. Vikram, G. C. S. Reddy

Abstract:

Time and cost are the two critical objectives of construction project management and are not independent but intricately related. Trade-off between project duration and cost are extensively discussed during project scheduling because of practical relevance. Generally when the project duration is compressed, the project calls for an increase in labor and more productive equipments, which increases the cost. Thus, the construction time-cost optimization is defined as a process to identify suitable construction activities for speeding up to attain the best possible savings in both time and cost. As there is hidden tradeoff relationship between project time and cost, it might be difficult to predict whether the total cost would increase or decrease as a result of compressing the schedule. Different combinations of duration and cost for the activities associated with the project determine the best set in the time-cost optimization. Therefore, the contractors need to select the best combination of time and cost to perform each activity, all of which will ultimately determine the project duration and cost. In this paper, the fuzzy set theory is used to model the uncertainties in the project environment for time-cost trade off analysis.

Keywords: fuzzy sets, uncertainty, qualitative factors, decision making

Procedia PDF Downloads 634
19652 Ice Load Measurements on Known Structures Using Image Processing Methods

Authors: Azam Fazelpour, Saeed R. Dehghani, Vlastimil Masek, Yuri S. Muzychka

Abstract:

This study employs a method based on image analyses and structure information to detect accumulated ice on known structures. The icing of marine vessels and offshore structures causes significant reductions in their efficiency and creates unsafe working conditions. Image processing methods are used to measure ice loads automatically. Most image processing methods are developed based on captured image analyses. In this method, ice loads on structures are calculated by defining structure coordinates and processing captured images. A pyramidal structure is designed with nine cylindrical bars as the known structure of experimental setup. Unsymmetrical ice accumulated on the structure in a cold room represents the actual case of experiments. Camera intrinsic and extrinsic parameters are used to define structure coordinates in the image coordinate system according to the camera location and angle. The thresholding method is applied to capture images and detect iced structures in a binary image. The ice thickness of each element is calculated by combining the information from the binary image and the structure coordinate. Averaging ice diameters from different camera views obtains ice thicknesses of structure elements. Comparison between ice load measurements using this method and the actual ice loads shows positive correlations with an acceptable range of error. The method can be applied to complex structures defining structure and camera coordinates.

Keywords: camera calibration, ice detection, ice load measurements, image processing

Procedia PDF Downloads 352
19651 Distributed Perceptually Important Point Identification for Time Series Data Mining

Authors: Tak-Chung Fu, Ying-Kit Hung, Fu-Lai Chung

Abstract:

In the field of time series data mining, the concept of the Perceptually Important Point (PIP) identification process is first introduced in 2001. This process originally works for financial time series pattern matching and it is then found suitable for time series dimensionality reduction and representation. Its strength is on preserving the overall shape of the time series by identifying the salient points in it. With the rise of Big Data, time series data contributes a major proportion, especially on the data which generates by sensors in the Internet of Things (IoT) environment. According to the nature of PIP identification and the successful cases, it is worth to further explore the opportunity to apply PIP in time series ‘Big Data’. However, the performance of PIP identification is always considered as the limitation when dealing with ‘Big’ time series data. In this paper, two distributed versions of PIP identification based on the Specialized Binary (SB) Tree are proposed. The proposed approaches solve the bottleneck when running the PIP identification process in a standalone computer. Improvement in term of speed is obtained by the distributed versions.

Keywords: distributed computing, performance analysis, Perceptually Important Point identification, time series data mining

Procedia PDF Downloads 414
19650 Second Time’s a Charm: The Intervention of the European Patent Office on the Strategic Use of Divisional Applications

Authors: Alissa Lefebre

Abstract:

It might seem intuitive to hope for a fast decision on the patent grant. After all, a granted patent provides you with a monopoly position, which allows you to obstruct others from using your technology. However, this does not take into account the strategic advantages one can obtain from keeping their patent applications pending. First, you have the financial advantage of postponing certain fees, although many applicants would probably agree that this is not the main benefit. As the scope of the patent protection is only decided upon at the grant, the pendency period introduces uncertainty amongst rivals. This uncertainty entails not knowing whether the patent will actually get granted and what the scope of protection will be. Consequently, rivals can only depend upon limited and uncertain information when deciding what technology is worth pursuing. One way to keep patent applications pending, is the use of divisional applications. These applicants can be filed out of a parent application as long as that parent application is still pending. This allows the applicant to pursue (part of) the content of the parent application in another application, as the divisional application cannot exceed the scope of the parent application. In a fast-moving and complex market such as the tele- and digital communications, it might allow applicants to obtain an actual monopoly position as competitors are discouraged to pursue a certain technology. Nevertheless, this practice also has downsides to it. First of all, it has an impact on the workload of the examiners at the patent office. As the number of patent filings have been increasing over the last decades, using strategies that increase this number even more, is not desirable from the patent examiners point of view. Secondly, a pending patent does not provide you with the protection of a granted patent, thus not only create uncertainty for the rivals, but also for the applicant. Consequently, the European patent office (EPO) has come up with a “raising the bar initiative” in which they have decided to tackle the strategic use of divisional applications. Over the past years, two rules have been implemented. The first rule in 2010 introduced a time limit, upon which divisional applications could only be filed within a 24-month limit after the first communication with the patent office. However, after carrying-out a user feedback survey, the EPO abolished the rule again in 2014 and replaced it by a fee mechanism. The fee mechanism is still in place today, which might be an indication of a better result compared to the first rule change. This study tests the impact of these rules on the strategic use of divisional applications in the tele- and digital communication industry and provides empirical evidence on their success. Upon using three different survival models, we find overall evidence that divisional applications prolong the pendency time and that only the second rule is able to tackle the strategic patenting and thus decrease the pendency time.

Keywords: divisional applications, regulatory changes, strategic patenting, EPO

Procedia PDF Downloads 110
19649 Harmonic Data Preparation for Clustering and Classification

Authors: Ali Asheibi

Abstract:

The rapid increase in the size of databases required to store power quality monitoring data has demanded new techniques for analysing and understanding the data. One suggested technique to assist in analysis is data mining. Preparing raw data to be ready for data mining exploration take up most of the effort and time spent in the whole data mining process. Clustering is an important technique in data mining and machine learning in which underlying and meaningful groups of data are discovered. Large amounts of harmonic data have been collected from an actual harmonic monitoring system in a distribution system in Australia for three years. This amount of acquired data makes it difficult to identify operational events that significantly impact the harmonics generated on the system. In this paper, harmonic data preparation processes to better understanding of the data have been presented. Underlying classes in this data has then been identified using clustering technique based on the Minimum Message Length (MML) method. The underlying operational information contained within the clusters can be rapidly visualised by the engineers. The C5.0 algorithm was used for classification and interpretation of the generated clusters.

Keywords: data mining, harmonic data, clustering, classification

Procedia PDF Downloads 230
19648 Retail Strategy to Reduce Waste Keeping High Profit Utilizing Taylor's Law in Point-of-Sales Data

Authors: Gen Sakoda, Hideki Takayasu, Misako Takayasu

Abstract:

Waste reduction is a fundamental problem for sustainability. Methods for waste reduction with point-of-sales (POS) data are proposed, utilizing the knowledge of a recent econophysics study on a statistical property of POS data. Concretely, the non-stationary time series analysis method based on the Particle Filter is developed, which considers abnormal fluctuation scaling known as Taylor's law. This method is extended for handling incomplete sales data because of stock-outs by introducing maximum likelihood estimation for censored data. The way for optimal stock determination with pricing the cost of waste reduction is also proposed. This study focuses on the examination of the methods for large sales numbers where Taylor's law is obvious. Numerical analysis using aggregated POS data shows the effectiveness of the methods to reduce food waste maintaining a high profit for large sales numbers. Moreover, the way of pricing the cost of waste reduction reveals that a small profit loss realizes substantial waste reduction, especially in the case that the proportionality constant  of Taylor’s law is small. Specifically, around 1% profit loss realizes half disposal at =0.12, which is the actual  value of processed food items used in this research. The methods provide practical and effective solutions for waste reduction keeping a high profit, especially with large sales numbers.

Keywords: food waste reduction, particle filter, point-of-sales, sustainable development goals, Taylor's law, time series analysis

Procedia PDF Downloads 114
19647 Early Identification and Early Intervention: Pre and Post Diagnostic Tests in Mathematics Courses

Authors: Kailash Ghimire, Manoj Thapa

Abstract:

This study focuses on early identification of deficiencies in pre-required areas of students who are enrolled in College Algebra and Calculus I classes. The students were given pre-diagnostic tests on the first day of the class before they are provided with the syllabus. The tests consist of prerequisite, uniform and advanced content outlined by the University System of Georgia (USG). The results show that 48% of students in College Algebra are lacking prerequisite skills while 52% of Calculus I students are lacking prerequisite skills but, interestingly these students are prior exposed to uniform content and advanced content. The study is still in progress and this paper contains the outcome from Fall 2017 and Spring 2018. In this paper, early intervention used in these classes: two days vs three days meeting a week and students’ self-assessment using exam wrappers and their effectiveness on students’ learning will also be discussed. A result of this study shows that there is an improvement on Drop, Fail and Withdraw (DFW) rates by 7%-10% compared to those in previous semesters.

Keywords: student at risk, diagnostic tests, identification, intervention, normalization gain, validity of tests

Procedia PDF Downloads 194
19646 Regular or Irregular: An Investigation of Medicine Consumption Pattern with Poisson Mixture Model

Authors: Lichung Jen, Yi Chun Liu, Kuan-Wei Lee

Abstract:

Fruitful data has been accumulated in database nowadays and is commonly used as support for decision-making. In the healthcare industry, hospital, for instance, ordering pharmacy inventory is one of the key decision. With large drug inventory, the current cost increases and its expiration dates might lead to future issue, such as drug disposal and recycle. In contrast, underestimating demand of the pharmacy inventory, particularly standing drugs, affects the medical treatment and possibly hospital reputation. Prescription behaviour of hospital physicians is one of the critical factor influencing this decision, particularly irregular prescription behaviour. If a drug’s usage amount in the month is irregular and less than the regular usage, it may cause the trend of subsequent stockpiling. On the contrary, if a drug has been prescribed often than expected, it may result in insufficient inventory. We proposed a hierarchical Bayesian mixture model with two components to identify physicians’ regular/irregular prescription patterns with probabilities. Heterogeneity of hospital is considered in our proposed hierarchical Bayes model. The result suggested that modeling the prescription patterns of physician is beneficial for estimating the order quantity of medication and pharmacy inventory management of the hospital. Managerial implication and future research are discussed.

Keywords: hierarchical Bayesian model, poission mixture model, medicines prescription behavior, irregular behavior

Procedia PDF Downloads 113
19645 Elasticity Model for Easing Peak Hour Demand for Metrorail Transport System

Authors: P. K. Sarkar, Amit Kumar Jain

Abstract:

The demand for Urban transportation is characterised by a large scale temporal and spatial variations which causes heavy congestion inside metro trains in peak hours near Centre Business District (CBD) of the city. The conventional approach to address peak hour congestion, metro trains has been to increase the supply by way of introduction of more trains, increasing the length of the trains, optimising the time table to increase the capacity of the system. However, there is a limitation of supply side measures determined by the design capacity of the systems beyond which any addition in the capacity requires huge capital investments. The demand side interventions are essentially required to actually spread the demand across the time and space. In this study, an attempt has been made to identify the potential Transport Demand Management tools applicable to Urban Rail Transportation systems with a special focus on differential pricing. A conceptual price elasticity model has been developed to analyse the effect of various combinations of peak and nonpeak hoursfares on demands. The elasticity values for peak hour, nonpeak hour and cross elasticity have been assumed from the relevant literature available in the field. The conceptual price elasticity model so developed is based on assumptions which need to be validated with actual values of elasticities for different segments of passengers. Once validated, the model can be used to determine the peak and nonpeak hour fares with an objective to increase overall ridership, revenue, demand levelling and optimal utilisation of assets.

Keywords: urban transport, differential fares, congestion, transport demand management, elasticity

Procedia PDF Downloads 293
19644 An Output Oriented Super-Efficiency Model for Considering Time Lag Effect

Authors: Yanshuang Zhang, Byungho Jeong

Abstract:

There exists some time lag between the consumption of inputs and the production of outputs. This time lag effect should be considered in calculating efficiency of decision making units (DMU). Recently, a couple of DEA models were developed for considering time lag effect in efficiency evaluation of research activities. However, these models can’t discriminate efficient DMUs because of the nature of basic DEA model in which efficiency scores are limited to ‘1’. This problem can be resolved a super-efficiency model. However, a super efficiency model sometimes causes infeasibility problem. This paper suggests an output oriented super-efficiency model for efficiency evaluation under the consideration of time lag effect. A case example using a long term research project is given to compare the suggested model with the MpO model

Keywords: DEA, Super-efficiency, Time Lag, research activities

Procedia PDF Downloads 638
19643 Examining the Links between Established Principles, Iranian Teachers' Perceptions of Reading Comprehension, and Their Actual Practice in English for Specific Purposes Courses

Authors: Zahra Alimorad

Abstract:

There is a strong belief that language teachers' actual practices in the classroom context are largely determined by the underlying perceptions they hold about the nature of language and language learning. That being so, it can be envisaged that teaching procedures of ESP (English for Specific Purposes) teachers teaching reading comprehension will mainly be driven by their perceptions about the nature of reading. To examine this issue, four Iranian university professors holding Ph.D. in either TEFL (Teaching English as a Foreign Language) or English Literature who were teaching English to Engineering and Sciences students were recruited to participate in this study. To collect the necessary data, classroom observations and follow-up semi-structured interviews were used. Furthermore, the materials utilized by the teachers such as textbooks, syllabuses, and tests were also examined. Although it can be argued that their perceptions were partially compatible with the established principles, results of the study pointed to a lack of congruence between these teachers' perceptions and their practices, on the one hand, and between the established principles and the practices, on the other. While the literature mostly supports a metacognitive-strategy approach to reading comprehension, the teachers were mainly adopting a skills-based approach to the teaching of reading. That is, they primarily focused on translation as the core activity in the classroom followed by reading aloud, defining words, and explaining grammatical structures. This divergence was partly attributed to the contextual constraints and partly to students' lack of motivation by the teachers.

Keywords: English teachers, perceptions, practice, principles, reading comprehension

Procedia PDF Downloads 235
19642 An Application of E-Learning Technology for Students with Deafness and Hearing Impairment

Authors: Eyup Bayram Guzel

Abstract:

There have been growing awareness that technology offers unique and promising advantages by offering up-to-data educational materials in promoting teaching and learning materials, new strategies for building enhanced communication environment for people with disabilities and specifically for this study concentrated on the students with deafness and hearing impairments. Creating e-learning environment where teachers and students work in collaboration to develop better educational outcomes is the foremost reason of conducting this research. This study examined the perspectives of special education teachers’ regarding an application of e-learning software called Multimedia Builder on the students with deafness and hearing impairments. Initial and follow up interviews were conducted with 15 special education teachers around the scope of qualitative case study. Grounded approach has been used to analyse and interpret the data. The research results revealed that application of Multimedia Builder software were influential on reading, sign language, vocabulary improvements, computer and ICT usage developments and on audio-visual learning achievements for the advantages of students with deafness and hearing impairments. The implications of the study encouraged the ways of using e-learning tools and strategies to promote unique and comprehensive learning experiences for the targeted students and their teachers.

Keywords: e-learning, special education, deafness and hearing impairment, computer-ICT usage.

Procedia PDF Downloads 421
19641 A Super-Efficiency Model for Evaluating Efficiency in the Presence of Time Lag Effect

Authors: Yanshuang Zhang, Byungho Jeong

Abstract:

In many cases, there is a time lag between the consumption of inputs and the production of outputs. This time lag effect should be considered in evaluating the performance of organizations. Recently, a couple of DEA models were developed for considering time lag effect in efficiency evaluation of research activities. Multi-periods input(MpI) and Multi-periods output(MpO) models are integrated models to calculate simple efficiency considering time lag effect. However, these models can’t discriminate efficient DMUs because of the nature of basic DEA model in which efficiency scores are limited to ‘1’. That is, efficient DMUs can’t be discriminated because their efficiency scores are same. Thus, this paper suggests a super-efficiency model for efficiency evaluation under the consideration of time lag effect based on the MpO model. A case example using a long-term research project is given to compare the suggested model with the MpO model.

Keywords: DEA, super-efficiency, time lag, multi-periods input

Procedia PDF Downloads 454
19640 Dual Mode “Turn On-Off-On” Photoluminescence Detection of EDTA and Lead Using Moringa Oleifera Gum-Derived Carbon Dots

Authors: Anisha Mandal, Swambabu Varanasi

Abstract:

Lead is one of the most prevalent toxic heavy metal ions, and its pollution poses a significant threat to the environment and human health. On the other hand, Ethylenediaminetetraacetic acid is a widely used metal chelating agent that, due to its poor biodegradability, is an incessant pollutant to the environment. For the first time, a green, simple, and cost-effective approach is used to hydrothermally synthesise photoluminescent carbon dots using Moringa Oleifera Gum in a single step. Then, using Moringa Oleifera Gum-derived carbon dots, a photoluminescent "ON-OFF-ON" mechanism for dual mode detection of trace Pb2+ and EDTA was proposed. MOG-CDs detect Pb2+ selectively and sensitively using a photoluminescence quenching mechanism, with a detection limit (LOD) of 0.000472 ppm. (1.24 nM). The quenched photoluminescence can be restored by adding EDTA to the MOG-CD+Pb2+ system; this strategy is used to quantify EDTA at a level of detection of 0.0026 ppm. (8.9 nM). The quantification of Pb2+ and EDTA in actual samples encapsulated the applicability and dependability of the proposed photoluminescent probe.

Keywords: carbon dots, photoluminescence, sensor, moringa oleifera gum

Procedia PDF Downloads 90
19639 Success of Trabeculectomy: May Not Always Depend on Mitomycin C

Authors: Sushma Tejwani, Shoruba Dinakaran, Rupa Rokhade, K. Bhujang Shetty

Abstract:

Introduction and aim: One of the major causes for failure of trabeculectomy is fibrosis and scarring of subconjunctival tissue around the bleb, and hence intra operative usage of anti-fibrotic agents like Mitomycin C (MMC) has become very popular. However, the long term effects of MMC like thin, avascular bleb, hypotony, bleb leaks and late onset endophthalmitis cannot be ignored, and may preclude its usage in routine trabeculectomy. In this particular study we aim to study the outcomes of trabeculectomy with and without MMC in uncomplicated glaucoma patients. Methods: Retrospective study of series of patients that underwent trabeculectomy with or without cataract surgery in glaucoma department of a tertiary eye care centre by a single surgeon for primary open angle glaucoma (POAG), angle closure glaucoma (PACG), Pseudoexfoliation glaucoma (PXF glaucoma). Patients with secondary glaucoma, juvenile and congenital glaucoma were excluded; also patients undergoing second trabeculectomy were excluded. The outcomes were studied in terms of IOP control at 1 month, 6 months, and 1 year and were analyzed separately for surgical outcomes with and without MMC. Success was considered if IOP was < 16 mmHg on applanation tonometry. Further, the necessity of medication, 5 fluorouracil (5FU) postoperative injections, needling post operatively was noted. Results: Eighty nine patient’s medical records were reviewed, of which 58 patients had undergone trabeculectomy without MMC and 31 with MMC. Mean age was 62.4 (95%CI 61- 64), 34 were females and 55 males. MMC group (n=31): Preoperative mean IOP was 21.1mmHg (95% CI: 17.6 -24.6), and 22 patients had IOP > 16. Three out of 33 patients were on single medication and rests were on multiple drugs. At 1 month (n=27) mean IOP was 12.4 mmHg (CI: 10.7-14), and 31/33 had success. At 6 months (n=18) mean IOP was 13mmHg (CI: 10.3-14.6) and 16/18 had good outcome, however at 1 year only 11 patients were available for follow up and 91% (10/11) had success. Overall, 3 patients required medication and one patient required postoperative injection of 5 FU. No MMC group (n=58): Preoperative mean IOP was 21.9 mmHg (CI: 19.8-24.2), and 42 had IOP > 16 mmHg. 12 out of 58 patients were on single medication and rests were on multiple drugs. At 1 month (n=52) mean IOP was14.6mmHg (CI: 13.2-15.9), and 45/ 58 had IOP < 16mmHg. At 6 months (n=31) mean IOP was 13.5 mmHg (CI: 11.9-15.2) and 26/31 had success, however at 1 year only 23 patients came for follow up and of these 87% (20/23) patients had success. Overall, 1 patient required needling, 5 required 5 FU injections and 5 patients required medication. The success rates at each follow up visit were not significantly different in both the groups. Conclusion: Intra-operative MMC usage may not be required in all patients undergoing trabeculectomy, and the ones without MMC also have fairly good outcomes in primary glaucoma.

Keywords: glaucoma filtration surgery, mitomycin C, outcomes of trabeculectomy, wound modulation

Procedia PDF Downloads 260
19638 A Study on Inverse Determination of Impact Force on a Honeycomb Composite Panel

Authors: Hamed Kalhori, Lin Ye

Abstract:

In this study, an inverse method was developed to reconstruct the magnitude and duration of impact forces exerted to a rectangular carbon fibre-epoxy composite honeycomb sandwich panel. The dynamic signals captured by Piezoelectric (PZT) sensors installed on the panel remotely from the impact locations were utilized to reconstruct the impact force generated by an instrumented hammer through an extended deconvolution approach. Two discretized forms of convolution integral are considered; the traditional one with an explicit transfer function and the modified one without an explicit transfer function. Deconvolution, usually applied to reconstruct the time history (e.g. magnitude) of a stochastic force at a defined location, is extended to identify both the location and magnitude of the impact force among a number of potential impact locations. It is assumed that a number of impact forces are simultaneously exerted to all potential locations, but the magnitude of all forces except one is zero, implicating that the impact occurs only at one location. The extended deconvolution is then applied to determine the magnitude as well as location (among the potential ones), incorporating the linear superposition of responses resulted from impact at each potential location. The problem can be categorized into under-determined (the number of sensors is less than that of impact locations), even-determined (the number of sensors equals that of impact locations), or over-determined (the number of sensors is greater than that of impact locations) cases. For an under-determined case, it comprises three potential impact locations and one PZT sensor for the rectangular carbon fibre-epoxy composite honeycomb sandwich panel. Assessments are conducted to evaluate the factors affecting the precision of the reconstructed force. Truncated Singular Value Decomposition (TSVD) and the Tikhonov regularization are independently chosen to regularize the problem to find the most suitable method for this system. The selection of optimal value of the regularization parameter is investigated through L-curve and Generalized Cross Validation (GCV) methods. In addition, the effect of different width of signal windows on the reconstructed force is examined. It is observed that the impact force generated by the instrumented impact hammer is sensitive to the impact locations of the structure, having a shape from a simple half-sine to a complicated one. The accuracy of the reconstructed impact force is evaluated using the correlation co-efficient between the reconstructed force and the actual one. Based on this criterion, it is concluded that the forces reconstructed by using the extended deconvolution without an explicit transfer function together with Tikhonov regularization match well with the actual forces in terms of magnitude and duration.

Keywords: honeycomb composite panel, deconvolution, impact localization, force reconstruction

Procedia PDF Downloads 519
19637 Optimization of Proton Exchange Membrane Fuel Cell Parameters Based on Modified Particle Swarm Algorithms

Authors: M. Dezvarei, S. Morovati

Abstract:

In recent years, increasing usage of electrical energy provides a widespread field for investigating new methods to produce clean electricity with high reliability and cost management. Fuel cells are new clean generations to make electricity and thermal energy together with high performance and no environmental pollution. According to the expansion of fuel cell usage in different industrial networks, the identification and optimization of its parameters is really significant. This paper presents optimization of a proton exchange membrane fuel cell (PEMFC) parameters based on modified particle swarm optimization with real valued mutation (RVM) and clonal algorithms. Mathematical equations of this type of fuel cell are presented as the main model structure in the optimization process. Optimized parameters based on clonal and RVM algorithms are compared with the desired values in the presence and absence of measurement noise. This paper shows that these methods can improve the performance of traditional optimization methods. Simulation results are employed to analyze and compare the performance of these methodologies in order to optimize the proton exchange membrane fuel cell parameters.

Keywords: clonal algorithm, proton exchange membrane fuel cell (PEMFC), particle swarm optimization (PSO), real-valued mutation (RVM)

Procedia PDF Downloads 332
19636 Insight-Based Evaluation of a Map-Based Dashboard

Authors: Anna Fredriksson Häägg, Charlotte Weil, Niklas Rönnberg

Abstract:

Map-based dashboards are used for data exploration every day. The present study used an insight-based methodology for evaluating a map-based dashboard that presents research findings of water management and ecosystem services in the Amazon. In addition to analyzing the insights gained from using the dashboard, the evaluation method was compared to standardized questionnaires and task-based evaluations. The result suggests that the dashboard enabled the participants to gain domain-relevant, complex insights regarding the topic presented. Furthermore, the insight-based analysis highlighted unexpected insights and hypotheses regarding causes and potential adaptation strategies for remediation. Although time- and resource-consuming, the insight-based methodology was shown to have the potential of thoroughly analyzing how end users can utilize map-based dashboards for data exploration and decision making. Finally, the insight-based methodology is argued to evaluate tools in scenarios more similar to real-life usage compared to task-based evaluation methods.

Keywords: visual analytics, dashboard, insight-based evaluation, geographic visualization

Procedia PDF Downloads 100
19635 Safe and Scalable Framework for Participation of Nodes in Smart Grid Networks in a P2P Exchange of Short-Term Products

Authors: Maciej Jedrzejczyk, Karolina Marzantowicz

Abstract:

Traditional utility value chain is being transformed during last few years into unbundled markets. Increased distributed generation of energy is one of considerable challenges faced by Smart Grid networks. New sources of energy introduce volatile demand response which has a considerable impact on traditional middlemen in E&U market. The purpose of this research is to search for ways to allow near-real-time electricity markets to transact with surplus energy based on accurate time synchronous measurements. A proposed framework evaluates the use of secure peer-2-peer (P2P) communication and distributed transaction ledgers to provide flat hierarchy, and allow real-time insights into present and forecasted grid operations, as well as state and health of the network. An objective is to achieve dynamic grid operations with more efficient resource usage, higher security of supply and longer grid infrastructure life cycle. Methods used for this study are based on comparative analysis of different distributed ledger technologies in terms of scalability, transaction performance, pluggability with external data sources, data transparency, privacy, end-to-end security and adaptability to various market topologies. An intended output of this research is a design of a framework for safer, more efficient and scalable Smart Grid network which is bridging a gap between traditional components of the energy network and individual energy producers. Results of this study are ready for detailed measurement testing, a likely follow-up in separate studies. New platforms for Smart Grid achieving measurable efficiencies will allow for development of new types of Grid KPI, multi-smart grid branches, markets, and businesses.

Keywords: autonomous agents, Distributed computing, distributed ledger technologies, large scale systems, micro grids, peer-to-peer networks, Self-organization, self-stabilization, smart grids

Procedia PDF Downloads 280
19634 Carbapenem Usage in Medical Wards: An Antibiotic Stewardship Feedback Project

Authors: Choon Seong Ng, P. Petrick, C. L. Lau

Abstract:

Background: Carbapenem-resistant isolates have been increasingly reported recently. Carbapenem stewardship is designed to optimize its usage particularly among medical wards with high prevalence of carbapenem prescriptions to combat such emerging resistance. Carbapenem stewardship programmes (CSP) can reduce antibiotic use but clinical outcome of such measures needs further evaluation. We examined this in a prospective manner using feedback mechanism. Methods: Our single-center prospective cohort study involved all carbapenem prescriptions across the medical wards (including medical patients admitted to intensive care unit) in a tertiary university hospital setting. The impact of such stewardship was analysed according to the accepted and the rejected groups. The primary endpoint was safety. Safety measure applied in this study was the death at 1 month. Secondary endpoints included length of hospitalisation and readmission. Results: Over the 19 months’ period, input from 144 carbapenem prescriptions was analysed on the basis of acceptance of our CSP recommendations on the use of carbapenems. Recommendations made were as follows : de-escalation of carbapenem; stopping the carbapenem; use for a short duration of 5-7 days; required prolonged duration in the case of carbapenem-sensitive Extended Spectrum Beta-Lactamases bacteremia; dose adjustment; and surgical intervention for removal of septic foci. De-escalation, shorten duration of carbapenem and carbapenem cessation comprised 79% of the recommendations. Acceptance rate was 57%. Those who accepted CSP recommendations had no increase in mortality (p = 0.92), had a shorter length of hospital stay (LOS) and had cost-saving. Infection-related deaths were found to be higher among those in the rejected group. Moreover, three rejected cases (6%) among all non-indicated cases (n = 50) were found to have developed carbapenem-resistant isolates. Lastly, Pitt’s bacteremia score appeared to be a key element affecting the carbapenem prescription’s behaviour in this trial. Conclusions: Carbapenem stewardship program in the medical wards not only saves money, but most importantly it is safe and does not harm the patients with added benefits of reducing the length of hospital stay. However, more time is needed to engage the primary clinical teams by formal clinical presentation and immediate personal feedback by senior Infectious Disease (ID) personnel to increase its acceptance.

Keywords: audit and feedback, carbapenem stewardship, medical wards, university hospital

Procedia PDF Downloads 192
19633 Enabling Enterprise Information System Interoperability: A Future Perspective

Authors: Mahdi Alkaeed, Adeel Ehsan

Abstract:

Enterprise information systems (EIS) act as the backbone of organizations that belong to different domains. These systems not only play a major role in the efficient usage of resources and time but also throw light on the future roadmap for the enterprise. In today's rapidly expanding world of business and technology, enterprise systems from various heterogenous environments have to exchange information at some point, be it within the same organization or between different organizations. This reality strengthens the importance of interoperability between these systems, which is one of the key enablers of systems collaboration. Both information technology infrastructure and business processes have to be aligned with each other to achieve this effect. This will be difficult to attain if traditional tightly coupled architecture is used. Instead, a more loosely coupled service-oriented architecture has to be used. That would enable an effective interoperability level between different EIS. This paper discusses and presents the current work that has been done in the field of EIS interoperability. Along the way, it also discusses the challenges, solutions to tackle those challenges presented in the studied literature, and limitations, if any.

Keywords: enterprise systems interoperability, collaboration and integration, service-based architecture, open system architecture

Procedia PDF Downloads 96
19632 E-Government Adoption in Zimbabwe's Local Government: Understanding the Influence of Attitudes and Perceptions of Residents in Selected Cases

Authors: Ricky Munyaradzi Mukonza

Abstract:

E-government literature continues to grow as scholars and practitioners endeavour to understand this phenomenon. There are many facets of e-government that have been written about including its definition, adoption, and implementation and so on. However, more still needs to be known particularly in relation to how e-government is being adopted in different contexts. There could be many context specific factors that have a bearing on e-government adoption and in this paper focus is on attitudes and perceptions. Association between usage of e-government services and various perceptions such as ease of use, transparency, security, ease of understanding, communication, reliability, relevancy, perceived usefulness and perceived trust is examined. Within the Zimbabwean context and in particular the country’s local government sphere, such a study has not been done. The main aim of the paper is therefore to establish perceptions and attitudes towards e-government services among residents in Zimbabwe’s two local authorities. In terms of research methodology the paper is based on a Mixed Methods Approach (MMA) to collect and analyse data giving the researcher a holistic picture of the phenomenon being investigated. A sample of 785 residents from the two local authorities was used and these were selected using a combination of cluster and purposive sampling methods. A key finding in this paper is that a majority of respondents who have had the opportunity to use e-government services perceive the services to be easy to use, transparent, secure, easy to understand, reliable, relevant, useful and trustworthy. The paper, therefore, makes an important contribution on the relationship between residents’ perceptions and attitudes and e-government usage within the chosen cases.

Keywords: adoption, attitudes, e-government, perceptions

Procedia PDF Downloads 296
19631 Time Compression in Engineer-to-Order Industry: A Case Study of a Norwegian Shipbuilding Industry

Authors: Tarek Fatouh, Chehab Elbelehy, Alaa Abdelsalam, Eman Elakkad, Alaa Abdelshafie

Abstract:

This paper aims to explore the possibility of time compression in Engineer to Order production networks. A case study research method is used in a Norwegian shipbuilding project by implementing a value stream mapping lean tool with total cycle time as a unit of analysis. The analysis resulted in demonstrating the time deviations for the planned tasks in one of the processes in the shipbuilding project. So, authors developed a future state map by removing time wastes from value stream process.

Keywords: engineer to order, total cycle time, value stream mapping, shipbuilding

Procedia PDF Downloads 142
19630 Time and Kinematics of Moving Bodies

Authors: Muhammad Omer Farooq Saeed

Abstract:

The purpose of the proposal is to find out what time actually is! And to understand the natural phenomenon of the behavior of time and light corresponding to the motion of the bodies at relatively high speeds. The utmost concern of the paper is to deal with the possible demerits in the equations of relativity, thereby providing some valuable extensions in those equations and concepts. The idea used develops the most basic conception of the relative motion of the body with respect to space and a real understanding of time and the variation of energy of the body in different frames of reference. The results show the development of a completely new understanding of time, relative motion and energy, along with some extensions in the equations of special relativity most importantly the time dilation and the mass-energy relationship that will explain all frames of a body, all in one go. The proposal also raises serious questions on the validity of the “Principle of Equivalence” on which the General Relativity is based, most importantly a serious case of the bending light that eventually goes against its own governing concepts of space-time being proposed in the theory. The results also predict the existence of a completely new field that explains the fact just how and why bodies acquire energy in space-time. This field explains the production of gravitational waves based on time. All in all, this proposal challenges the formulas and conceptions of Special and General Relativity, respectively.

Keywords: time, relative motion, energy, speed, frame of reference, photon, curvature, space-time, time –differentials

Procedia PDF Downloads 50
19629 Perceptions of Climate Change and Adaptation of Climate-Smart Technology by the Paddy Farmers: A Case Study of Kandy District in Sri Lanka

Authors: W. A. D. P. Wanigasundera, P. C. B. Alahakoon

Abstract:

Kandy district in Sri Lanka has small scale and rain-fed paddy farming, and highly vulnerable to climate change. In this study, the status of climate change was assessed using meteorological data and compared with the perceptions of paddy farming community. Factors affecting the adaptation to the climate smart farming were also assessed. Meteorological data for 33 years were collected and the changes over time compared with the perceptions of farmers. The temperature, rainfall and number of rainy days have increased in both locations. The onset of rains also has shifted. The perceptions of the majority of the farmers were in line with the actual changes. The knowledge and attitudes about the causes of climate change and adaptation were medium and related to level of adoption. Formulating effective communication strategies, and a collaborative approach involving state, private sector, civil society to make Sri Lankan agriculture ‘climate-smart’ is urgently needed.

Keywords: adaptation of climate-smart technology, climate change, perception, rain-fed paddy

Procedia PDF Downloads 309
19628 Performance of HVOF Sprayed Ni-20CR and Cr3C2-NiCr Coatings on Fe-Based Superalloy in an Actual Industrial Environment of a Coal Fired Boiler

Authors: Tejinder Singh Sidhu

Abstract:

Hot corrosion has been recognized as a severe problem in steam-powered electricity generation plants and industrial waste incinerators as it consumes the material at an unpredictably rapid rate. Consequently, the load-carrying ability of the components reduces quickly, eventually leading to catastrophic failure. The inability to either totally prevent hot corrosion or at least detect it at an early stage has resulted in several accidents, leading to loss of life and/or destruction of infrastructures. A number of countermeasures are currently in use or under investigation to combat hot corrosion, such as using inhibitors, controlling the process parameters, designing a suitable industrial alloy, and depositing protective coatings. However, the protection system to be selected for a particular application must be practical, reliable, and economically viable. Due to the continuously rising cost of the materials as well as increased material requirements, the coating techniques have been given much more importance in recent times. Coatings can add value to products up to 10 times the cost of the coating. Among the different coating techniques, thermal spraying has grown into a well-accepted industrial technology for applying overlay coatings onto the surfaces of engineering components to allow them to function under extreme conditions of wear, erosion-corrosion, high-temperature oxidation, and hot corrosion. In this study, the hot corrosion performances of Ni-20Cr and Cr₃C₂-NiCr coatings developed by High Velocity Oxy-Fuel (HVOF) process have been studied. The coatings were developed on a Fe-based superalloy, and experiments were performed in an actual industrial environment of a coal-fired boiler. The cyclic study was carried out around the platen superheater zone where the temperature was around 1000°C. The study was conducted for 10 cycles, and one cycle was consisting of 100 hours of heating followed by 1 hour of cooling at ambient temperature. Both the coatings deposited on Fe-based superalloy imparted better hot corrosion resistance than the uncoated one. The Ni-20Cr coated superalloy performed better than the Cr₃C₂-NiCr coated in the actual working conditions of the coal fired boiler. It is found that the formation of chromium oxide at the boundaries of Ni-rich splats of the coating blocks the inward permeation of oxygen and other corrosive species to the substrate.

Keywords: hot corrosion, coating, HVOF, oxidation

Procedia PDF Downloads 66
19627 The Use of Smartphones as a News Resource by Female University Students in the UAE

Authors: Mahinaz Saad

Abstract:

Little empirical data exists regarding smartphone usage for news consumption in the UAE, and no previous research explored undergraduate female university students’ usage of smartphones. This represents a gap in the professional literature and makes it an important area to examine. Uses and Gratifications theory is used to study the motivations of consumers for adopting a particular type of communication tool. This theory is an audience-centred approach to understanding mass communication that assumes audiences are active consumers of media and explains why and how people seek out specific media to satisfy needs. This theory is particularly relevant given the rapid development of new communication technologies. Situated within this theoretical framework, this study utilised a quantitative research design to explore respondents’ (N=488) how and why respondents use their smartphones. Further, this study explored the relationship between mobile news use and the use of other mediums for news access and how different gratifications predict mobile hard news use and mobile soft news use. Results revealed that smartphones often replace traditional media as a news source and have become students’ primary source of news. Results also revealed that different gratifications can be used as a predictor of mobile hard news and soft news and that most students use their smartphones to access soft news. These results are fundamental in allowing us to predict possible future trends relating to news consumption in the UAE and the myriad ways in which the media landscape is changing.

Keywords: uses and gratifications, smartphones, university students, news consumption

Procedia PDF Downloads 128
19626 Water Balance Components under Climate Change in Croatia

Authors: Jelena Bašić, Višnjica Vučetić, Mislav Anić, Tomislav Bašić

Abstract:

Lack of precipitation combined with high temperatures causes great damage to the agriculture and economy in Croatia. Therefore, it is important to understand water circulation and balance. We decided to gain a better insight into the spatial distribution of water balance components (WBC) and their long-term changes in Croatia. WBC are precipitation (P), potential evapotranspiration (PET), actual evapotranspiration (ET), soil moisture content (S), runoff (RO), recharge (R), and soil moisture loss (L). Since measurements of the mentioned components in Croatia are very rare, the Palmer model has been applied to estimate them. We refined method by setting into the account the corrective factor to include influence effects of the wind as well as a maximum soil capacity for specific soil types. We will present one hundred years’ time series of PET and ET showing the trends at few meteorological stations and a comparison of components of two climatological periods. The meteorological data from 109 stations have been used for the spatial distribution map of the WBC of Croatia.

Keywords: croatia, long-term trends, the palmer method, water balance components

Procedia PDF Downloads 124
19625 Assessing Acute Toxicity and Endocrine Disruption Potential of Selected Packages Internal Layers Extracts

Authors: N. Szczepanska, B. Kudlak, G. Yotova, S. Tsakovski, J. Namiesnik

Abstract:

In the scientific literature related to the widely understood issue of packaging materials designed to have contact with food (food contact materials), there is much information on raw materials used for their production, as well as their physiochemical properties, types, and parameters. However, not much attention is given to the issues concerning migration of toxic substances from packaging and its actual influence on the health of the final consumer, even though health protection and food safety are the priority tasks. The goal of this study was to estimate the impact of particular foodstuff packaging type, food production, and storage conditions on the degree of leaching of potentially toxic compounds and endocrine disruptors to foodstuffs using the acute toxicity test Microtox and XenoScreen YES YAS assay. The selected foodstuff packaging materials were metal cans used for fish storage and tetrapak. Five stimulants respectful to specific kinds of food were chosen in order to assess global migration: distilled water for aqueous foods with a pH above 4.5; acetic acid at 3% in distilled water for acidic aqueous food with pH below 4.5; ethanol at 5% for any food that may contain alcohol; dimethyl sulfoxide (DMSO) and artificial saliva were used in regard to the possibility of using it as an simulation medium. For each packaging three independent variables (temperature and contact time) factorial design simulant was performed. Xenobiotics migration from epoxy resins was studied at three different temperatures (25°C, 65°C, and 121°C) and extraction time of 12h, 48h and 2 weeks. Such experimental design leads to 9 experiments for each food simulant as conditions for each experiment are obtained by combination of temperature and contact time levels. Each experiment was run in triplicate for acute toxicity and in duplicate for estrogen disruption potential determination. Multi-factor analysis of variation (MANOVA) was used to evaluate the effects of the three main factors solvent, temperature (temperature regime for cup), contact time and their interactions on the respected dependent variable (acute toxicity or estrogen disruption potential). From all stimulants studied the most toxic were can and tetrapak lining acetic acid extracts that are indication for significant migration of toxic compounds. This migration increased with increase of contact time and temperature and justified the hypothesis that food products with low pH values cause significant damage internal resin filling. Can lining extracts of all simulation medias excluding distilled water and artificial saliva proved to contain androgen agonists even at 25°C and extraction time of 12h. For tetrapak extracts significant endocrine potential for acetic acid, DMSO and saliva were detected.

Keywords: food packaging, extraction, migration, toxicity, biotest

Procedia PDF Downloads 165