Search results for: peak measures.
1071 A Combinatorial Model for ECG Interpretation
Authors: Costas S. Iliopoulos, Spiros Michalakopoulos
Abstract:
A new, combinatorial model for analyzing and inter- preting an electrocardiogram (ECG) is presented. An application of the model is QRS peak detection. This is demonstrated with an online algorithm, which is shown to be space as well as time efficient. Experimental results on the MIT-BIH Arrhythmia database show that this novel approach is promising. Further uses for this approach are discussed, such as taking advantage of its small memory requirements and interpreting large amounts of pre-recorded ECG data.Keywords: Combinatorics, ECG analysis, MIT-BIH Arrhythmia Database, QRS Detection, String Algorithms
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 19391070 Philosophy, Geometry, and Purpose in Islamic and Gothic Architecture as Two Religious-Based Styles
Authors: P. Nafisi Poor, P. Javid
Abstract:
Religion and divinity have always held important meaning to humans, and therefore it affects different aspects of life including art and architecture. Numerous works of art are related to religion whether supporting or denying it. Religion and religious scholars have influenced and changed art throughout history. This paper focuses on Islam and Christianity because these two religions have been the most discussed and most popular of all time, starting from the birth of Jesus to the arrival of Mohammad. Based on this popularity, these religions have influenced the arts and especially architecture. Islam on one hand changed Iranian and Arabian architecture and they applied it in different places around the world. From the appearance of Islam at 622 AD to this day, Islamic architecture has been evolving; however, one of the most important periods for this style was between 1501 AD and 1736 AD in Iran. Christianity, on the other hand, changed European architecture especially between 1150 AD and 1450 AD or the so-called "Gothic" era, which begins at medieval time and reaches its peak at International Gothic ages. At both of these periods, designing buildings based on spiritual concepts and divine statements reached its peak, and architects were considering God and religion as their center of attention. This article studies the focus on the religions of Islam and Christianity in terms of architecture and presents a general philosophy of both styles to comprehend the idea behind each one, followed by an analysis of their geometry and architectural aspects derived from the best examples, all to understand the purpose of each style and to realize, which one was more successful in reaching their purpose. Subsequently, a comprehensive review of each building is provided including 3D visualizations to help achieve the goal of the article. These studies can support diverse inquiries about both Islamic and Gothic architecture and can be used as a resource to support studies and research towards designing based on religion or for divine purposes.
Keywords: Architecture, gothic, Islamic, religion.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 9621069 Ankh Key Broadband Array Antenna for 5G Applications
Authors: Noha M. Rashad, W. Swelam, M. H. Abd ElAzeem
Abstract:
A simple design of array antenna is presented in this paper, supporting millimeter wave applications which can be used in short range wireless communications such as 5G applications. This design enhances the use of V-band, according to IEEE standards, as the antenna works in the 70 GHz band with bandwidth more than 11 GHz and peak gain more than 13 dBi. The design is simulated using different numerical techniques achieving a very good agreement.
Keywords: 5G Technology, array antenna, microstrip, millimeter wave.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 9641068 A Multiclass BCMP Queueing Modeling and Simulation-Based Road Traffic Flow Analysis
Authors: Jouhra Dad, Mohammed Ouali, Yahia Lebbah
Abstract:
Urban road network traffic has become one of the most studied research topics in the last decades. This is mainly due to the enlargement of the cities and the growing number of motor vehicles traveling in this road network. One of the most sensitive problems is to verify if the network is congestion-free. Another related problem is the automatic reconfiguration of the network without building new roads to alleviate congestions. These problems require an accurate model of the traffic to determine the steady state of the system. An alternative is to simulate the traffic to see if there are congestions and when and where they occur. One key issue is to find an adequate model for road intersections. Once the model established, either a large scale model is built or the intersection is represented by its performance measures and simulation for analysis. In both cases, it is important to seek the queueing model to represent the road intersection. In this paper, we propose to model the road intersection as a BCMP queueing network and we compare this analytical model against a simulation model for validation.Keywords: Queueing theory, transportation systems, BCMPqueueing network, performance measures, modeling, simulation
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 24421067 Semi-Automatic Method to Assist Expert for Association Rules Validation
Authors: Amdouni Hamida, Gammoudi Mohamed Mohsen
Abstract:
In order to help the expert to validate association rules extracted from data, some quality measures are proposed in the literature. We distinguish two categories: objective and subjective measures. The first one depends on a fixed threshold and on data quality from which the rules are extracted. The second one consists on providing to the expert some tools in the objective to explore and visualize rules during the evaluation step. However, the number of extracted rules to validate remains high. Thus, the manually mining rules task is very hard. To solve this problem, we propose, in this paper, a semi-automatic method to assist the expert during the association rule's validation. Our method uses rule-based classification as follow: (i) We transform association rules into classification rules (classifiers), (ii) We use the generated classifiers for data classification. (iii) We visualize association rules with their quality classification to give an idea to the expert and to assist him during validation process.Keywords: Association rules, Rule-based classification, Classification quality, Validation.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 17911066 Selection of Intensity Measure in Probabilistic Seismic Risk Assessment of a Turkish Railway Bridge
Authors: M. F. Yilmaz, B. Ö. Çağlayan
Abstract:
Fragility curve is an effective common used tool to determine the earthquake performance of structural and nonstructural components. Also, it is used to determine the nonlinear behavior of bridges. There are many historical bridges in the Turkish railway network; the earthquake performances of these bridges are needed to be investigated. To derive fragility curve Intensity measures (IMs) and Engineering demand parameters (EDP) are needed to be determined. And the relation between IMs and EDP are needed to be derived. In this study, a typical simply supported steel girder riveted railway bridge is studied. Fragility curves of this bridge are derived by two parameters lognormal distribution. Time history analyses are done for selected 60 real earthquake data to determine the relation between IMs and EDP. Moreover, efficiency, practicality, and sufficiency of three different IMs are discussed. PGA, Sa(0.2s) and Sa(1s), the most common used IMs parameters for fragility curve in the literature, are taken into consideration in terms of efficiency, practicality and sufficiency.
Keywords: Railway bridges, earthquake performance, fragility analyses, selection of intensity measures.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 9031065 Spreading Dynamics of a Viral Infection in a Complex Network
Authors: Khemanand Moheeput, Smita S. D. Goorah, Satish K. Ramchurn
Abstract:
We report a computational study of the spreading dynamics of a viral infection in a complex (scale-free) network. The final epidemic size distribution (FESD) was found to be unimodal or bimodal depending on the value of the basic reproductive number R0 . The FESDs occurred on time-scales long enough for intermediate-time epidemic size distributions (IESDs) to be important for control measures. The usefulness of R0 for deciding on the timeliness and intensity of control measures was found to be limited by the multimodal nature of the IESDs and by its inability to inform on the speed at which the infection spreads through the population. A reduction of the transmission probability at the hubs of the scale-free network decreased the occurrence of the larger-sized epidemic events of the multimodal distributions. For effective epidemic control, an early reduction in transmission at the index cell and its neighbors was essential.
Keywords: Basic reproductive number, epidemic control, scalefree network, viral infection.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 17201064 Using Axiomatic Design for Developing a Framework of Manufacturing Cloud Service Composition in the Equilibrium State
Authors: Ehsan Vaziri Goodarzi, Mahmood Houshmand, Omid Fatahi Valilai, Vahidreza Ghezavati, Shahrooz Bamdad
Abstract:
One important paradigm of industry 4.0 is Cloud Manufacturing (CM). In CM everything is considered as a service, therefore, the CM platform should consider all service provider's capabilities and tries to integrate services in an equilibrium state. This research develops a framework for implementing manufacturing cloud service composition in the equilibrium state. The developed framework using well-known tools called axiomatic design (AD) and game theory. The research has investigated the factors for forming equilibrium for measures of the manufacturing cloud service composition. Functional requirements (FRs) represent the measures of manufacturing cloud service composition in the equilibrium state. These FRs satisfied by related Design Parameters (DPs). The FRs and DPs are defined by considering the game theory, QoS, consumer needs, parallel and cooperative services. Ultimately, four FRs and DPs represent the framework. To insure the validity of the framework, the authors have used the first AD’s independent axiom.Keywords: Axiomatic design, manufacturing cloud service composition, cloud manufacturing, Industry 4.0.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 7061063 Study on Seismic Performance of Reinforced Soil Walls to Modify the Pseudo Static Method
Authors: Majid Yazdandoust
Abstract:
This study, tries to suggest a design method based on displacement using finite difference numerical modeling in reinforcing soil retaining wall with steel strip. In this case, dynamic loading characteristics such as duration, frequency, peak ground acceleration, geometrical characteristics of reinforced soil structure and type of the site are considered to correct the pseudo static method and finally introduce the pseudo static coefficient as a function of seismic performance level and peak ground acceleration. For this purpose, the influence of dynamic loading characteristics, reinforcement length, height of reinforced system and type of the site are investigated on seismic behavior of reinforcing soil retaining wall with steel strip. Numerical results illustrate that the seismic response of this type of wall is highly dependent to cumulative absolute velocity, maximum acceleration, and height and reinforcement length so that the reinforcement length can be introduced as the main factor in shape of failure. Considering the loading parameters, geometric parameters of the wall and type of the site showed that the used method in this study leads to efficient designs in comparison with other methods, which are usually based on limit-equilibrium concept. The outputs show the over-estimation of equilibrium design methods in comparison with proposed displacement based methods here.Keywords: Pseudo static coefficient, seismic performance design, numerical modeling, steel strip reinforcement, retaining walls, cumulative absolute velocity, failure shape.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 21531062 Nebulized Magnesium Sulfate in Acute Moderate to Severe Asthma in Pediatric Patients
Authors: Lubna M. Zakaryia Mahmoud, Mohammed A. Dawood, Doaa A. Heiba
Abstract:
A prospective double-blind placebo controlled trial carried out on 60 children known to be asthmatic who presented to the emergency department at Alexandria University of Children’s Hospital at El-Shatby with acute asthma exacerbations to assess the efficacy of adding inhaled magnesium sulfate to β-agonist, compared with β-agonist in saline, in the management of acute asthma exacerbations in children. The participants in the study were divided in two groups; Group A (study group) received inhaled salbutamol solution (0.15 ml/kg) plus isotonic magnesium sulfate 2 ml in a nebulizer chamber. Group B (control group): received nebulized salbutamol solution (0.15 ml/kg) diluted with placebo (2 ml normal saline). Both groups received inhaled solution every 20 minutes that was repeated for three doses. They were evaluated using the Pediatric Asthma Severity Score (PASS), oxygen saturation using portable pulse oximetry and peak expiratory flow rate using a portable peak expiratory flow meter at initially recorded as zero-minute assessment and every 20 minutes from the end of each nebulization (nebulization lasts 5-10 minutes) recorded as 20, 40 and 60-minute assessments. Regarding PASS, comparison showed non-significant difference with p-value 0.463, 0.472, 0.0766 at 20, 40 and 60 minutes. Regarding oxygen saturation, improvement was more significant towards group A starting from 40 min with significant p-value=0.000. At 60 min p-value=0.000. Although mean PEFR significantly improved from zero-min in both groups; however, improvement was more significant in group A with significant p-value = 0.015, 0.001, 0.001 at 20 min, 40 min and 60 min, respectively. The conclusion this study suggests is that inhaled magnesium sulfate is an efficient add on drug to standard β- agonist inhalation used in the treatment of moderate to severe asthma exacerbations.
Keywords: Nebulized, magnesium sulfate, acute asthma, pediatric.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 16601061 Exploring the Spatial Characteristics of Mortality Map: A Statistical Area Perspective
Authors: Jung-Hong Hong, Jing-Cen Yang, Cai-Yu Ou
Abstract:
The analysis of geographic inequality heavily relies on the use of location-enabled statistical data and quantitative measures to present the spatial patterns of the selected phenomena and analyze their differences. To protect the privacy of individual instance and link to administrative units, point-based datasets are spatially aggregated to area-based statistical datasets, where only the overall status for the selected levels of spatial units is used for decision making. The partition of the spatial units thus has dominant influence on the outcomes of the analyzed results, well known as the Modifiable Areal Unit Problem (MAUP). A new spatial reference framework, the Taiwan Geographical Statistical Classification (TGSC), was recently introduced in Taiwan based on the spatial partition principles of homogeneous consideration of the number of population and households. Comparing to the outcomes of the traditional township units, TGSC provides additional levels of spatial units with finer granularity for presenting spatial phenomena and enables domain experts to select appropriate dissemination level for publishing statistical data. This paper compares the results of respectively using TGSC and township unit on the mortality data and examines the spatial characteristics of their outcomes. For the mortality data between the period of January 1st, 2008 and December 31st, 2010 of the Taitung County, the all-cause age-standardized death rate (ASDR) ranges from 571 to 1757 per 100,000 persons, whereas the 2nd dissemination area (TGSC) shows greater variation, ranged from 0 to 2222 per 100,000. The finer granularity of spatial units of TGSC clearly provides better outcomes for identifying and evaluating the geographic inequality and can be further analyzed with the statistical measures from other perspectives (e.g., population, area, environment.). The management and analysis of the statistical data referring to the TGSC in this research is strongly supported by the use of Geographic Information System (GIS) technology. An integrated workflow that consists of the tasks of the processing of death certificates, the geocoding of street address, the quality assurance of geocoded results, the automatic calculation of statistic measures, the standardized encoding of measures and the geo-visualization of statistical outcomes is developed. This paper also introduces a set of auxiliary measures from a geographic distribution perspective to further examine the hidden spatial characteristics of mortality data and justify the analyzed results. With the common statistical area framework like TGSC, the preliminary results demonstrate promising potential for developing a web-based statistical service that can effectively access domain statistical data and present the analyzed outcomes in meaningful ways to avoid wrong decision making.
Keywords: Mortality map, spatial patterns, statistical area, variation.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 9901060 Complexity Analysis of Some Known Graph Coloring Instances
Authors: Jeffrey L. Duffany
Abstract:
Graph coloring is an important problem in computer science and many algorithms are known for obtaining reasonably good solutions in polynomial time. One method of comparing different algorithms is to test them on a set of standard graphs where the optimal solution is already known. This investigation analyzes a set of 50 well known graph coloring instances according to a set of complexity measures. These instances come from a variety of sources some representing actual applications of graph coloring (register allocation) and others (mycieleski and leighton graphs) that are theoretically designed to be difficult to solve. The size of the graphs ranged from ranged from a low of 11 variables to a high of 864 variables. The method used to solve the coloring problem was the square of the adjacency (i.e., correlation) matrix. The results show that the most difficult graphs to solve were the leighton and the queen graphs. Complexity measures such as density, mobility, deviation from uniform color class size and number of block diagonal zeros are calculated for each graph. The results showed that the most difficult problems have low mobility (in the range of .2-.5) and relatively little deviation from uniform color class size.Keywords: graph coloring, complexity, algorithm.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 14011059 Assessing Efficiency Trends in the Indian Sugar Industry
Authors: S. P. Singh
Abstract:
This paper measures technical and scale efficiencies of 40 Indian sugar companies for the period from 2004-05 to 2013-14. The efficiencies are estimated through input-oriented DEA models using one output variable—value of output (VOP) and five input variables—capital cost (CA), employee cost (EMP), raw material (RW), energy & fuel (E&F) and other manufacturing expenses (OME). The sugar companies are classified into integrated and non-integrated categories to know which one achieves higher level of efficiency. Sources of inefficiency in the industry are identified through decomposing the overall technical efficiency (TE) into pure technical efficiency (PTE) and scale efficiency (SE). The paper also estimates input-reduction targets for relatively inefficient companies and suggests measures to improve their efficiency level. The findings reveal that the TE does not evince any trend rather it shows fluctuations across years, largely due to erratic and cyclical pattern of sugar production. Further, technical inefficiency in the industry seems to be driven more by the managerial inefficiency than the scale inefficiency, which implies that TE can be improved through better conversion of inputs into output.
Keywords: Sugar industry, companies, technical efficiency, data envelopment analysis, targets.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 13441058 A Black-box Approach for Response Quality Evaluation of Conversational Agent Systems
Authors: Ong Sing Goh, C. Ardil, Wilson Wong, Chun Che Fung
Abstract:
The evaluation of conversational agents or chatterbots question answering systems is a major research area that needs much attention. Before the rise of domain-oriented conversational agents based on natural language understanding and reasoning, evaluation is never a problem as information retrieval-based metrics are readily available for use. However, when chatterbots began to become more domain specific, evaluation becomes a real issue. This is especially true when understanding and reasoning is required to cater for a wider variety of questions and at the same time to achieve high quality responses. This paper discusses the inappropriateness of the existing measures for response quality evaluation and the call for new standard measures and related considerations are brought forward. As a short-term solution for evaluating response quality of conversational agents, and to demonstrate the challenges in evaluating systems of different nature, this research proposes a blackbox approach using observation, classification scheme and a scoring mechanism to assess and rank three example systems, AnswerBus, START and AINI.
Keywords: Evaluation, conversational agents, Response Quality, chatterbots
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 19271057 Main Cause of Children's Deaths in Indigenous Wayuu Community from Department of La Guajira: A Research Developed through Data Mining Use
Authors: Isaura Esther Solano Núñez, David Suarez
Abstract:
The main purpose of this research is to discover what causes death in children of the Wayuu community, and deeply analyze those results in order to take corrective measures to properly control infant mortality. We consider important to determine the reasons that are producing early death in this specific type of population, since they are the most vulnerable to high risk environmental conditions. In this way, the government, through competent authorities, may develop prevention policies and the right measures to avoid an increase of this tragic fact. The methodology used to develop this investigation is data mining, which consists in gaining and examining large amounts of data to produce new and valuable information. Through this technique it has been possible to determine that the child population is dying mostly from malnutrition. In short, this technique has been very useful to develop this study; it has allowed us to transform large amounts of information into a conclusive and important statement, which has made it easier to take appropriate steps to resolve a particular situation.
Keywords: Malnutrition, datamining, analytical, descriptive, population, wayuu, indigenous.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 6961056 Proposing Enterprise Wide Information Systems Business Performance Model
Authors: Vineet Kansal
Abstract:
Enterprise Wide Information Systems (EWIS) implementation involves the entire business and will require changes throughout the firm. Because of the scope, complexity and continuous nature of ERP, the project-based approach to managing the implementation process resulted in failure rates of between 60% and 80%. In recent years ERP systems have received much attention. The organizational relevance and risk of ERP projects make it important for organizations to focus on ways to make ERP implementation successful. Once these systems are in place, however, their performance depends on the identified macro variables viz. 'Business Process', 'Decision Making' and 'Individual / Group working'. The questionnaire was designed and administered. The responses from 92 organizations were compiled. The relationship of these variables with EWIS performance is analyzed using inferential statistical measurements. The study helps to understand the performance of model presented. The study suggested in keeping away from the calamities and thereby giving the necessary competitive edge. Whenever some discrepancy is identified during the process of performance appraisal care has to be taken to draft necessary preventive measures. If all these measures are taken care off then the EWIS performance will definitely deliver the results.Keywords: Enterprise Systems, performance, technology
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 13041055 Personal Factors and Career Adaptability in a Call Centre Work Environment: The Mediating Effects of Professional Efficacy
Authors: Nisha Harry
Abstract:
The study discussed in this article sought to assess whether a sense of professional efficacy mediates the relationship between personal factors and career adaptability. A quantitative cross-sectional survey approach was followed. A non–probability sample of (N = 409) of which predominantly early career and permanently employed black females in call centres in Africa participated in this study. In order to assess personal factors, the participants completed sense of meaningfulness and emotional intelligence measures. Measures of professional efficacy and career adaptability were also completed. The results of the mediational analysis revealed that professional efficacy significantly mediates the meaningfulness (sense of coherence) and career adaptability relationship, but not the emotional intelligence–career adaptability relationship. Call centre agents with professional efficacy are likely to be more work engaged as a result of their sense of meaningfulness and emotional intelligence.
Keywords: Call centre, professional efficacy, career adaptability, emotional intelligence.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 6371054 Talent Management through Integration of Talent Value Chain and Human Capital Analytics Approaches
Authors: Wuttigrai Ngamsirijit
Abstract:
Talent management in today’s modern organizations has become data-driven due to a demand for objective human resource decision making and development of analytics technologies. HR managers have been faced with some obstacles in exploiting data and information to obtain their effective talent management decisions. These include process-based data and records; insufficient human capital-related measures and metrics; lack of capabilities in data modeling in strategic manners; and, time consuming to add up numbers and make decisions. This paper proposes a framework of talent management through integration of talent value chain and human capital analytics approaches. It encompasses key data, measures, and metrics regarding strategic talent management decisions along the organizational and talent value chain. Moreover, specific predictive and prescriptive models incorporating these data and information are recommended to help managers in understanding the state of talent, gaps in managing talent and the organization, and the ways to develop optimized talent strategies.Keywords: Decision making, human capital analytics, talent management, talent value chain.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 9661053 The Forensic Swing of Things: The Current Legal and Technical Challenges of IoT Forensics
Authors: Pantaleon Lutta, Mohamed Sedky, Mohamed Hassan
Abstract:
The inability of organizations to put in place management control measures for Internet of Things (IoT) complexities persists to be a risk concern. Policy makers have been left to scamper in finding measures to combat these security and privacy concerns. IoT forensics is a cumbersome process as there is no standardization of the IoT products, no or limited historical data are stored on the devices. This paper highlights why IoT forensics is a unique adventure and brought out the legal challenges encountered in the investigation process. A quadrant model is presented to study the conflicting aspects in IoT forensics. The model analyses the effectiveness of forensic investigation process versus the admissibility of the evidence integrity; taking into account the user privacy and the providers’ compliance with the laws and regulations. Our analysis concludes that a semi-automated forensic process using machine learning, could eliminate the human factor from the profiling and surveillance processes, and hence resolves the issues of data protection (privacy and confidentiality).
Keywords: Cloud forensics, data protection laws, GDPR, IoT forensics, machine learning.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 10991052 The Impact of Financial Risks on Profitability of Malaysian Commercial Banks: 1996-2005
Authors: Fauziah Hanim Tafri, Zarinah Hamid, Ahamed Kameel Mydin Meera, Mohd Azmi Omar
Abstract:
This paper examines the relationship between financial risks and profitability of the conventional and Islamic banks in Malaysia for the period between 1996 and 2005. The measures of profitability that have been used in the study are the return on equity (ROE) and return on assets (ROA) while the financial risks are credit risk, interest rate risk and liquidity risks. This study employs panel data regression analysis of Generalised Least Squares of fixed effects and random effects models. It was found that credit risk has a significant impact on ROA and ROE for the conventional as well as the Islamic banks. The relationship between interest rate risk and ROE were found to be weakly significant for the conventional banks and insignificant for the Islamic banks. The effect of interest rate risk on ROA is significant for the conventional banks. Liquidity risk was found to have an insignificant impact on both profitability measures.Keywords: Credit risk, interest rate risk, liquidity risk, market risk, profitability.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 58341051 Photovoltaic Array Sizing for PV-Electrolyzer
Authors: Panhathai Buasri
Abstract:
Hydrogen that used as fuel in fuel cell vehicles can be produced from renewable sources such as wind, solar, and hydro technologies. PV-electrolyzer is one of the promising methods to produce hydrogen with zero pollution emission. Hydrogen production from a PV-electrolyzer system depends on the efficiency of the electrolyzer and photovoltaic array, and sun irradiance at that site. In this study, the amount of hydrogen is obtained using mathematical equations for difference driving distance and sun peak hours. The results show that the minimum of 99 PV modules are used to generate 1.75 kgH2 per day for two vehicles.Keywords: About four key words or phrases in alphabetical order, separated by commas.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 17531050 Urban Corridor Management Strategy Based on Intelligent Transportation System
Authors: Sourabh Jain, Sukhvir Singh Jain, Gaurav V. Jain
Abstract:
Intelligent Transportation System (ITS) is the application of technology for developing a user–friendly transportation system for urban areas in developing countries. The goal of urban corridor management using ITS in road transport is to achieve improvements in mobility, safety, and the productivity of the transportation system within the available facilities through the integrated application of advanced monitoring, communications, computer, display, and control process technologies, both in the vehicle and on the road. This paper attempts to present the past studies regarding several ITS available that have been successfully deployed in urban corridors of India and abroad, and to know about the current scenario and the methodology considered for planning, design, and operation of Traffic Management Systems. This paper also presents the endeavor that was made to interpret and figure out the performance of the 27.4 Km long study corridor having eight intersections and four flyovers. The corridor consisting of 6 lanes as well as 8 lanes divided road network. Two categories of data were collected on February 2016 such as traffic data (traffic volume, spot speed, delay) and road characteristics data (no. of lanes, lane width, bus stops, mid-block sections, intersections, flyovers). The instruments used for collecting the data were video camera, radar gun, mobile GPS and stopwatch. From analysis, the performance interpretations incorporated were identification of peak hours and off peak hours, congestion and level of service (LOS) at mid blocks, delay followed by the plotting speed contours and recommending urban corridor management strategies. From the analysis, it is found that ITS based urban corridor management strategies will be useful to reduce congestion, fuel consumption and pollution so as to provide comfort and efficiency to the users. The paper presented urban corridor management strategies based on sensors incorporated in both vehicles and on the roads.
Keywords: Congestion, ITS Strategies, Mobility, Safety.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 16591049 Performance Management of Tangible Assets within the Balanced Scorecard and Interactive Business Decision Tools
Authors: Raymond K. Jonkers
Abstract:
The present study investigated approaches and techniques to enhance strategic management governance and decision making within the framework of a performance-based balanced scorecard. The review of best practices from strategic, program, process, and systems engineering management provided for a holistic approach toward effective outcome-based capability management. One technique, based on factorial experimental design methods, was used to develop an empirical model. This model predicted the degree of capability effectiveness and is dependent on controlled system input variables and their weightings. These variables represent business performance measures, captured within a strategic balanced scorecard. The weighting of these measures enhances the ability to quantify causal relationships within balanced scorecard strategy maps. The focus in this study was on the performance of tangible assets within the scorecard rather than the traditional approach of assessing performance of intangible assets such as knowledge and technology. Tangible assets are represented in this study as physical systems, which may be thought of as being aboard a ship or within a production facility. The measures assigned to these systems include project funding for upgrades against demand, system certifications achieved against those required, preventive maintenance to corrective maintenance ratios, and material support personnel capacity against that required for supporting respective systems. The resultant scorecard is viewed as complimentary to the traditional balanced scorecard for program and performance management. The benefits from these scorecards are realized through the quantified state of operational capabilities or outcomes. These capabilities are also weighted in terms of priority for each distinct system measure and aggregated and visualized in terms of overall state of capabilities achieved. This study proposes the use of interactive controls within the scorecard as a technique to enhance development of alternative solutions in decision making. These interactive controls include those for assigning capability priorities and for adjusting system performance measures, thus providing for what-if scenarios and options in strategic decision-making. In this holistic approach to capability management, several cross functional processes were highlighted as relevant amongst the different management disciplines. In terms of assessing an organization’s ability to adopt this approach, consideration was given to the P3M3 management maturity model.
Keywords: Outcome based management, performance management, lifecycle costs, balanced scorecard.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 13531048 Using Teager Energy Cepstrum and HMM distancesin Automatic Speech Recognition and Analysis of Unvoiced Speech
Authors: Panikos Heracleous
Abstract:
In this study, the use of silicon NAM (Non-Audible Murmur) microphone in automatic speech recognition is presented. NAM microphones are special acoustic sensors, which are attached behind the talker-s ear and can capture not only normal (audible) speech, but also very quietly uttered speech (non-audible murmur). As a result, NAM microphones can be applied in automatic speech recognition systems when privacy is desired in human-machine communication. Moreover, NAM microphones show robustness against noise and they might be used in special systems (speech recognition, speech conversion etc.) for sound-impaired people. Using a small amount of training data and adaptation approaches, 93.9% word accuracy was achieved for a 20k Japanese vocabulary dictation task. Non-audible murmur recognition in noisy environments is also investigated. In this study, further analysis of the NAM speech has been made using distance measures between hidden Markov model (HMM) pairs. It has been shown the reduced spectral space of NAM speech using a metric distance, however the location of the different phonemes of NAM are similar to the location of the phonemes of normal speech, and the NAM sounds are well discriminated. Promising results in using nonlinear features are also introduced, especially under noisy conditions.Keywords: Speech recognition, unvoiced speech, nonlinear features, HMM distance measures
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 16471047 Measuring Risk Levels and Efficacy of Risk Management Strategies in Vietnamese Catfish Farming
Authors: Tru C. Le, France Cheong
Abstract:
Although the Vietnamese catfish farming has grown at very high rates in recent years, the industry has also faced many problems affecting its sustainability. This paper studies the perceptions of catfish farmers regarding risk and risk management strategies in their production activities. Specifically, the study aims to measure the consequences, likelihoods, and levels of risks as well as the efficacy of risk management in Vietnamese catfish farming. Data for the study were collected through a sample of 261 catfish farmers in the Mekong Delta, Vietnam using a questionnaire survey in 2008. Results show that, in general, price and production risks were perceived as the most important risks. Farm management and technical measures were perceived more effective than other kinds of risk management strategies in risk reduction. Although price risks were rated as important risks, price risk management strategies were not perceived as important measures for risk mitigation. The results of the study are discussed to provide implications for various industry stakeholders, including policy makers, processors, advisors, and developers of new risk management strategies.Keywords: Aquaculture, catfish farming, sources of risk, riskmanagement, risk strategies, risk mitigation.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 19861046 An Improved k Nearest Neighbor Classifier Using Interestingness Measures for Medical Image Mining
Authors: J. Alamelu Mangai, Satej Wagle, V. Santhosh Kumar
Abstract:
The exponential increase in the volume of medical image database has imposed new challenges to clinical routine in maintaining patient history, diagnosis, treatment and monitoring. With the advent of data mining and machine learning techniques it is possible to automate and/or assist physicians in clinical diagnosis. In this research a medical image classification framework using data mining techniques is proposed. It involves feature extraction, feature selection, feature discretization and classification. In the classification phase, the performance of the traditional kNN k nearest neighbor classifier is improved using a feature weighting scheme and a distance weighted voting instead of simple majority voting. Feature weights are calculated using the interestingness measures used in association rule mining. Experiments on the retinal fundus images show that the proposed framework improves the classification accuracy of traditional kNN from 78.57 % to 92.85 %.
Keywords: Medical Image Mining, Data Mining, Feature Weighting, Association Rule Mining, k nearest neighbor classifier.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 33081045 Measures for Limiting Corruption upon Migration Wave in Europe
Authors: Jordan Georgiev Deliversky
Abstract:
Fight against migrant smuggling has been put as a priority issues at the European Union policy agenda for more than a decade. The trafficked person, who has been targeted as the object of criminal exploitation, is specifically unique for human trafficking. Generally, the beginning of human trafficking activities is related to profit from the victim’s exploitation. The objective of this paper is to present measures that could result in the limitation of corruption mainly through analyzing the existing legislation framework against corruption in Europe. The analysis is focused on exploring the multiple origins of factors influencing migration processes in Europe, as corruption could be characterized as one of the most significant reasons for refugees to flee their countries. The main results show that law enforcement must turn the focus on the financing of the organized crime groups that are involved in migrant smuggling activities. Corruption has a significant role in managing smuggling operations and in particular when criminal organizations and networks are involved. Illegal migrants and refugees usually represent significant sources of additional income for officials involved in the process of boarding protection and immigration control within the European Union borders.
Keywords: Corruption, influence, human smuggling, legislation, migration.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 10921044 Risks and Mitigation Measures in Build-Operate-Transfer Projects
Authors: Syed Kamarul Bakri Syed Ahmad Bokharey, Kalaikumar Vallyutham, Narayanan Sambu Potty, Nabilah Abu Bakar
Abstract:
Infrastructure investments are important in developing countries, it will not only help to foster the economic growth of a nation, but it will also act as a platform in which new forms of partnership and collaboration can be developed mainly in East Asian countries. Since the last two decades, many infrastructure projects had been completed through build-operate-transfer (BOT) type of procurement. The developments of BOT have attracted participation of local and foreign private sector investor to secure funding and to deliver projects on time, within the budget and to the required specifications. Private sectors are preferred by the government in East Asia to participate in BOT projects due to lack of public funding. The finding has resulted that the private sector or promoter of the BOT projects is exposed to multiple risks which have been discussed in this paper. Effective risk management methods and good managerial skills are required in ensuring the success of the project. The review indicated that mitigation measures should be employed by the promoter throughout the concession period and support from the host government is also required in ensuring the success of the BOT project.Keywords: BOT project, risks management, concessionaire, consortium.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 82051043 Union is Strength in Lossy Image Compression
Authors: Mario Mastriani
Abstract:
In this work, we present a comparison between different techniques of image compression. First, the image is divided in blocks which are organized according to a certain scan. Later, several compression techniques are applied, combined or alone. Such techniques are: wavelets (Haar's basis), Karhunen-Loève Transform, etc. Simulations show that the combined versions are the best, with minor Mean Squared Error (MSE), and higher Peak Signal to Noise Ratio (PSNR) and better image quality, even in the presence of noise.Keywords: Haar's basis, Image compression, Karhunen-LoèveTransform, Morton's scan, row-rafter scan.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 17461042 A Follow up Study on the Elderly Survivors - Mental Health Two Years after the Wenchuan Earthquake
Authors: Ting Wang, Huiqin Yang, Buxin Han
Abstract:
Background: This investigated the mental health of the elderly survivors six months, ten months and two years after the “5.12 Wenchuan" earthquake. Methods: Two hundred and thirty-two physically healthy older survivors from earthquake-affected Mianyang County were interviewed. The measures included the Revised Impact of Event Scale (IES-R, Chinese version, for PTSD) and a Chinese Mental Health Inventory for the Elderly (MHIE). A repeated measures ANOVA test was used for statistical analysis. Results: The follow-up group had a statistically significant lower IES-R score and lower MHIE score than the initial group ten months after the earthquake. Two years later, the score of IES-R in follow-up group were still lower than that of non-follow-up group, but no differences were significant on the score of MHIE between groups. Furthermore, a negative relationship was found between scores of IES-R and MHIE. Conclusion: The earthquake has had a persistent negative impact on older survivors- mental health within the two-year period and that although the PTSD level declined significantly with time, it did not disappear completely.
Keywords: Elderly survivors, follow-up, mental health, post-Wenchuan earthquake.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2025