Search results for: real time data.
11055 Forensic Speaker Verification in Noisy Environmental by Enhancing the Speech Signal Using ICA Approach
Authors: Ahmed Kamil Hasan Al-Ali, Bouchra Senadji, Ganesh Naik
Abstract:
We propose a system to real environmental noise and channel mismatch for forensic speaker verification systems. This method is based on suppressing various types of real environmental noise by using independent component analysis (ICA) algorithm. The enhanced speech signal is applied to mel frequency cepstral coefficients (MFCC) or MFCC feature warping to extract the essential characteristics of the speech signal. Channel effects are reduced using an intermediate vector (i-vector) and probabilistic linear discriminant analysis (PLDA) approach for classification. The proposed algorithm is evaluated by using an Australian forensic voice comparison database, combined with car, street and home noises from QUT-NOISE at a signal to noise ratio (SNR) ranging from -10 dB to 10 dB. Experimental results indicate that the MFCC feature warping-ICA achieves a reduction in equal error rate about (48.22%, 44.66%, and 50.07%) over using MFCC feature warping when the test speech signals are corrupted with random sessions of street, car, and home noises at -10 dB SNR.Keywords: Noisy forensic speaker verification, ICA algorithm, MFCC, MFCC feature warping.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 99111054 Solution Approaches for Some Scheduling Problems with Learning Effect and Job Dependent Delivery Times
Authors: M. Duran Toksarı, B. Uçarkuş
Abstract:
In this paper, we propose two algorithms to optimally solve makespan and total completion time scheduling problems with learning effect and job dependent delivery times in a single machine environment. The delivery time is the extra time to eliminate adverse effect between the main processing and delivery to the customer. In this paper, we introduce the job dependent delivery times for some single machine scheduling problems with position dependent learning effect, which are makespan are total completion. The results with respect to two algorithms proposed for solving of the each problem are compared with LINGO solutions for 50-jobs, 100-jobs and 150- jobs problems. The proposed algorithms can find the same results in shorter time.Keywords: Delivery times, learning effect, makespan, scheduling, total completion time.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 155111053 Moving Data Mining Tools toward a Business Intelligence System
Authors: Nittaya Kerdprasop, Kittisak Kerdprasop
Abstract:
Data mining (DM) is the process of finding and extracting frequent patterns that can describe the data, or predict unknown or future values. These goals are achieved by using various learning algorithms. Each algorithm may produce a mining result completely different from the others. Some algorithms may find millions of patterns. It is thus the difficult job for data analysts to select appropriate models and interpret the discovered knowledge. In this paper, we describe a framework of an intelligent and complete data mining system called SUT-Miner. Our system is comprised of a full complement of major DM algorithms, pre-DM and post-DM functionalities. It is the post-DM packages that ease the DM deployment for business intelligence applications.Keywords: Business intelligence, data mining, functionalprogramming, intelligent system.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 174211052 Edit Distance Algorithm to Increase Storage Efficiency of Javanese Corpora
Authors: Aji P. Wibawa, Andrew Nafalski, Neil Murray, Wayan F. Mahmudy
Abstract:
Since the one-to-one word translator does not have the facility to translate pragmatic aspects of Javanese, the parallel text alignment model described uses a phrase pair combination. The algorithm aligns the parallel text automatically from the beginning to the end of each sentence. Even though the results of the phrase pair combination outperform the previous algorithm, it is still inefficient. Recording all possible combinations consume more space in the database and time consuming. The original algorithm is modified by applying the edit distance coefficient to improve the data-storage efficiency. As a result, the data-storage consumption is 90% reduced as well as its learning period (42s).Keywords: edit distance coefficient, Javanese, parallel text alignment, phrase pair combination
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 172811051 Analysis of Diverse Clustering Tools in Data Mining
Authors: S. Sarumathi, N. Shanthi, M. Sharmila
Abstract:
Clustering in data mining is an unsupervised learning technique of aggregating the data objects into meaningful groups such that the intra cluster similarity of objects are maximized and inter cluster similarity of objects are minimized. Over the past decades several clustering tools were emerged in which clustering algorithms are inbuilt and are easier to use and extract the expected results. Data mining mainly deals with the huge databases that inflicts on cluster analysis and additional rigorous computational constraints. These challenges pave the way for the emergence of powerful expansive data mining clustering softwares. In this survey, a variety of clustering tools used in data mining are elucidated along with the pros and cons of each software.
Keywords: Cluster Analysis, Clustering Algorithms, Clustering Techniques, Association, Visualization.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 220111050 The Role of People and Data in Complex Spatial-Related Long-Term Decisions: A Case Study of Capital Project Management Groups
Authors: Peter Boyes, Sarah Sharples, Paul Tennent, Gary Priestnall, Jeremy Morley
Abstract:
Significant long-term investment projects can involve complex decisions. These are often described as capital projects and the factors that contribute to their complexity include budgets, motivating reasons for investment, stakeholder involvement, interdependent projects, and the delivery phases required. The complexity of these projects often requires management groups to be established involving stakeholder representatives, these teams are inherently multidisciplinary. This study uses two university campus capital projects as case studies for this type of management group. Due to the interaction of projects with wider campus infrastructure and users, decisions are made at varying spatial granularity throughout the project lifespan. This spatial-related context brings complexity to the group decisions. Sensemaking is the process used to achieve group situational awareness of a complex situation, enabling the team to arrive at a consensus and make a decision. The purpose of this study is to understand the role of people and data in complex spatial related long-term decision and sensemaking processes. The paper aims to identify and present issues experienced in practical settings of these types of decision. A series of exploratory semi-structured interviews with members of the two projects elicit an understanding of their operation. From two stages of thematic analysis, inductive and deductive, emergent themes are identified around the group structure, the data usage, and the decision making within these groups. When data were made available to the group, there were commonly issues with perception of veracity and validity of the data presented; this impacted the ability of the group to reach consensus and therefore for decision to be made. Similarly, there were different responses to forecasted or modelled data, shaped by the experience and occupation of the individuals within the multidisciplinary management group. This paper provides an understanding of further support required for team sensemaking and decision making in complex capital projects. The paper also discusses the barriers found to effective decision making in this setting and suggests opportunities to develop decision support systems in this team strategic decision-making process. Recommendations are made for further research into the sensemaking and decision-making process of this complex spatial-related setting.
Keywords: decision making, decisions under uncertainty, real decisions, sensemaking, spatial, team decision making
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 49211049 A Monte Carlo Method to Data Stream Analysis
Authors: Kittisak Kerdprasop, Nittaya Kerdprasop, Pairote Sattayatham
Abstract:
Data stream analysis is the process of computing various summaries and derived values from large amounts of data which are continuously generated at a rapid rate. The nature of a stream does not allow a revisit on each data element. Furthermore, data processing must be fast to produce timely analysis results. These requirements impose constraints on the design of the algorithms to balance correctness against timely responses. Several techniques have been proposed over the past few years to address these challenges. These techniques can be categorized as either dataoriented or task-oriented. The data-oriented approach analyzes a subset of data or a smaller transformed representation, whereas taskoriented scheme solves the problem directly via approximation techniques. We propose a hybrid approach to tackle the data stream analysis problem. The data stream has been both statistically transformed to a smaller size and computationally approximated its characteristics. We adopt a Monte Carlo method in the approximation step. The data reduction has been performed horizontally and vertically through our EMR sampling method. The proposed method is analyzed by a series of experiments. We apply our algorithm on clustering and classification tasks to evaluate the utility of our approach.Keywords: Data Stream, Monte Carlo, Sampling, DensityEstimation.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 141711048 Determinants of Extra Charges for Container Shipments: A Case Study of Nexus Zone Logistics
Authors: Zety Shakila Binti Mohd Yusof, Muhammad Adib Bin Ishak, Hajah Fatimah Binti Hussein
Abstract:
The international shipping business is related to numerous controls or regulations of export and import shipments. It is costly and time consuming, and when something goes wrong or when the buyer or seller fails to comply with the regulations, it can result in penalties, delays, and unexpected costs etc. For the focus of this study, the researchers have selected a local forwarder that provides forwarding and clearance services, Nexus Zone Logistics. It was identified that this company currently has many extra costs to be paid including local and detention charges, which negatively impacts the flow of income and reduces overall stability. Two variables have been identified as factors of extra charges; loaded containers entering the port by exceeded closing time and late delivery of empty containers to the container yard. This study is a qualitative in nature and the secondary data collected was analyzed using self-administered observation. The findings of this study were covered by one selected case for each export and import shipment between July and December 2014. The data were analyzed using frequency analysis based on tables and graphs. The researcher recommends Nexus Zone Logistics impose a 1% deposit payment per container for each shipment (export and import) to its customers.Keywords: International shipping, export and import, detention charges, container shipment.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 118411047 Low-Cost Inertial Sensors Modeling Using Allan Variance
Authors: A. A. Hussen, I. N. Jleta
Abstract:
Micro-electromechanical system (MEMS) accelerometers and gyroscopes are suitable for the inertial navigation system (INS) of many applications due to low price, small dimensions and light weight. The main disadvantage in a comparison with classic sensors is a worse long term stability. The estimation accuracy is mostly affected by the time-dependent growth of inertial sensor errors, especially the stochastic errors. In order to eliminate negative effects of these random errors, they must be accurately modeled. In this paper, the Allan variance technique will be used in modeling the stochastic errors of the inertial sensors. By performing a simple operation on the entire length of data, a characteristic curve is obtained whose inspection provides a systematic characterization of various random errors contained in the inertial-sensor output data.Keywords: Allan variance, accelerometer, gyroscope, stochastic errors.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 527711046 Assessing Chemo-Radiotherapy Induced Toxicity and Quality of Life of Cancer Patients
Authors: A. Taqaddas
Abstract:
Chemotherapy and radiotherapy are one of the major treatment modalities that play important role in the management of a number of different cancers. This study for the first time evaluates the toxicity of these treatment modalities and its impact on quality of life of cancer patients in Pakistan. The study also for the first time determines what cancer patients of different ages and cancer stages believe would be an effective intervention to manage their psychosocial needs and treatment induced toxicity. The article also provides evidence based approach for the use of variety of interventions to mange cancer treatment induced morbidity and toxicity. In light of the present study and reviewed research data, evidence based recommendations are also made for selection of appropriate interventions to manage Pain, Nausea and Vomiting, Anxiety and Depression, Fatigue and Overall QOL of cancer survivors.
Keywords: Chemotherapy Toxicity, Psycho-Social Interventions, Quality of Life, Radiotherapy Toxicity.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 223111045 Uneven Development: Structural Changes and Income Outcomes across States in Malaysia
Authors: Siti Aiysyah Tumin
Abstract:
This paper looks at the nature of structural changes—the transition of employment from agriculture, to manufacturing, then to different types of services—in different states in Malaysia and links it to income outcomes for households and workers. Specifically, this paper investigates the conditional association between the concentration of different economic activities and income outcomes (household incomes and employee wages) in almost four decades. Using publicly available state-level employment and income data, we found that significant wage premium was associated with “modern” services (finance, real estate, professional, information and communication), which are urban-based services sectors that employ a larger proportion of skilled and educated workers. However, employment in manufacturing and other services subsectors was significantly associated with a lower income dispersion and inequality, alluding to their importance in welfare improvements.Keywords: Employment, labour market, structural change, wages.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 53311044 Parallel Computation in Hypersonic Aerodynamic Heating Problem
Authors: Ding Guo-hao, Li Hua, Wang Wen-long
Abstract:
A parallel computational fluid dynamics code has been developed for the study of aerodynamic heating problem in hypersonic flows. The code employs the 3D Navier-Stokes equations as the basic governing equations to simulate the laminar hypersonic flow. The cell centered finite volume method based on structured grid is applied for spatial discretization. The AUSMPW+ scheme is used for the inviscid fluxes, and the MUSCL approach is used for higher order spatial accuracy. The implicit LU-SGS scheme is applied for time integration to accelerate the convergence of computations in steady flows. A parallel programming method based on MPI is employed to shorten the computing time. The validity of the code is demonstrated by comparing the numerical calculation result with the experimental data of a hypersonic flow field around a blunt body.Keywords: Aerodynamic Heating, AUSMPW+, MPI, ParallelComputation
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 196511043 Exponential Passivity Criteria for BAM Neural Networks with Time-Varying Delays
Authors: Qingqing Wang, Baocheng Chen, Shouming Zhong
Abstract:
In this paper,the exponential passivity criteria for BAM neural networks with time-varying delays is studied.By constructing new Lyapunov-Krasovskii functional and dividing the delay interval into multiple segments,a novel sufficient condition is established to guarantee the exponential stability of the considered system.Finally,a numerical example is provided to illustrate the usefulness of the proposed main results
Keywords: BAM neural networks, Exponential passivity, LMI approach, Time-varying delays.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 190811042 Finding Pareto Optimal Front for the Multi- Mode Time, Cost Quality Trade-off in Project Scheduling
Authors: H. Iranmanesh, M. R. Skandari, M. Allahverdiloo
Abstract:
Project managers are the ultimate responsible for the overall characteristics of a project, i.e. they should deliver the project on time with minimum cost and with maximum quality. It is vital for any manager to decide a trade-off between these conflicting objectives and they will be benefited of any scientific decision support tool. Our work will try to determine optimal solutions (rather than a single optimal solution) from which the project manager will select his desirable choice to run the project. In this paper, the problem in project scheduling notated as (1,T|cpm,disc,mu|curve:quality,time,cost) will be studied. The problem is multi-objective and the purpose is finding the Pareto optimal front of time, cost and quality of a project (curve:quality,time,cost), whose activities belong to a start to finish activity relationship network (cpm) and they can be done in different possible modes (mu) which are non-continuous or discrete (disc), and each mode has a different cost, time and quality . The project is constrained to a non-renewable resource i.e. money (1,T). Because the problem is NP-Hard, to solve the problem, a meta-heuristic is developed based on a version of genetic algorithm specially adapted to solve multi-objective problems namely FastPGA. A sample project with 30 activities is generated and then solved by the proposed method.Keywords: FastPGA, Multi-Execution Activity Mode, Pareto Optimality, Project Scheduling, Time-Cost-Quality Trade-Off.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 180711041 Effects of the Sintering Process on Properties of Triaxial Electrical Porcelain from Ugandan Ceramic Minerals
Authors: Peter W. Olupot, Stefan Jonsson, Joseph K. Byaruhanga
Abstract:
Porcelain specimens were fired at 6C/min to 1250C (dwell time 0.5-3h) and cooled at 6C/min to room temperature. Additionally, three different slower firing/cooling cycles were tried. Sintering profile and effects on MOR, crystalline phase content and morphology were investigated using dilatometry, 4-point bending strength, XRD and FEG-SEM respectively. Industrial-sized specimens prepared using the promising cycle were tested basing on the ANSI standards. Increasing dwell time from 1h to 3h at peak temperature of 1250C resulted in neither a significant effect on the quartz and mullite content nor MOR. Reducing the firing/cooling rate to below 6C/min, for peak temperature of 1250C (dwell time of 1h) does not result in improvement of strength of porcelain. The industrial sized specimen exhibited flashover voltages of 20.3kV (dry) and 9.3kV (wet) respectively, transverse strength of 12.5kN and bulk density of 2.27g/cm3, which are satisfactory. There was however dye penetration during porosity test. KeywordsDwell time, Microstructure, Porcelain, Strength.Keywords: Dwell time, Microstructure, Porcelain, Strength.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 299011040 Centralized Resource Management for Network Infrastructure Including Ip Telephony by Integrating a Mediator Between the Heterogeneous Data Sources
Authors: Mohammed Fethi Khalfi, Malika Kandouci
Abstract:
Over the past decade, mobile has experienced a revolution that will ultimately change the way we communicate.All these technologies have a common denominator exploitation of computer information systems, but their operation can be tedious because of problems with heterogeneous data sources.To overcome the problems of heterogeneous data sources, we propose to use a technique of adding an extra layer interfacing applications of management or supervision at the different data sources.This layer will be materialized by the implementation of a mediator between different host applications and information systems frequently used hierarchical and relational manner such that the heterogeneity is completely transparent to the VoIP platform.Keywords: TOIP, Data Integration, Mediation, informationcomputer system, heterogeneous data sources
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 133211039 Empirical Mode Decomposition Based Denoising by Customized Thresholding
Authors: Wahiba Mohguen, Raïs El’hadi Bekka
Abstract:
This paper presents a denoising method called EMD-Custom that was based on Empirical Mode Decomposition (EMD) and the modified Customized Thresholding Function (Custom) algorithms. EMD was applied to decompose adaptively a noisy signal into intrinsic mode functions (IMFs). Then, all the noisy IMFs got threshold by applying the presented thresholding function to suppress noise and to improve the signal to noise ratio (SNR). The method was tested on simulated data and real ECG signal, and the results were compared to the EMD-Based signal denoising methods using the soft and hard thresholding. The results showed the superior performance of the proposed EMD-Custom denoising over the traditional approach. The performances were evaluated in terms of SNR in dB, and Mean Square Error (MSE).
Keywords: Customized thresholding, ECG signal, EMD, hard thresholding, Soft-thresholding.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 108511038 Secure Multiparty Computations for Privacy Preserving Classifiers
Authors: M. Sumana, K. S. Hareesha
Abstract:
Secure computations are essential while performing privacy preserving data mining. Distributed privacy preserving data mining involve two to more sites that cannot pool in their data to a third party due to the violation of law regarding the individual. Hence in order to model the private data without compromising privacy and information loss, secure multiparty computations are used. Secure computations of product, mean, variance, dot product, sigmoid function using the additive and multiplicative homomorphic property is discussed. The computations are performed on vertically partitioned data with a single site holding the class value.Keywords: Homomorphic property, secure product, secure mean and variance, secure dot product, vertically partitioned data.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 92011037 Color Image Edge Detection using Pseudo-Complement and Matrix Operations
Authors: T. N. Janakiraman, P. V. S. S. R. Chandra Mouli
Abstract:
A color image edge detection algorithm is proposed in this paper using Pseudo-complement and matrix rotation operations. First, pseudo-complement method is applied on the image for each channel. Then, matrix operations are applied on the output image of the first stage. Dominant pixels are obtained by image differencing between the pseudo-complement image and the matrix operated image. Median filtering is carried out to smoothen the image thereby removing the isolated pixels. Finally, the dominant or core pixels occurring in at least two channels are selected. On plotting the selected edge pixels, the final edge map of the given color image is obtained. The algorithm is also tested in HSV and YCbCr color spaces. Experimental results on both synthetic and real world images show that the accuracy of the proposed method is comparable to other color edge detectors. All the proposed procedures can be applied to any image domain and runs in polynomial time.Keywords: Color edge detection, dominant pixels, matrixrotation/shift operations, pseudo-complement.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 233111036 On the Accuracy of Basic Modal Displacement Method Considering Various Earthquakes
Authors: Seyed Sadegh Naseralavi, Sadegh Balaghi, Ehsan Khojastehfar
Abstract:
Time history seismic analysis is supposed to be the most accurate method to predict the seismic demand of structures. On the other hand, the required computational time of this method toward achieving the result is its main deficiency. While being applied in optimization process, in which the structure must be analyzed thousands of time, reducing the required computational time of seismic analysis of structures makes the optimization algorithms more practical. Apparently, the invented approximate methods produce some amount of errors in comparison with exact time history analysis but the recently proposed method namely, Complete Quadratic Combination (CQC) and Sum Root of the Sum of Squares (SRSS) drastically reduces the computational time by combination of peak responses in each mode. In the present research, the Basic Modal Displacement (BMD) method is introduced and applied towards estimation of seismic demand of main structure. Seismic demand of sampled structure is estimated by calculation of modal displacement of basic structure (in which the modal displacement has been calculated). Shear steel sampled structures are selected as case studies. The error applying the introduced method is calculated by comparison of the estimated seismic demands with exact time history dynamic analysis. The efficiency of the proposed method is demonstrated by application of three types of earthquakes (in view of time of peak ground acceleration).Keywords: Time history dynamic analysis, basic modal displacement, earthquake induced demands, shear steel structures.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 142011035 Design of Buffer Management for Industry to Avoid Sensor Data- Conflicts
Authors: Dae-ho Won, Jong-wook Hong, Yeon-Mo Yang, Jinung An
Abstract:
To reduce accidents in the industry, WSNs(Wireless Sensor networks)- sensor data is used. WSNs- sensor data has the persistence and continuity. therefore, we design and exploit the buffer management system that has the persistence and continuity to avoid and delivery data conflicts. To develop modules, we use the multi buffers and design the buffer management modules that transfer sensor data through the context-aware methods.Keywords: safe management system, buffer management, context-aware, input data stream
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 155411034 Evaluating Spectral Relationships between Signals by Removing the Contribution of a Common, Periodic Source A Partial Coherence-based Approach
Authors: Antonio Mauricio F. L. Miranda de Sá
Abstract:
Partial coherence between two signals removing the contribution of a periodic, deterministic signal is proposed for evaluating the interrelationship in multivariate systems. The estimator expression was derived and shown to be independent of such periodic signal. Simulations were used for obtaining its critical value, which were found to be the same as those for Gaussian signals, as well as for evaluating the technique. An Illustration with eletroencephalografic (EEG) signals during photic stimulation is also provided. The application of the proposed technique in both simulation and real EEG data indicate that it seems to be very specific in removing the contribution of periodic sources. The estimate independence of the periodic signal may widen partial coherence application to signal analysis, since it could be used together with simple coherence to test for contamination in signals by a common, periodic noise source.
Keywords: Partial coherence, periodic input, spectral analysis, statistical signal processing.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 146411033 Probability Distribution of Rainfall Depth at Hourly Time-Scale
Authors: S. Dan'azumi, S. Shamsudin, A. A. Rahman
Abstract:
Rainfall data at fine resolution and knowledge of its characteristics plays a major role in the efficient design and operation of agricultural, telecommunication, runoff and erosion control as well as water quality control systems. The paper is aimed to study the statistical distribution of hourly rainfall depth for 12 representative stations spread across Peninsular Malaysia. Hourly rainfall data of 10 to 22 years period were collected and its statistical characteristics were estimated. Three probability distributions namely, Generalized Pareto, Exponential and Gamma distributions were proposed to model the hourly rainfall depth, and three goodness-of-fit tests, namely, Kolmogorov-Sminov, Anderson-Darling and Chi-Squared tests were used to evaluate their fitness. Result indicates that the east cost of the Peninsular receives higher depth of rainfall as compared to west coast. However, the rainfall frequency is found to be irregular. Also result from the goodness-of-fit tests show that all the three models fit the rainfall data at 1% level of significance. However, Generalized Pareto fits better than Exponential and Gamma distributions and is therefore recommended as the best fit.Keywords: Goodness-of-fit test, Hourly rainfall, Malaysia, Probability distribution.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 292011032 Mean Velocity Modeling of Open-Channel Flow with Submerged Rigid Vegetation
Authors: M. Morri, A. Soualmia, P. Belleudy
Abstract:
Vegetation affects the mean and turbulent flow structure. It may increase flood risks and sediment transport. Therefore, it is important to develop analytical approaches for the bed shear stress on vegetated bed, to predict resistance caused by vegetation. In the recent years, experimental and numerical models have both been developed to model the effects of submerged vegetation on open-channel flow. In this paper, different analytic models are compared and tested using the criteria of deviation, to explore their capacity for predicting the mean velocity and select the suitable one that will be applied in real case of rivers. The comparison between the measured data in vegetated flume and simulated mean velocities indicated, a good performance, in the case of rigid vegetation, whereas, Huthoff model shows the best agreement with a high coefficient of determination (R2=80%) and the smallest error in the prediction of the average velocities.
Keywords: Analytic Models, Comparison, Mean Velocity, Vegetation.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 253811031 Multiple Periodic Solutions for a Delayed Predator-prey System on Time Scales
Authors: Xiaoquan Ding, Jianmin Hao, Changwen Liu
Abstract:
This paper is devoted to a delayed periodic predatorprey system with non-monotonic numerical response on time scales. With the help of a continuation theorem based on coincidence degree theory, we establish easily verifiable criteria for the existence of multiple periodic solutions. As corollaries, some applications are listed. In particular, our results improve and generalize some known ones.
Keywords: Predator-prey system, periodic solution, time scale, delay, coincidence degree.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 137011030 Injury Prediction for Soccer Players Using Machine Learning
Authors: Amiel Satvedi, Richard Pyne
Abstract:
Injuries in professional sports occur on a regular basis. Some may be minor while others can cause huge impact on a player’s career and earning potential. In soccer, there is a high risk of players picking up injuries during game time. This research work seeks to help soccer players reduce the risk of getting injured by predicting the likelihood of injury while playing in the near future and then providing recommendations for intervention. The injury prediction tool will use a soccer player’s number of minutes played on the field, number of appearances, distance covered and performance data for the current and previous seasons as variables to conduct statistical analysis and provide injury predictive results using a machine learning linear regression model.
Keywords: Injury predictor, soccer injury prevention, machine learning in soccer, big data in soccer.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 174711029 Non-Homogeneous Layered Fiber Reinforced Concrete
Authors: Vitalijs Lusis, Andrejs Krasnikovs
Abstract:
Fiber reinforced concrete is important material for load bearing structural elements. Usually fibers are homogeneously distributed in a concrete body having arbitrary spatial orientations. At the same time, in many situations, fiber concrete with oriented fibers is more optimal. Is obvious, that is possible to create constructions with oriented short fibers in them, in different ways. Present research is devoted to one of such approaches- fiber reinforced concrete prisms having dimensions 100mm ×100mm ×400mmwith layers of non-homogeneously distributed fibers inside them were fabricated.
Simultaneously prisms with homogeneously dispersed fibers were produced for reference as well. Prisms were tested under four point bending conditions. During the tests vertical deflection at the center of every prism and crack opening were measured (using linear displacements transducers in real timescale). Prediction results were discussed.
Keywords: Fiber reinforced concrete, 4-point bending, steel fiber.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 300911028 Order Partitioning in Hybrid MTS/MTO Contexts using Fuzzy ANP
Authors: H. Rafiei, M. Rabbani
Abstract:
A novel concept to balance and tradeoff between make-to-stock and make-to-order has been hybrid MTS/MTO production context. One of the most important decisions involved in the hybrid MTS/MTO environment is determining whether a product is manufactured to stock, to order, or hybrid MTS/MTO strategy. In this paper, a model based on analytic network process is developed to tackle the addressed decision. Since the regarded decision deals with the uncertainty and ambiguity of data as well as experts- and managers- linguistic judgments, the proposed model is equipped with fuzzy sets theory. An important attribute of the model is its generality due to diverse decision factors which are elicited from the literature and developed by the authors. Finally, the model is validated by applying to a real case study to reveal how the proposed model can actually be implemented.Keywords: Fuzzy analytic network process, Hybrid make-tostock/ make-to-order, Order partitioning, Production planning.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 217611027 Identification of Differentially Expressed Gene(DEG) in Atherosclerotic Lesion by Annealing Control Primer (ACP)-Based Genefishing™ PCR
Authors: M. Maimunah, G. A. Froemming, H. Nawawi, M. I. Nafeeza, O. Effat, M. Y. Rosmadi, M. S. Mohamed Saifulaman
Abstract:
Atherosclerosis was identified as a chronic inflammatory process resulting from interactions between plasma lipoproteins, cellular components (monocyte, macrophages, T lymphocytes, endothelial cells and smooth muscle cells) and the extracellular matrix of the arterial wall. Several types of genes were known to express during formation of atherosclerosis. This study is carried out to identify unknown differentially expressed gene (DEG) in atherogenesis. Rabbit’s aorta tissues were stained by H&E for histomorphology. GeneFishing™ PCR analysis was performed from total RNA extracted from the aorta tissues. The DNA fragment from DEG was cloned, sequenced and validated by Real-time PCR. Histomorphology showed intimal thickening in the aorta. DEG detected from ACP-41 was identified as cathepsin B gene and showed upregulation at week-8 and week-12 of atherogenesis. Therefore, ACP-based GeneFishing™ PCR facilitated identification of cathepsin B gene which was differentially expressed during development of atherosclerosis.
Keywords: Atherosclerosis, GeneFishing™ PCR, cathepsin B gene.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 195611026 The Investigation of the Role of Institutions in the Process of Growth and Development of Economy
Authors: Seyed Mohammad Reza Hosseini
Abstract:
The new institutional Economics helps generalization and expansion of new classic by adding the institution theories to Economic. It is clear that the appropriate institution is among the factors that lead to success in Economic programs. If the institutional are appropriate, the society will save the source and when we make use of time to apply the program, there will be welfare and average revenue product will also increase. In Economy, one should not expect the real manifestation of Economic programs only with a model for estimating and predicting rather institutions of the same purpose and along with production are needed to form the process of growth and development costs. In this research, the institution role in transaction costs, financial markets, distribution of revenue and capital and its influence on the process of growth and development are investigated so that handicaps and problems of Iran Economic Institutions can be recognized. In other words, incapability, non productivity and ambiguity of the institution in Iran Economic are some of the factors that handicap Economic growth and development. For example, Iran government as an important institution while having 20 ministries,83 organizations and 60 years of programming could not go along the growth and development but why?Keywords: Institution, New institutional economics, Transaction costs.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1616