Search results for: Family size intention
534 Petrology Investigation of Apatite Minerals in the Esfordi Mine, Yazd, Iran
Authors: Haleh Rezaei Zanjirabadi, Fatemeh Saberi, Bahman Rahimzadeh, Fariborz Masoudi, Mohammad Rahgosha
Abstract:
In this study, apatite minerals from the iron-phosphate deposit of Yazd have been investigated within the microcontinent zone of Iran in the Zagros structural zone. The geological units in the Esfordi area belong to the pre-Cambrian to lower-Cambrian age, consisting of a succession of carbonate rocks (dolomite), shale, tuff, sandstone, and volcanic rocks. In addition to the mentioned sedimentary and volcanic rocks, the granitoid mass of Bahabad, which is the largest intrusive mass in the region, has intruded into the eastern part of this series and has caused its metamorphism and alteration. After collecting the available data, various samples of Esfordi’s apatite were prepared, and their mineralogy and crystallography were investigated using laboratory methods such as petrographic microscopy, Raman spectroscopy, EDS (Energy Dispersive Spectroscopy), and Scanning Electron Microscopy (SEM). In non-destructive Raman spectroscopy, the molecular structure of apatite minerals was revealed in four distinct spectral ranges. Initially, the spectra of phosphate and aluminum bonds with O2HO, OH, were observed, followed by the identification of Cl, OH, Al, Na, Ca and hydroxyl units depending on the type of apatite mineral family. In SEM analysis, based on various shapes and different phases of apatites, their constituent major elements were identified through EDS, indicating that the samples from the Esfordi mining area exhibit a dense and coherent texture with smooth surfaces. Based on the elemental analysis results by EDS, the apatites in the Esfordi area are classified into the calcic apatite group.
Keywords: Petrology, apatite, Esfordi, EDS, SEM, Scanning Electron Microscopy, Raman spectroscopy.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 161533 Application of Legendre Transformation to Portfolio Optimization
Authors: Peter Benneth, Tsaroh N. Theophilus, Prince Benjamin
Abstract:
This research work aims at studying the application of Legendre Transformation Method (LTM) to Hamilton Jacobi Bellman (HJB) equation which is an example of optimal control problem. We discuss the steps involved in modelling the HJB equation as it relates to mathematical finance by applying the Ito’s lemma and maximum principle theorem. By applying the LTM and dual theory, the resultant HJB equation is transformed to a linear Partial Differential Equation (PDE). Also, the Optimal Investment Strategy (OIS) and the optimal value function were obtained under the exponential utility function. Furthermore, some numerical results were also presented with observations that the OIS under exponential utility is directly proportional to the appreciation rate of the risky asset and inversely proportional to the instantaneous volatility, predetermined interest rate, risk averse coefficient. Finally, it was observed that the optimal fund size is an increasing function of the risk free interest rate. This result is consistent with some existing results.
Keywords: Legendre transformation method, Optimal investment strategy, Ito’s lemma, Hamilton Jacobi Bellman equation, Geometric Brownian motion, financial market.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 67532 HaskellFL: A Tool for Detecting Logical Errors in Haskell
Authors: Vanessa Vasconcelos, Mariza A. S. Bigonha
Abstract:
Understanding and using the functional paradigm is a challenge for many programmers. Looking for logical errors in code may take a lot of a developer’s time when a program grows in size. In order to facilitate both processes, this paper presents HaskellFL, a tool that uses fault localization techniques to locate a logical error in Haskell code. The Haskell subset used in this work is sufficiently expressive for those studying Functional Programming to get immediate help debugging their code and to answer questions about key concepts associated with the functional paradigm. HaskellFL was tested against Functional Programming assignments submitted by students enrolled at the Functional Programming class at the Federal University of Minas Gerais and against exercises from the Exercism Haskell track that are publicly available in GitHub. This work also evaluated the effectiveness of two fault localization techniques, Tarantula and Ochiai, in the Haskell context. Furthermore, the EXAM score was chosen to evaluate the tool’s effectiveness, and results showed that HaskellFL reduced the effort needed to locate an error for all tested scenarios. The results also showed that the Ochiai method was more effective than Tarantula.
Keywords: Debug, fault localization, functional programming, Haskell.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 731531 Finite Element Application to Estimate Inservice Material Properties using Miniature Specimen
Authors: G. Partheepan, D.K. Sehgal, R.K. Pandey
Abstract:
This paper presents a method for determining the uniaxial tensile properties such as Young-s modulus, yield strength and the flow behaviour of a material in a virtually non-destructive manner. To achieve this, a new dumb-bell shaped miniature specimen has been designed. This helps in avoiding the removal of large size material samples from the in-service component for the evaluation of current material properties. The proposed miniature specimen has an advantage in finite element modelling with respect to computational time and memory space. Test fixtures have been developed to enable the tension tests on the miniature specimen in a testing machine. The studies have been conducted in a chromium (H11) steel and an aluminum alloy (AR66). The output from the miniature test viz. load-elongation diagram is obtained and the finite element simulation of the test is carried out using a 2D plane stress analysis. The results are compared with the experimental results. It is observed that the results from the finite element simulation corroborate well with the miniature test results. The approach seems to have potential to predict the mechanical properties of the materials, which could be used in remaining life estimation of the various in-service structures.Keywords: ABAQUS, finite element, miniature test, tensileproperties
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1730530 Distributed Estimation Using an Improved Incremental Distributed LMS Algorithm
Authors: Amir Rastegarnia, Mohammad Ali Tinati, Azam Khalili
Abstract:
In this paper we consider the problem of distributed adaptive estimation in wireless sensor networks for two different observation noise conditions. In the first case, we assume that there are some sensors with high observation noise variance (noisy sensors) in the network. In the second case, different variance for observation noise is assumed among the sensors which is more close to real scenario. In both cases, an initial estimate of each sensor-s observation noise is obtained. For the first case, we show that when there are such sensors in the network, the performance of conventional distributed adaptive estimation algorithms such as incremental distributed least mean square (IDLMS) algorithm drastically decreases. In addition, detecting and ignoring these sensors leads to a better performance in a sense of estimation. In the next step, we propose a simple algorithm to detect theses noisy sensors and modify the IDLMS algorithm to deal with noisy sensors. For the second case, we propose a new algorithm in which the step-size parameter is adjusted for each sensor according to its observation noise variance. As the simulation results show, the proposed methods outperforms the IDLMS algorithm in the same condition.
Keywords: Distributes estimation, sensor networks, adaptive filter, IDLMS.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1446529 A Hybrid Multi Objective Algorithm for Flexible Job Shop Scheduling
Authors: Parviz Fattahi
Abstract:
Scheduling for the flexible job shop is very important in both fields of production management and combinatorial optimization. However, it quit difficult to achieve an optimal solution to this problem with traditional optimization approaches owing to the high computational complexity. The combining of several optimization criteria induces additional complexity and new problems. In this paper, a Pareto approach to solve the multi objective flexible job shop scheduling problems is proposed. The objectives considered are to minimize the overall completion time (makespan) and total weighted tardiness (TWT). An effective simulated annealing algorithm based on the proposed approach is presented to solve multi objective flexible job shop scheduling problem. An external memory of non-dominated solutions is considered to save and update the non-dominated solutions during the solution process. Numerical examples are used to evaluate and study the performance of the proposed algorithm. The proposed algorithm can be applied easily in real factory conditions and for large size problems. It should thus be useful to both practitioners and researchers.Keywords: Flexible job shop, Scheduling, Hierarchical approach, simulated annealing, tabu search, multi objective.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2010528 Performance Analysis of Software Reliability Models using Matrix Method
Authors: RajPal Garg, Kapil Sharma, Rajive Kumar, R. K. Garg
Abstract:
This paper presents a computational methodology based on matrix operations for a computer based solution to the problem of performance analysis of software reliability models (SRMs). A set of seven comparison criteria have been formulated to rank various non-homogenous Poisson process software reliability models proposed during the past 30 years to estimate software reliability measures such as the number of remaining faults, software failure rate, and software reliability. Selection of optimal SRM for use in a particular case has been an area of interest for researchers in the field of software reliability. Tools and techniques for software reliability model selection found in the literature cannot be used with high level of confidence as they use a limited number of model selection criteria. A real data set of middle size software project from published papers has been used for demonstration of matrix method. The result of this study will be a ranking of SRMs based on the Permanent value of the criteria matrix formed for each model based on the comparison criteria. The software reliability model with highest value of the Permanent is ranked at number – 1 and so on.Keywords: Matrix method, Model ranking, Model selection, Model selection criteria, Software reliability models.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2319527 A Digital Pulse-Width Modulation Controller for High-Temperature DC-DC Power Conversion Application
Authors: Jingjing Lan, Jun Yu, Muthukumaraswamy Annamalai Arasu
Abstract:
This paper presents a digital non-linear pulse-width modulation (PWM) controller in a high-voltage (HV) buck-boost DC-DC converter for the piezoelectric transducer of the down-hole acoustic telemetry system. The proposed design controls the generation of output signal with voltage higher than the supply voltage and is targeted to work under high temperature. To minimize the power consumption and silicon area, a simple and efficient design scheme is employed to develop the PWM controller. The proposed PWM controller consists of serial to parallel (S2P) converter, data assign block, a mode and duty cycle controller (MDC), linearly PWM (LPWM) and noise shaper, pulse generator and clock generator. To improve the reliability of circuit operation at higher temperature, this design is fabricated with the 1.0-μm silicon-on-insulator (SOI) CMOS process. The implementation results validated that the proposed design has the advantages of smaller size, lower power consumption and robust thermal stability.
Keywords: DC-DC power conversion, digital control, high temperatures, pulse-width modulation.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1711526 Deoxygenation of Beef Fat over Pd Supported Mesoporous TiO2 Catalyst Prepared by Single-Step Sol-Gel Process with Surfactant Template
Authors: Tossaporn Jindarat, Siriporn Jongpatiwut, Somchai Osuwan, Suchada Butnark
Abstract:
Deoxygenation of beef fat for the production of hydrogenated biodiesel is investigated in a high pressure continuous flow fixed bed reactor over palladium-supported mesoporous titania catalyst synthesized via a combined single-step sol-gel process with surfactant-assisted templating method (SATM). The catalyst possessed a mesoporous charactheristic with high surface area and narrow pore size distribution. The main products of all Pd/TiO2 catalysts are n-heptadecane (n-C17) and n-pentadecane (n-C15) resulting from decarbonylation reaction. Pd/TiO2 catalyst synthesized via a combined single-step sol-gel process with SATM (SSSG) gave higher activity and selectivity to the desired products when compared to IWI/SG-TiO2 and IWI/P25-TiO2, respectively. SSSG catalyst gave the average conversion up to 80-90 % and 80 % for the selectivity in diesel range hydrocarbons. This result may cause by the higher surface area and the ability in dispersion of palladium ion in mesoporous of TiO2 during sol-gel process.
Keywords: Beef fat, Deoxygenation, Hydrogenated biodiesel, Pd/TiO2
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1444525 Advanced Neural Network Learning Applied to Pulping Modeling
Authors: Z. Zainuddin, W. D. Wan Rosli, R. Lanouette, S. Sathasivam
Abstract:
This paper reports work done to improve the modeling of complex processes when only small experimental data sets are available. Neural networks are used to capture the nonlinear underlying phenomena contained in the data set and to partly eliminate the burden of having to specify completely the structure of the model. Two different types of neural networks were used for the application of pulping problem. A three layer feed forward neural networks, using the Preconditioned Conjugate Gradient (PCG) methods were used in this investigation. Preconditioning is a method to improve convergence by lowering the condition number and increasing the eigenvalues clustering. The idea is to solve the modified odified problem M-1 Ax= M-1b where M is a positive-definite preconditioner that is closely related to A. We mainly focused on Preconditioned Conjugate Gradient- based training methods which originated from optimization theory, namely Preconditioned Conjugate Gradient with Fletcher-Reeves Update (PCGF), Preconditioned Conjugate Gradient with Polak-Ribiere Update (PCGP) and Preconditioned Conjugate Gradient with Powell-Beale Restarts (PCGB). The behavior of the PCG methods in the simulations proved to be robust against phenomenon such as oscillations due to large step size.
Keywords: Convergence, pulping modeling, neural networks, preconditioned conjugate gradient.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1409524 Clogging Reduction Design Factor for Geosynthetics Used in Sustainable Urban Drainage Systems and Roads
Authors: Jaime Carpio-García, Elena Blanco-Fernández, Javier González-Fernández, Daniel Castro-Fresno
Abstract:
Sustainable urban drainage systems (SUDS) are more often used in order to prevent floods, water treatment, fight against pollution, urban heat island effect, and global warming in applications like green roofs, permeable pavements, and others. Furthermore, geosynthetics are also worldwide used as a part of drainage systems in road construction. Geotextiles are an essential part of both, and one of the main geotextile properties in those applications is permeability, whose behavior is not well established along its service life. In this paper, clogging reduction design factors for an estimated service life of 25 years are experimentally obtained for five different geotextiles used in SUDS and roads combined with two different soils and with two pollutants, motor oil, and lime, in order to evaluate chemical clogging, too. The effect of characteristic opening size and other characteristics of the geosynthetics are also discussed in order to give civil engineers, together with the clogging reduction factors, a better long-time design of geotextiles used in their SUDS and roads.
Keywords: Geotextiles, drainage, clogging, reduction factor.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 21523 The Relationship between Human Values and Service Quality with the Mediating Role of Motivation: A Quantitative Study on Malaysian Commercial Banks
Authors: Kuppusamy Singaravelloo, Zahra Kamiab
Abstract:
This study explores the mediating effects of motivation in the relationship between human values and service quality. To examine the fundamental relationships among human values, motivation, and service quality, a Structural Equation Model (SEM) with a full mediation model was adopted. The model was designed and subsequently analyzed by utilizing the Partial Least Squares (PLS) procedure on data collected from a survey that yielded 936 usable questionnaires. The survey was sent to all 117 branches of two local commercial banks (CIMB and Maybank) operating in Kuala Lumpur, Malaysia. Both banks ranked the highest in terms of asset size and market capitalization. The mediating role of motivation was examined in the relationship between four dimensions of human values and bank service quality, whereby human values initiate service quality. The results support the notion that employee motivation fully mediates the relationship between self-enhancement values and service quality in commercial banks. These results demonstrate the unpredictable role of the interaction structures of human values on bank service quality.
Keywords: Human values, service quality, mediation, motivation, commercial bank.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 878522 Photocatalytic Degradation of Organic Pollutant Reacting with Tungstates: Role of Microstructure and Size Effect on Oxidation Kinetics
Authors: A. Taoufyq, B. Bakiz, A. Benlhachemi, L. Patout, D. V. Chokouadeua, F. Guinneton, G. Nolibe, A. Lyoussi, J-R. Gavarri
Abstract:
The aim of this study was to investigate the photocatalytic activity of polycrystalline phases of bismuth tungstate of formula Bi2WO6. Polycrystalline samples were elaborated using a coprecipitation technique followed by a calcination process at different temperatures (300, 400, 600 and 900°C). The obtained polycrystalline phases have been characterized by X-ray diffraction (XRD), scanning electron microscopy (SEM), and transmission electron microscopy (TEM). Crystal cell parameters and cell volume depend on elaboration temperature. High-resolution electron microscopy images and image simulations, associated with X-ray diffraction data, allowed confirming the lattices and space groups Pca21. The photocatalytic activity of the as-prepared samples was studied by irradiating aqueous solutions of Rhodamine B, associated with Bi2WO6 additives having variable crystallite sizes. The photocatalytic activity of such bismuth tungstates increased as the crystallite sizes decreased. The high specific area of the photocatalytic particles obtained at 300°C seems to condition the degradation kinetics of RhB.Keywords: Bismuth tungstate, crystallite sizes, electron microscopy, photocatalytic activity, X-ray diffraction.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2070521 A Fragile Watermarking Scheme for Color Image Authentication
Authors: M. Hamad Hassan, S.A.M. Gilani
Abstract:
In this paper, a fragile watermarking scheme is proposed for color image specified object-s authentication. The color image is first transformed from RGB to YST color space, suitable for watermarking the color media. The T channel corresponds to the chrominance component of a color image andYS ÔèÑ T , therefore selected for embedding the watermark. The T channel is first divided into 2×2 non-overlapping blocks and the two LSBs are set to zero. The object that is to be authenticated is also divided into 2×2 nonoverlapping blocks and each block-s intensity mean is computed followed by eight bit encoding. The generated watermark is then embedded into T channel randomly selected 2×2 block-s LSBs using 2D-Torus Automorphism. Selection of block size is paramount for exact localization and recovery of work. The proposed scheme is blind, efficient and secure with ability to detect and locate even minor tampering applied to the image with full recovery of original work. The quality of watermarked media is quite high both subjectively and objectively. The technique is suitable for class of images with format such as gif, tif or bitmap.
Keywords: Image Authentication, LSBs, PSNR, 2D-Torus Automorphism, YST Color Space.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1888520 The Relationship between Representational Conflicts, Generalization, and Encoding Requirements in an Instance Memory Network
Authors: Mathew Wakefield, Matthew Mitchell, Lisa Wise, Christopher McCarthy
Abstract:
This paper aims to provide an interpretation of artificial neural networks (ANNs) and explore some of its implications. The interpretation views ANNs as a memory which encodes instances of experience. An experiment explores the behavior of encoding and retrieval of instances from memory. A localised representation ANN is created that allows control over encoding and retrieved memory sample size and is experimented with using the MNIST digits dataset. The relationship between input familiarity, conflict within retrieved samples, and error rates is described and demonstrated to be an effective driver for memory encoding. Results indicate that selective encoding and retrieval samples that allow detection of memory conflicts produce optimal performance, and that error rates are normally distributed with input familiarity and conflict. By using input familiarity and sample consistency to guide memory encoding, the number of encoding trials on the dataset were reduced to 18.33% of the training data while maintaining good recognition performance on the test data.
Keywords: Artificial Neural Networks, ANNs, representation, memory, conflict monitoring, confidence.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 507519 Mechanical Characterization and Impact Study on the Environment of Raw Sediments and Sediments Dehydrated by Addition of Polymer
Authors: A. Kasmi, N. E. Abriak, M. Benzerzour, I. Shahrour
Abstract:
Large volumes of river sediments are dredged each year in Europe in order to maintain harbour activities and prevent floods. The management of this sediment has become increasingly complex. Several European projects were implemented to find environmentally sound solutions for these materials. The main objective of this study is to show the ability of river sediment to be used in road. Since sediments contain a high amount of water, then a dehydrating treatment by addition of the flocculation aid has been used. Firstly, a lot of physical characteristics are measured and discussed for a better identification of the raw sediment and this dehydrated sediment by addition the flocculation aid. The identified parameters are, for example, the initial water content, the density, the organic matter content, the grain size distribution, the liquid limit and plastic limit and geotechnical parameters. The environmental impacts of the used material were evaluated. The results obtained show that there is a slight change on the physical-chemical and geotechnical characteristics of sediment after dehydration by the addition of polymer. However, these sediments cannot be used in road construction.
Keywords: River sediment, dehydration, flocculation aid, characteristics, environmental impacts, road construction.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1298518 Formulation of Extended-Release Gliclazide Tablet Using a Mathematical Model for Estimation of Hypromellose
Authors: Farzad Khajavi, Farzaneh Jalilfar, Faranak Jafari, Leila Shokrani
Abstract:
Formulation of gliclazide in the form of extended-release tablet in 30 and 60 mg dosage forms was performed using hypromellose (HPMC K4M) as a retarding agent. Drug-release profiles were investigated in comparison with references Diamicron MR 30 and 60 mg tablets. The effect of size of powder particles, the amount of hypromellose in formulation, hardness of tablets, and also the effect of halving the tablets were investigated on drug release profile. A mathematical model which describes hypromellose behavior in initial times of drug release was proposed for the estimation of hypromellose content in modified-release gliclazide 60 mg tablet. This model is based on erosion of hypromellose in dissolution media. The model is applicable to describe release profiles of insoluble drugs. Therefore, by using dissolved amount of drug in initial times of dissolution and the model, the amount of hypromellose in formulation can be predictable. The model was used to predict the HPMC K4M content in modified-release gliclazide 30 mg and extended-release quetiapine 200 mg tablets.
Keywords: Hypromellose, gliclazide, drug release, modified-release tablet, mathematical model.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2304517 Efficient Tuning Parameter Selection by Cross-Validated Score in High Dimensional Models
Authors: Yoonsuh Jung
Abstract:
As DNA microarray data contain relatively small sample size compared to the number of genes, high dimensional models are often employed. In high dimensional models, the selection of tuning parameter (or, penalty parameter) is often one of the crucial parts of the modeling. Cross-validation is one of the most common methods for the tuning parameter selection, which selects a parameter value with the smallest cross-validated score. However, selecting a single value as an ‘optimal’ value for the parameter can be very unstable due to the sampling variation since the sample sizes of microarray data are often small. Our approach is to choose multiple candidates of tuning parameter first, then average the candidates with different weights depending on their performance. The additional step of estimating the weights and averaging the candidates rarely increase the computational cost, while it can considerably improve the traditional cross-validation. We show that the selected value from the suggested methods often lead to stable parameter selection as well as improved detection of significant genetic variables compared to the tradition cross-validation via real data and simulated data sets.Keywords: Cross Validation, Parameter Averaging, Parameter Selection, Regularization Parameter Search.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1572516 Wind-tunnel Measurement of the Drag-reducing Effect of Compliant Coating
Authors: Inwon Lee, Victor M. Kulik, Andrey V. Boiko, Ho Hwan Chun
Abstract:
A specially designed flat plate was mounted vertically over the axial line in the wind tunnel of the Aerospace Department of the Pusan National University. The plate is 2 m long, 0.8 m high and 8 cm thick. The measurements were performed in velocity range from 15 to 60 m/s. A sand paper turbulizer was placed close to the plate nose to provide fully developed turbulent boundary layer over the most part of the plate. Strain balances were mounted in the trailing part of the plate to measure the skin friction drag over removable insertions of 0.55×0.25m2 size. A set of the insertions was designed and manufactured: 3mm thick polished metal surface and three compliant surfaces. The compliant surfaces were manufactured of a silicone rubber Silastic® S2 (Dow Corning company). To modify the viscoelastic properties of the rubber, its composition was varied: 90% of the rubber + 10% catalyst (standard), 92.5% + 7.5% (weak), 85% + 15% (strong). Modulus of elasticity and the loss tangent were measured accurately for these materials in the frequency range from 40 Hz to 3 KHz using the unique proposed technique.Keywords: boundary layer, compliant coating, drag reduction, hot wire, wind tunnel.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1686515 Corporate Governance Mechanisms, Whistle-Blowing Policy and Earnings Management Practices of Firms in Malaysia
Authors: Mujeeb Saif Mohsen Al-Absy, Ku Nor Izah Ku Ismail, Sitraselvi Chandren
Abstract:
This study examines whether corporate governance (CG) mechanisms in firms that have a whistle-blowing policy (WHBLP) are more effective in constraining earnings management (EM), than those without. A sample of 288 Malaysian firms for the years 2013 to 2015, amounting to 864 firm-years were grouped into firms with and without WHBLP. Results show that for firms without WHBLP, the board chairman tenure would minimize EM activities. Meanwhile, for firms with WHBLP, board chairman independence, board chairman tenure, audit committee size, audit committee meeting and women in the audit committees are found to be associated with less EM activities. Further, it is found that ownership concentration and Big 4 auditing firms help to reduce EM activities in firms with WHBLP, while not in firms without WHBLP. Hence, functional and effective governance can be achieved by having a WHBLP, which is in line with agency and resource dependent theories. Therefore, this study suggests that firms should have a WHBLP in place, and policymakers should come up with enhanced criteria to strengthen the mechanisms of WHBLP.
Keywords: Corporate governance, earnings management, whistle-blowing policy, audit committee, board of directors.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1270514 Circular Patch Microstrip Array Antenna for KU-band
Authors: T.F.Lai, Wan Nor Liza Mahadi, Norhayati Soin
Abstract:
This paper present a circular patch microstrip array antenna operate in KU-band (10.9GHz – 17.25GHz). The proposed circular patch array antenna will be in light weight, flexible, slim and compact unit compare with current antenna used in KU-band. The paper also presents the detail steps of designing the circular patch microstrip array antenna. An Advance Design System (ADS) software is used to compute the gain, power, radiation pattern, and S11 of the antenna. The proposed Circular patch microstrip array antenna basically is a phased array consisting of 'n' elements (circular patch antennas) arranged in a rectangular grid. The size of each element is determined by the operating frequency. The incident wave from satellite arrives at the plane of the antenna with equal phase across the surface of the array. Each 'n' element receives a small amount of power in phase with the others. There are feed network connects each element to the microstrip lines with an equal length, thus the signals reaching the circular patches are all combined in phase and the voltages add up. The significant difference of the circular patch array antenna is not come in the phase across the surface but in the magnitude distribution.
Keywords: Circular patch microstrip array antenna, gain, radiation pattern, S-Parameter.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3114513 Empirical Process Monitoring Via Chemometric Analysis of Partially Unbalanced Data
Authors: Hyun-Woo Cho
Abstract:
Real-time or in-line process monitoring frameworks are designed to give early warnings for a fault along with meaningful identification of its assignable causes. In artificial intelligence and machine learning fields of pattern recognition various promising approaches have been proposed such as kernel-based nonlinear machine learning techniques. This work presents a kernel-based empirical monitoring scheme for batch type production processes with small sample size problem of partially unbalanced data. Measurement data of normal operations are easy to collect whilst special events or faults data are difficult to collect. In such situations, noise filtering techniques can be helpful in enhancing process monitoring performance. Furthermore, preprocessing of raw process data is used to get rid of unwanted variation of data. The performance of the monitoring scheme was demonstrated using three-dimensional batch data. The results showed that the monitoring performance was improved significantly in terms of detection success rate of process fault.
Keywords: Process Monitoring, kernel methods, multivariate filtering, data-driven techniques, quality improvement.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1746512 Obtaining Constants of Johnson-Cook Material Model Using a Combined Experimental, Numerical Simulation and Optimization Method
Authors: F. Rahimi Dehgolan, M. Behzadi, J. Fathi Sola
Abstract:
In this article, the Johnson-Cook material model’s constants for structural steel ST.37 have been determined by a method which integrates experimental tests, numerical simulation, and optimization. In the first step, a quasi-static test was carried out on a plain specimen. Next, the constants were calculated for it by minimizing the difference between the results acquired from the experiment and numerical simulation. Then, a quasi-static tension test was performed on three notched specimens with different notch radii. At last, in order to verify the results, they were used in numerical simulation of notched specimens and it was observed that experimental and simulation results are in good agreement. Changing the diameter size of the plain specimen in the necking area was set as the objective function in the optimization step. For final validation of the proposed method, diameter variation was considered as a parameter and its sensitivity to a change in any of the model constants was examined and the results were completely corroborating.
Keywords: Constants, Johnson-Cook material model, notched specimens, quasi-static test, sensitivity.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3617511 Examination of Flood Runoff Reproductivity for Different Rainfall Sources in Central Vietnam
Authors: Do Hoai Nam, Keiko Udo, Akira Mano
Abstract:
This paper presents the combination of different precipitation data sets and the distributed hydrological model, in order to examine the flood runoff reproductivity of scattered observation catchments. The precipitation data sets were obtained from observation using rain-gages, satellite based estimate (TRMM), and numerical weather prediction model (NWP), then were coupled with the super tank model. The case study was conducted in three basins (small, medium, and large size) located in Central Vietnam. Calculated hydrographs based on ground observation rainfall showed best fit to measured stream flow, while those obtained from TRMM and NWP showed high uncertainty of peak discharges. However, calculated hydrographs using the adjusted rainfield depicted a promising alternative for the application of TRMM and NWP in flood modeling for scattered observation catchments, especially for the extension of forecast lead time.
Keywords: Flood forecast, rainfall-runoff model, satellite rainfall estimate, numerical weather prediction, quantitative precipitation forecasting.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1609510 A Comparison of Some Thresholding Selection Methods for Wavelet Regression
Authors: Alsaidi M. Altaher, Mohd T. Ismail
Abstract:
In wavelet regression, choosing threshold value is a crucial issue. A too large value cuts too many coefficients resulting in over smoothing. Conversely, a too small threshold value allows many coefficients to be included in reconstruction, giving a wiggly estimate which result in under smoothing. However, the proper choice of threshold can be considered as a careful balance of these principles. This paper gives a very brief introduction to some thresholding selection methods. These methods include: Universal, Sure, Ebays, Two fold cross validation and level dependent cross validation. A simulation study on a variety of sample sizes, test functions, signal-to-noise ratios is conducted to compare their numerical performances using three different noise structures. For Gaussian noise, EBayes outperforms in all cases for all used functions while Two fold cross validation provides the best results in the case of long tail noise. For large values of signal-to-noise ratios, level dependent cross validation works well under correlated noises case. As expected, increasing both sample size and level of signal to noise ratio, increases estimation efficiency.
Keywords: wavelet regression, simulation, Threshold.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1767509 OCR for Script Identification of Hindi (Devnagari) Numerals using Error Diffusion Halftoning Algorithm with Neural Classifier
Authors: Banashree N. P., Andhe Dharani, R. Vasanta, P. S. Satyanarayana
Abstract:
The applications on numbers are across-the-board that there is much scope for study. The chic of writing numbers is diverse and comes in a variety of form, size and fonts. Identification of Indian languages scripts is challenging problems. In Optical Character Recognition [OCR], machine printed or handwritten characters/numerals are recognized. There are plentiful approaches that deal with problem of detection of numerals/character depending on the sort of feature extracted and different way of extracting them. This paper proposes a recognition scheme for handwritten Hindi (devnagiri) numerals; most admired one in Indian subcontinent our work focused on a technique in feature extraction i.e. Local-based approach, a method using 16-segment display concept, which is extracted from halftoned images & Binary images of isolated numerals. These feature vectors are fed to neural classifier model that has been trained to recognize a Hindi numeral. The archetype of system has been tested on varieties of image of numerals. Experimentation result shows that recognition rate of halftoned images is 98 % compared to binary images (95%).
Keywords: OCR, Halftoning, Neural classifier, 16-segmentdisplay concept.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1716508 On Developing an Automatic Speech Recognition System for Standard Arabic Language
Authors: R. Walha, F. Drira, H. El-Abed, A. M. Alimi
Abstract:
The Automatic Speech Recognition (ASR) applied to Arabic language is a challenging task. This is mainly related to the language specificities which make the researchers facing multiple difficulties such as the insufficient linguistic resources and the very limited number of available transcribed Arabic speech corpora. In this paper, we are interested in the development of a HMM-based ASR system for Standard Arabic (SA) language. Our fundamental research goal is to select the most appropriate acoustic parameters describing each audio frame, acoustic models and speech recognition unit. To achieve this purpose, we analyze the effect of varying frame windowing (size and period), acoustic parameter number resulting from features extraction methods traditionally used in ASR, speech recognition unit, Gaussian number per HMM state and number of embedded re-estimations of the Baum-Welch Algorithm. To evaluate the proposed ASR system, a multi-speaker SA connected-digits corpus is collected, transcribed and used throughout all experiments. A further evaluation is conducted on a speaker-independent continue SA speech corpus. The phonemes recognition rate is 94.02% which is relatively high when comparing it with another ASR system evaluated on the same corpus.Keywords: ASR, HMM, acoustical analysis, acoustic modeling, Standard Arabic language
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1779507 Chikungunya Protease Domain–High Throughput Virtual Screening
Authors: Surender Singh Jadav, Venkatesan Jayaprakash, Arijit Basu, Barij Nayan Sinha
Abstract:
Chikungunya virus (CHICKV) is an arboviruses belonging to family Tagoviridae and is transmitted to human through by mosquito (Aedes aegypti and Aedes albopictus) bite. A large outbreak of chikungunya has been reported in India between 2006 and 2007, along with several other countries from South-East Asia and for the first time in Europe. It was for the first time that the CHICKV outbreak has been reported with mortality from Reunion Island and increased mortality from Asian countries. CHICKV affects all age groups, and currently there are no specific drugs or vaccine to cure the disease. The need of antiviral agents for the treatment of CHICKV infection and the success of virtual screening against many therapeutically valuable targets led us to carry out the structure based drug design against Chikungunya nSP2 protease (PDB: 3TRK). Highthroughput virtual screening of publicly available databases, ZINC12 and BindingDB, has been carried out using the Openeye tools and Schrodinger LLC software packages. Openeye Filter program has been used to filter the database and the filtered outputs were docked using HTVS protocol implemented in GLIDE package of Schrodinger LLC. The top HITS were further used for enriching the similar molecules from the database through vROCS; a shape based screening protocol implemented in Openeye. The approach adopted has provided different scaffolds as HITS against CHICKV protease. Three scaffolds: Indole, Pyrazole and Sulphone derivatives were selected based on the docking score and synthetic feasibility. Derivatives of Pyrazole were synthesized and submitted for antiviral screening against CHICKV.
Keywords: Chikungunya, nsP2 protease, ADME filter, HTVS, Docking, Active site.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2496506 A Simplified Approach for Load Flow Analysis of Radial Distribution Network
Authors: K. Vinoth Kumar, M.P. Selvan
Abstract:
This paper presents a simple approach for load flow analysis of a radial distribution network. The proposed approach utilizes forward and backward sweep algorithm based on Kirchoff-s current law (KCL) and Kirchoff-s voltage law (KVL) for evaluating the node voltages iteratively. In this approach, computation of branch current depends only on the current injected at the neighbouring node and the current in the adjacent branch. This approach starts from the end nodes of sub lateral line, lateral line and main line and moves towards the root node during branch current computation. The node voltage evaluation begins from the root node and moves towards the nodes located at the far end of the main, lateral and sub lateral lines. The proposed approach has been tested using four radial distribution systems of different size and configuration and found to be computationally efficient.Keywords: constant current load, constant impedance load, constant power load, forward–backward sweep, load flow analysis, radial distribution system.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2675505 Finite Element Analysis of Different Architectures for Bone Scaffold
Authors: Nimisha R. Shirbhate, Sanjay Bokade
Abstract:
Bone Scaffolds are fundamental architecture or a support structure that allows the regeneration of lost or damaged tissues and they are developed as a crucial tool in biomedical engineering. The structure of bone scaffolds plays an important role in treating bone defects. The shape of the bone scaffold performs a vital role, specifically pore size and shape, which help understand the behavior and strength of the scaffold. In this article, first, fundamental aspects of bone scaffold design are established. Second, the behavior of each architecture of the bone scaffold with biomaterials is discussed. Finally, for each structure, the stress analysis was carried out. This study aimed to design a porous and mechanically strong bone regeneration scaffold that can be successfully manufactured. Four porous architectures of the bone scaffold were designed using Rhinoceros solid modelling software. The structure model consisted of repeatable unit cells arranged in layers to fill the chosen scaffold volume. The mechanical behavior of used biocompatible material is studied with the help of ANSYS 19.2 software. It is also playing significant role to predict the strength of defined structures or 3 dimensional models.
Keywords: Bone scaffold, stress analysis, porous structure, static loading.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 534