Search results for: Heat transfer performance
296 Cyclic Behaviour of Wide Beam-Column Joints with Shear Strength Ratios of 1.0 and 1.7
Authors: Roy Y. C. Huang, J. S. Kuang, Hamdolah Behnam
Abstract:
Beam-column connections play an important role in the reinforced concrete moment resisting frame (RCMRF), which is one of the most commonly used structural systems around the world. The premature failure of such connections would severely limit the seismic performance and increase the vulnerability of RCMRF. In the past decades, researchers primarily focused on investigating the structural behaviour and failure mechanisms of conventional beam-column joints, the beam width of which is either smaller than or equal to the column width, while studies in wide beam-column joints were scarce. This paper presents the preliminary experimental results of two full-scale exterior wide beam-column connections, which are mainly designed and detailed according to ACI 318-14 and ACI 352R-02, under reversed cyclic loading. The ratios of the design shear force to the nominal shear strength of these specimens are 1.0 and 1.7, respectively, so as to probe into differences of the joint shear strength between experimental results and predictions by design codes of practice. Flexural failure dominated in the specimen with ratio of 1.0 in which full-width plastic hinges were observed, while both beam hinges and post-peak joint shear failure occurred for the other specimen. No sign of premature joint shear failure was found which is inconsistent with ACI codes’ prediction. Finally, a modification of current codes of practice is provided to accurately predict the joint shear strength in wide beam-column joint.
Keywords: Joint shear strength, reversed cyclic loading, seismic codes, wide beam-column joints.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1071295 Analyzing the Effect of Materials’ Selection on Energy Saving and Carbon Footprint: A Case Study Simulation of Concrete Structure Building
Authors: M. Kouhirostamkolaei, M. Kouhirostami, M. Sam, J. Woo, A. T. Asutosh, J. Li, C. Kibert
Abstract:
Construction is one of the most energy consumed activities in the urban environment that results in a significant amount of greenhouse gas emissions around the world. Thus, the impact of the construction industry on global warming is undeniable. Thus, reducing building energy consumption and mitigating carbon production can slow the rate of global warming. The purpose of this study is to determine the amount of energy consumption and carbon dioxide production during the operation phase and the impact of using new shells on energy saving and carbon footprint. Therefore, a residential building with a re-enforced concrete structure is selected in Babolsar, Iran. DesignBuilder software has been used for one year of building operation to calculate the amount of carbon dioxide production and energy consumption in the operation phase of the building. The primary results show the building use 61750 kWh of energy each year. Computer simulation analyzes the effect of changing building shells -using XPS polystyrene and new electrochromic windows- as well as changing the type of lighting on energy consumption reduction and subsequent carbon dioxide production. The results show that the amount of energy and carbon production during building operation has been reduced by approximately 70% by applying the proposed changes. The changes reduce CO2e to 11345 kg CO2/yr. The result of this study helps designers and engineers to consider material selection’s process as one of the most important stages of design for improving energy performance of buildings.
Keywords: Construction materials, green construction, energy simulation, carbon footprint, energy saving, concrete structure, DesignBuilder.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 992294 The Impact of Surface Roughness and PTFE/TiF3/FeF3 Additives in Plain ZDDP Oil on the Friction and Wear Behavior Using Thermal and Tribological Analysis under Extreme Pressure Condition
Authors: Gabi N. Nehme, Saeed Ghalambor
Abstract:
The use of titanium fluoride and iron fluoride (TiF3/FeF3) catalysts in combination with polutetrafluoroethylene (PTFE) in plain zinc- dialkyldithiophosphate (ZDDP) oil is important for the study of engine tribocomponents and is increasingly a strategy to improve the formation of tribofilm and provide low friction and excellent wear protection in reduced phosphorus plain ZDDP oil. The influence of surface roughness and the concentration of TiF3/FeF3/PTFE were investigated using bearing steel samples dipped in lubricant solution at 100°C for two different heating time durations. This paper addresses the effects of water drop contact angle using different surface; finishes after treating them with different lubricant combination. The calculated water drop contact angles were analyzed using Design of Experiment software (DOE) and it was determined that a 0.05 μm Ra surface roughness would provide an excellent TiF3/FeF3/PTFE coating for antiwear resistance as reflected in the Scanning electron microscopy (SEM) images and the tribological testing under extreme pressure conditions. Both friction and wear performance depend greatly on the PTFE/and catalysts in plain ZDDP oil with 0.05 % phosphorous and on the surface finish of bearing steel. The friction and wear reducing effects, which was observed in the tribological tests, indicated a better micro lubrication effect of the 0.05 μm Ra surface roughness treated at 100°C for 24 hours when compared to the 0.1 μm Ra surface roughness with the same treatment.
Keywords: Scanning Electron Microscopy (SEM), ZDDP, catalysts, PTFE, friction, wear.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1628293 Prediction Modeling of Alzheimer’s Disease and Its Prodromal Stages from Multimodal Data with Missing Values
Authors: M. Aghili, S. Tabarestani, C. Freytes, M. Shojaie, M. Cabrerizo, A. Barreto, N. Rishe, R. E. Curiel, D. Loewenstein, R. Duara, M. Adjouadi
Abstract:
A major challenge in medical studies, especially those that are longitudinal, is the problem of missing measurements which hinders the effective application of many machine learning algorithms. Furthermore, recent Alzheimer's Disease studies have focused on the delineation of Early Mild Cognitive Impairment (EMCI) and Late Mild Cognitive Impairment (LMCI) from cognitively normal controls (CN) which is essential for developing effective and early treatment methods. To address the aforementioned challenges, this paper explores the potential of using the eXtreme Gradient Boosting (XGBoost) algorithm in handling missing values in multiclass classification. We seek a generalized classification scheme where all prodromal stages of the disease are considered simultaneously in the classification and decision-making processes. Given the large number of subjects (1631) included in this study and in the presence of almost 28% missing values, we investigated the performance of XGBoost on the classification of the four classes of AD, NC, EMCI, and LMCI. Using 10-fold cross validation technique, XGBoost is shown to outperform other state-of-the-art classification algorithms by 3% in terms of accuracy and F-score. Our model achieved an accuracy of 80.52%, a precision of 80.62% and recall of 80.51%, supporting the more natural and promising multiclass classification.
Keywords: eXtreme Gradient Boosting, missing data, Alzheimer disease, early mild cognitive impairment, late mild cognitive impairment, multiclass classification, ADNI, support vector machine, random forest.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 958292 Invasion of Pectinatella magnifica in Freshwater Resources of the Czech Republic
Authors: J. Pazourek, K. Šmejkal, P. Kollár, J. Rajchard, J. Šinko, Z. Balounová, E. Vlková, H. Salmonová
Abstract:
Pectinatella magnifica (Leidy, 1851) is an invasive freshwater animal that lives in colonies. A colony of Pectinatella magnifica (a gelatinous blob) can be up to several feet in diameter large and under favorable conditions it exhibits an extreme growth rate. Recently European countries around rivers of Elbe, Oder, Danube, Rhine and Vltava have confirmed invasion of Pectinatella magnifica, including freshwater reservoirs in South Bohemia (Czech Republic). Our project (Czech Science Foundation, GAČR P503/12/0337) is focused onto biology and chemistry of Pectinatella magnifica. We monitor the organism occurrence in selected South Bohemia ponds and sandpits during the last years, collecting information about physical properties of surrounding water, and sampling the colonies for various analyses (classification, maps of secondary metabolites, toxicity tests). Because the gelatinous matrix is during the colony lifetime also a host for algae, bacteria and cyanobacteria (co-habitants), in this contribution, we also applied a high performance liquid chromatography (HPLC) method for determination of potentially present cyanobacterial toxins (microcystin-LR, microcystin-RR, nodularin). Results from the last 3-year monitoring show that these toxins are under limit of detection (LOD), so that they do not represent a danger yet. The final goal of our study is to assess toxicity risks related to fresh water resources invaded by Pectinatella magnifica, and to understand the process of invasion, which can enable to control it.
Keywords: Cyanobacteria, freshwater resources, Pectinatella magnifica invasion, toxicity monitoring.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1876291 Study of Pipes Scaling of Purified Wastewater Intended for the Irrigation of Agadir Golf Grass
Authors: A. Driouiche, S. Mohareb, A. Hadfi
Abstract:
In Morocco’s Agadir region, the reuse of treated wastewater for irrigation of green spaces has faced the problem of scaling of the pipes of these waters. This research paper aims at studying the phenomenon of scaling caused by the treated wastewater from the Mzar sewage treatment plant. These waters are used in the irrigation of golf turf for the Ocean Golf Resort. Ocean Golf, located about 10 km from the center of the city of Agadir, is one of the most important recreation centers in Morocco. The course is a Belt Collins design with 27 holes, and is quite open with deep challenging bunkers. The formation of solid deposits in the irrigation systems has led to a decrease in their lifetime and, consequently, a loss of load and performance. Thus, the sprinklers used in golf turf irrigation are plugged in the first weeks of operation. To study this phenomenon, the wastewater used for the irrigation of the golf turf was taken and analyzed at various points, and also samples of scale formed in the circuits of the passage of these waters were characterized. This characterization of the scale was performed by X-ray fluorescence spectrometry, X-ray diffraction (XRD), thermogravimetric analysis (TGA), differential thermal analysis (DTA), and scanning electron microscopy (SEM). The results of the physicochemical analysis of the waters show that they are full of bicarbonates (653 mg/L), chloride (478 mg/L), nitrate (412 mg/L), sodium (425 mg/L) and calcium (199mg/L). Their pH is slightly alkaline. The analysis of the scale reveals that it is rich in calcium and phosphorus. It is formed of calcium carbonate (CaCO₃), silica (SiO₂), calcium silicate (Ca₂SiO₄), hydroxylapatite (Ca₁₀P₆O₂₆), calcium carbonate and phosphate (Ca₁₀(PO₄) 6CO₃) and silicate calcium and magnesium (Ca₅MgSi₃O₁₂).
Keywords: Agadir, irrigation, scaling water, wastewater.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 723290 Effect of Mixing Process on Polypropylene Modified Bituminous Concrete Mix Properties
Authors: Noor Zainab Habib, Ibrahim Kamaruddin, Madzalan Napiah, Isa Mohd Tan
Abstract:
This paper presents a research conducted to investigate the effect of mixing process on polypropylene (PP) modified bitumen mixed with well graded aggregate to form modified bituminous concrete mix. Two mode of mixing, namely dry and wet with different concentration of polymer polypropylene was used with 80/100 pen bitumen, to evaluate the bituminous concrete mix properties. Three percentages of polymer varying from 1-3% by the weight of bitumen was used in this study. Three mixes namely control mix, wet mix and dry mix were prepared. Optimum binder content was calculated considering Marshall Stability, flow, air voids and Marshall Quotient at different bitumen content varying from 4% - 6.5% for control, dry and wet mix. Engineering properties thus obtained at the calculated optimum bitumen content revealed that wet mixing process is advantageous in comparison to dry mixing as it increases the stiffness of the mixture with the increase in polymer content in bitumen. Stiffness value for wet mix increases with the increase in polymer content which is beneficial in terms of rutting. 1% PP dry mix also shows enhanced stiffness, with the air void content limited to 4%.The flow behaviour of dry mix doesn't indicate any major difference with the increase in polymer content revealing that polymer acting as an aggregate only without affecting the viscosity of the binder in the mix. Polypropylene (PP) when interacted with 80 pen base bitumen enhances its performance characteristics which were brought about by altered rheological properties of the modified bitumen. The decrease in flow with the increase in binder content reflects the increase in viscosity of binder which induces the plastic flow in the mix. Workability index indicates that wet mix were easy to compact up to desired void ratio in comparison to dry mix samples.
Keywords: Marshall Flow, Marshall Stability, Polymer modified bitumen, Polypropylene, Stiffness.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 4497289 Investigating the Formation of Nano-Hydroxyapatite on a Biocompatible and Antibacterial Cu/Mg-Substituted Bioglass
Authors: Elhamalsadat Ghaffari, Moghan Amirhosseinian, Amir Khaleghipour
Abstract:
Multifunctional bioactive glasses (BGs) are designed with a focus on the provision of bactericidal and biological properties desired for angiogenesis, osteogenesis, and ultimately potential applications in bone tissue engineering. To achieve these, six sol-gel copper/magnesium substituted derivatives of 58S-BG, i.e. a mol% series of 60SiO2-4P2O5-5CuO-(31-x) CaO/xMgO (where x=0, 1, 3, 5, 8, and 10), were synthesized. Afterwards, the effect of MgO/CaO substitution on the in vitro formation of nano-hydroxyapatite (HA), osteoblast-like cell responses and BGs antibacterial performance were studied. During the BGs synthesis, the elimination of nitrates was achieved at 700 °C that prevented the BGs crystallization and stabilized the obtained dried gels. The structural and morphological evaluations were performed with X-ray diffraction (XRD), Fourier transform infrared spectroscopy (FTIR), and scanning electron microscopy (SEM). These characterizations revealed that Cu-substituted 58S-BG consisting of 5 mol% MgO (BG-5/5) slightly had retarded the formation of HA. In addition, Cu-substituted 58S-BGs consisting 8 mol% and 10 mol% MgO (BG-5/8 and BG-5/10) displayed lower bioactivity probably due to the lower ion release rate of Ca–Si into the simulated body fluid (SBF). The determination of 3-(4, 5 dimethylthiazol-2-yl)-2, 5-diphenyltetrazolium bromide (MTT) and alkaline phosphate (ALP) activities proved that the highest values of both differentiation and proliferation of MC3T3-E1 cells can be obtained from a 5 mol% MgO substituted BG, while the over addition of MgO (8 mol% and 10 mol%) decreased the bioactivity. Furthermore, these novel Cu/Mg-substituted 58S-BGs displayed antibacterial effect against methicillin-resistant Staphylococcus aureus bacteria. Taken together, the results suggest the equally-substituted BG-5/5 (i.e. the one consists of 5 mol% of both CuO and MgO) as a promising candidate for bone tissue engineering, among all newly designed BGs in this work, owing to its desirable cell proliferation, ALP activity and antibacterial properties.Keywords: Apatite, bioactivity, biomedical applications sol-gel processes.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 832288 Inquiry on the Improvement Teaching Quality in the Classroom with Meta-Teaching Skills
Authors: Shahlan Surat, Saemah Rahman, Saadiah Kummin
Abstract:
When teachers reflect and evaluate whether their teaching methods actually have an impact on students’ learning, they will adjust their practices accordingly. This inevitably improves their students’ learning and performance. The approach in meta-teaching can invigorate and create a passion for teaching. It thus helps to increase the commitment and love for the teaching profession. This study was conducted to determine the level of metacognitive thinking of teachers in the process of teaching and learning in the classroom. Metacognitive thinking teachers include the use of metacognitive knowledge which consists of different types of knowledge: declarative, procedural and conditional. The ability of the teachers to plan, monitor and evaluate the teaching process can also be determined. This study was conducted on 377 graduate teachers in Klang Valley, Malaysia. The stratified sampling method was selected for the purpose of this study. The metacognitive teaching inventory consisting of 24 items is called InKePMG (Teacher Indicators of Effectiveness Meta-Teaching). The results showed the level of mean is high for two components of metacognitive knowledge; declarative knowledge (mean = 4.16) and conditional (mean = 4.11) whereas, the mean of procedural knowledge is 4.00 (moderately high). Similarly, the level of knowledge in monitoring (mean = 4.11), evaluating (mean = 4.00) which indicate high score and planning (mean = 4.00) are moderately high score among teachers. In conclusion, this study shows that the planning and procedural knowledge is an important element in improving the quality of teachers teaching in the classroom. Thus, the researcher recommended that further studies should focus on training programs for teachers on metacognitive skills and also on developing creative thinking among teachers.Keywords: Metacognitive thinking skills, procedural knowledge, conditional knowledge, declarative knowledge, meta-teaching and regulation of cognitive.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1436287 Multiaxial Fatigue Analysis of a High Performance Nickel-Based Superalloy
Authors: P. Selva, B. Lorrain, J. Alexis, A. Seror, A. Longuet, C. Mary, F. Denard
Abstract:
Over the past four decades, the fatigue behavior of nickel-based alloys has been widely studied. However, in recent years, significant advances in the fabrication process leading to grain size reduction have been made in order to improve fatigue properties of aircraft turbine discs. Indeed, a change in particle size affects the initiation mode of fatigue cracks as well as the fatigue life of the material. The present study aims to investigate the fatigue behavior of a newly developed nickel-based superalloy under biaxial-planar loading. Low Cycle Fatigue (LCF) tests are performed at different stress ratios so as to study the influence of the multiaxial stress state on the fatigue life of the material. Full-field displacement and strain measurements as well as crack initiation detection are obtained using Digital Image Correlation (DIC) techniques. The aim of this presentation is first to provide an in-depth description of both the experimental set-up and protocol: the multiaxial testing machine, the specific design of the cruciform specimen and performances of the DIC code are introduced. Second, results for sixteen specimens related to different load ratios are presented. Crack detection, strain amplitude and number of cycles to crack initiation vs. triaxial stress ratio for each loading case are given. Third, from fractographic investigations by scanning electron microscopy it is found that the mechanism of fatigue crack initiation does not depend on the triaxial stress ratio and that most fatigue cracks initiate from subsurface carbides.
Keywords: Cruciform specimen, multiaxial fatigue, Nickelbased superalloy.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2193286 In-situ Chemical Oxidation of Residual TCE by Permanganate in Epikarst
Authors: Nihat Hakan Akyol, Irfan Yolcubal
Abstract:
In-situ chemical oxidation (ISCO) has been widely used for source zone remediation of Dense Nonaqueous Phase Liquids (DNAPLs) in subsurface environments. DNAPL source zones for karst aquifers are generally located in epikarst where the DNAPL mass is trapped either in karst soil or at the regolith contact with carbonate bedrock. This study aims to investigate the performance of oxidation of residual trichloroethylene found in such environments by potassium permanganate. Batch and flow cell experiments were conducted to determine the kinetics and the mass removal rate of TCE. pH change, Cl production, TCE and MnO4 destruction were monitored routinely during experiments. Nonreactive tracer tests were also conducted prior and after the oxidation process to determine the influence of oxidation on flow conditions. The results show that oxidant consumption rate of the calcareous epikarst soil was significant and the oxidant demand was determined to be 20 g KMnO4/kg soil. Oxidation rate of residual TCE (1.26x10-3 s-1) was faster than the oxidant consumption rate of the soil (2.54 - 2.92x10-4 s-1) at only high oxidant concentrations (> 40 mM KMnO4). Half life of TCE oxidation ranged from 7.9 to 10.7 min. Although highly significant fraction of residual TCE mass in the system was destroyed by permanganate oxidation, TCE concentration in the effluent remained above its MCL. Flow interruption tests indicate that efficiency of ISCO was limited by the rate of TCE dissolution and the rate-limited desorption of TCE. The residence time and the initial concentration of the oxidant in the source zone also controlled the efficiency of ISCO in epikarst.Keywords: Epikarst, in-situ chemical oxidation, permanganate.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2027285 Job Satisfaction of Midwives Working in Labor Ward of the Lady Dufferin Hospital: A Cross-Sectional Study
Authors: B. Muhammadani
Abstract:
Health workforce is a fundamental component of health system and plays a significant role in delivering effective health care services. However, there is a crucial shortage of skilled personnel which make them prone to work in stressful conditions. In spite of excessively high workload and burnout among the staff, little attention is given to their job satisfaction level which has serious implications on the productivity and effective performance of staff to achieve organizational goals. Therefore, this study aims to explore the job satisfaction of midwives working in the labor ward of the Lady Dufferin Hospital, Karachi. A cross-sectional survey was conducted. The short version of Minnesota Job Satisfaction Questionnaire was administered on a convenient sample group of 22 midwives to gather information on their job satisfaction. The results demonstrated that midwives were overall satisfied with their job. The level of job satisfaction was however found different in various positions within midwifery cadre. The head of midwives was highly satisfied as compared to midwifery staff who works under the supervision of head. The level of satisfaction of team leaders fall between the head and staff of midwifery. Similar trends were observed for both intrinsic and extrinsic job satisfaction. Such evidences on these issues are essential and useful as it helps explore the attitudes of individuals towards work which has direct implications on access to quality care services. Strategic interventions are required at organizational level to provide motivators and satisfiers to health workers for their work related satisfaction and enhanced motivation.Keywords: Health workforce, job satisfaction, motivation, workload, burnout, midwives, health system.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2304284 Scatterer Density in Edge and Coherence Enhancing Nonlinear Anisotropic Diffusion for Medical Ultrasound Speckle Reduction
Authors: Ahmed Badawi, J. Michael Johnson, Mohamed Mahfouz
Abstract:
This paper proposes new enhancement models to the methods of nonlinear anisotropic diffusion to greatly reduce speckle and preserve image features in medical ultrasound images. By incorporating local physical characteristics of the image, in this case scatterer density, in addition to the gradient, into existing tensorbased image diffusion methods, we were able to greatly improve the performance of the existing filtering methods, namely edge enhancing (EE) and coherence enhancing (CE) diffusion. The new enhancement methods were tested using various ultrasound images, including phantom and some clinical images, to determine the amount of speckle reduction, edge, and coherence enhancements. Scatterer density weighted nonlinear anisotropic diffusion (SDWNAD) for ultrasound images consistently outperformed its traditional tensor-based counterparts that use gradient only to weight the diffusivity function. SDWNAD is shown to greatly reduce speckle noise while preserving image features as edges, orientation coherence, and scatterer density. SDWNAD superior performances over nonlinear coherent diffusion (NCD), speckle reducing anisotropic diffusion (SRAD), adaptive weighted median filter (AWMF), wavelet shrinkage (WS), and wavelet shrinkage with contrast enhancement (WSCE), make these methods ideal preprocessing steps for automatic segmentation in ultrasound imaging.Keywords: Nonlinear anisotropic diffusion, ultrasound imaging, speckle reduction, scatterer density estimation, edge based enhancement, coherence enhancement.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1906283 Software Product Quality Evaluation Model with Multiple Criteria Decision Making Analysis
Authors: C. Ardil
Abstract:
This paper presents a software product quality evaluation model based on the ISO/IEC 25010 quality model. The evaluation characteristics and sub characteristics were identified from the ISO/IEC 25010 quality model. The multidimensional structure of the quality model is based on characteristics such as functional suitability, performance efficiency, compatibility, usability, reliability, security, maintainability, and portability, and associated sub characteristics. Random numbers are generated to establish the decision maker’s importance weights for each sub characteristics. Also, random numbers are generated to establish the decision matrix of the decision maker’s final scores for each software product against each sub characteristics. Thus, objective criteria importance weights and index scores for datasets were obtained from the random numbers. In the proposed model, five different software product quality evaluation datasets under three different weight vectors were applied to multiple criteria decision analysis method, preference analysis for reference ideal solution (PARIS) for comparison, and sensitivity analysis procedure. This study contributes to provide a better understanding of the application of MCDMA methods and ISO/IEC 25010 quality model guidelines in software product quality evaluation process.
Keywords: ISO/IEC 25010 quality model, multiple criteria decisions making, multiple criteria decision making analysis, MCDMA, PARIS, Software Product Quality Evaluation Model, Software Product Quality Evaluation, Software Evaluation, Software Selection, Software
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 448282 Discovery of Quantified Hierarchical Production Rules from Large Set of Discovered Rules
Authors: Tamanna Siddiqui, M. Afshar Alam
Abstract:
Automated discovery of Rule is, due to its applicability, one of the most fundamental and important method in KDD. It has been an active research area in the recent past. Hierarchical representation allows us to easily manage the complexity of knowledge, to view the knowledge at different levels of details, and to focus our attention on the interesting aspects only. One of such efficient and easy to understand systems is Hierarchical Production rule (HPRs) system. A HPR, a standard production rule augmented with generality and specificity information, is of the following form: Decision If < condition> Generality
Keywords: Knowledge discovery in database, quantification, dempster shafer theory, genetic programming, hierarchy, subsumption matrix.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1527281 An Improved Total Variation Regularization Method for Denoising Magnetocardiography
Authors: Yanping Liao, Congcong He, Ruigang Zhao
Abstract:
The application of magnetocardiography signals to detect cardiac electrical function is a new technology developed in recent years. The magnetocardiography signal is detected with Superconducting Quantum Interference Devices (SQUID) and has considerable advantages over electrocardiography (ECG). It is difficult to extract Magnetocardiography (MCG) signal which is buried in the noise, which is a critical issue to be resolved in cardiac monitoring system and MCG applications. In order to remove the severe background noise, the Total Variation (TV) regularization method is proposed to denoise MCG signal. The approach transforms the denoising problem into a minimization optimization problem and the Majorization-minimization algorithm is applied to iteratively solve the minimization problem. However, traditional TV regularization method tends to cause step effect and lacks constraint adaptability. In this paper, an improved TV regularization method for denoising MCG signal is proposed to improve the denoising precision. The improvement of this method is mainly divided into three parts. First, high-order TV is applied to reduce the step effect, and the corresponding second derivative matrix is used to substitute the first order. Then, the positions of the non-zero elements in the second order derivative matrix are determined based on the peak positions that are detected by the detection window. Finally, adaptive constraint parameters are defined to eliminate noises and preserve signal peak characteristics. Theoretical analysis and experimental results show that this algorithm can effectively improve the output signal-to-noise ratio and has superior performance.Keywords: Constraint parameters, derivative matrix, magnetocardiography, regular term, total variation.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 697280 Efficient Design Optimization of Multi-State Flow Network for Multiple Commodities
Authors: Yu-Cheng Chou, Po Ting Lin
Abstract:
The network of delivering commodities has been an important design problem in our daily lives and many transportation applications. The delivery performance is evaluated based on the system reliability of delivering commodities from a source node to a sink node in the network. The system reliability is thus maximized to find the optimal routing. However, the design problem is not simple because (1) each path segment has randomly distributed attributes; (2) there are multiple commodities that consume various path capacities; (3) the optimal routing must successfully complete the delivery process within the allowable time constraints. In this paper, we want to focus on the design optimization of the Multi-State Flow Network (MSFN) for multiple commodities. We propose an efficient approach to evaluate the system reliability in the MSFN with respect to randomly distributed path attributes and find the optimal routing subject to the allowable time constraints. The delivery rates, also known as delivery currents, of the path segments are evaluated and the minimal-current arcs are eliminated to reduce the complexity of the MSFN. Accordingly, the correct optimal routing is found and the worst-case reliability is evaluated. It has been shown that the reliability of the optimal routing is at least higher than worst-case measure. Two benchmark examples are utilized to demonstrate the proposed method. The comparisons between the original and the reduced networks show that the proposed method is very efficient.
Keywords: Multiple Commodities, Multi-State Flow Network (MSFN), Time Constraints, Worst-Case Reliability (WCR)
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1449279 A Vehicle Monitoring System Based on the LoRa Technique
Authors: Chao-Linag Hsieh, Zheng-Wei Ye, Chen-Kang Huang, Yeun-Chung Lee, Chih-Hong Sun, Tzai-Hung Wen, Jehn-Yih Juang, Joe-Air Jiang
Abstract:
Air pollution and climate warming become more and more intensified in many areas, especially in urban areas. Environmental parameters are critical information to air pollution and weather monitoring. Thus, it is necessary to develop a suitable air pollution and weather monitoring system for urban areas. In this study, a vehicle monitoring system (VMS) based on the IoT technique is developed. Cars are selected as the research tool because it can reach a greater number of streets to collect data. The VMS can monitor different environmental parameters, including ambient temperature and humidity, and air quality parameters, including PM2.5, NO2, CO, and O3. The VMS can provide other information, including GPS signals and the vibration information through driving a car on the street. Different sensor modules are used to measure the parameters and collect the measured data and transmit them to a cloud server through the LoRa protocol. A user interface is used to show the sensing data storing at the cloud server. To examine the performance of the system, a researcher drove a Nissan x-trail 1998 to the area close to the Da’an District office in Taipei to collect monitoring data. The collected data are instantly shown on the user interface. The four kinds of information are provided by the interface: GPS positions, weather parameters, vehicle information, and air quality information. With the VMS, users can obtain the information regarding air quality and weather conditions when they drive their car to an urban area. Also, government agencies can make decisions on traffic planning based on the information provided by the proposed VMS.
Keywords: Vehicle, monitoring system, LoRa, smart city.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3101278 Efficiency of Robust Heuristic Gradient Based Enumerative and Tunneling Algorithms for Constrained Integer Programming Problems
Authors: Vijaya K. Srivastava, Davide Spinello
Abstract:
This paper presents performance of two robust gradient-based heuristic optimization procedures based on 3n enumeration and tunneling approach to seek global optimum of constrained integer problems. Both these procedures consist of two distinct phases for locating the global optimum of integer problems with a linear or non-linear objective function subject to linear or non-linear constraints. In both procedures, in the first phase, a local minimum of the function is found using the gradient approach coupled with hemstitching moves when a constraint is violated in order to return the search to the feasible region. In the second phase, in one optimization procedure, the second sub-procedure examines 3n integer combinations on the boundary and within hypercube volume encompassing the result neighboring the result from the first phase and in the second optimization procedure a tunneling function is constructed at the local minimum of the first phase so as to find another point on the other side of the barrier where the function value is approximately the same. In the next cycle, the search for the global optimum commences in both optimization procedures again using this new-found point as the starting vector. The search continues and repeated for various step sizes along the function gradient as well as that along the vector normal to the violated constraints until no improvement in optimum value is found. The results from both these proposed optimization methods are presented and compared with one provided by popular MS Excel solver that is provided within MS Office suite and other published results.
Keywords: Constrained integer problems, enumerative search algorithm, Heuristic algorithm, tunneling algorithm.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 800277 Incorporating Lexical-Semantic Knowledge into Convolutional Neural Network Framework for Pediatric Disease Diagnosis
Authors: Xiaocong Liu, Huazhen Wang, Ting He, Xiaozheng Li, Weihan Zhang, Jian Chen
Abstract:
The utilization of electronic medical record (EMR) data to establish the disease diagnosis model has become an important research content of biomedical informatics. Deep learning can automatically extract features from the massive data, which brings about breakthroughs in the study of EMR data. The challenge is that deep learning lacks semantic knowledge, which leads to impracticability in medical science. This research proposes a method of incorporating lexical-semantic knowledge from abundant entities into a convolutional neural network (CNN) framework for pediatric disease diagnosis. Firstly, medical terms are vectorized into Lexical Semantic Vectors (LSV), which are concatenated with the embedded word vectors of word2vec to enrich the feature representation. Secondly, the semantic distribution of medical terms serves as Semantic Decision Guide (SDG) for the optimization of deep learning models. The study evaluates the performance of LSV-SDG-CNN model on four kinds of Chinese EMR datasets. Additionally, CNN, LSV-CNN, and SDG-CNN are designed as baseline models for comparison. The experimental results show that LSV-SDG-CNN model outperforms baseline models on four kinds of Chinese EMR datasets. The best configuration of the model yielded an F1 score of 86.20%. The results clearly demonstrate that CNN has been effectively guided and optimized by lexical-semantic knowledge, and LSV-SDG-CNN model improves the disease classification accuracy with a clear margin.
Keywords: lexical semantics, feature representation, semantic decision, convolutional neural network, electronic medical record
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 594276 Overview Studies of High Strength Self-Consolidating Concrete
Authors: Raya Harkouss, Bilal Hamad
Abstract:
Self-Consolidating Concrete (SCC) is considered as a relatively new technology created as an effective solution to problems associated with low quality consolidation. A SCC mix is defined as successful if it flows freely and cohesively without the intervention of mechanical compaction. The construction industry is showing high tendency to use SCC in many contemporary projects to benefit from the various advantages offered by this technology.
At this point, a main question is raised regarding the effect of enhanced fluidity of SCC on the structural behavior of high strength self-consolidating reinforced concrete.
A three phase research program was conducted at the American University of Beirut (AUB) to address this concern. The first two phases consisted of comparative studies conducted on concrete and mortar mixes prepared with second generation Sulphonated Naphtalene-based superplasticizer (SNF) or third generation Polycarboxylate Ethers-based superplasticizer (PCE). The third phase of the research program investigates and compares the structural performance of high strength reinforced concrete beam specimens prepared with two different generations of superplasticizers that formed the unique variable between the concrete mixes. The beams were designed to test and exhibit flexure, shear, or bond splitting failure.
The outcomes of the experimental work revealed comparable resistance of beam specimens cast using self-compacting concrete and conventional vibrated concrete. The dissimilarities in the experimental values between the SCC and the control VC beams were minimal, leading to a conclusion, that the high consistency of SCC has little effect on the flexural, shear and bond strengths of concrete members.
Keywords: Self-consolidating concrete (SCC), high-strength concrete, concrete admixtures, mechanical properties of hardened SCC, structural behavior of reinforced concrete beams.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2970275 Cirrhosis Mortality Prediction as Classification Using Frequent Subgraph Mining
Authors: Abdolghani Ebrahimi, Diego Klabjan, Chenxi Ge, Daniela Ladner, Parker Stride
Abstract:
In this work, we use machine learning and data analysis techniques to predict the one-year mortality of cirrhotic patients. Data from 2,322 patients with liver cirrhosis are collected at a single medical center. Different machine learning models are applied to predict one-year mortality. A comprehensive feature space including demographic information, comorbidity, clinical procedure and laboratory tests is being analyzed. A temporal pattern mining technic called Frequent Subgraph Mining (FSM) is being used. Model for End-stage liver disease (MELD) prediction of mortality is used as a comparator. All of our models statistically significantly outperform the MELD-score model and show an average 10% improvement of the area under the curve (AUC). The FSM technic itself does not improve the model significantly, but FSM, together with a machine learning technique called an ensemble, further improves the model performance. With the abundance of data available in healthcare through electronic health records (EHR), existing predictive models can be refined to identify and treat patients at risk for higher mortality. However, due to the sparsity of the temporal information needed by FSM, the FSM model does not yield significant improvements. Our work applies modern machine learning algorithms and data analysis methods on predicting one-year mortality of cirrhotic patients and builds a model that predicts one-year mortality significantly more accurate than the MELD score. We have also tested the potential of FSM and provided a new perspective of the importance of clinical features.
Keywords: machine learning, liver cirrhosis, subgraph mining, supervised learning
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 449274 A Novel Neighborhood Defined Feature Selection on Phase Congruency Images for Recognition of Faces with Extreme Variations
Authors: Satyanadh Gundimada, Vijayan K Asari
Abstract:
A novel feature selection strategy to improve the recognition accuracy on the faces that are affected due to nonuniform illumination, partial occlusions and varying expressions is proposed in this paper. This technique is applicable especially in scenarios where the possibility of obtaining a reliable intra-class probability distribution is minimal due to fewer numbers of training samples. Phase congruency features in an image are defined as the points where the Fourier components of that image are maximally inphase. These features are invariant to brightness and contrast of the image under consideration. This property allows to achieve the goal of lighting invariant face recognition. Phase congruency maps of the training samples are generated and a novel modular feature selection strategy is implemented. Smaller sub regions from a predefined neighborhood within the phase congruency images of the training samples are merged to obtain a large set of features. These features are arranged in the order of increasing distance between the sub regions involved in merging. The assumption behind the proposed implementation of the region merging and arrangement strategy is that, local dependencies among the pixels are more important than global dependencies. The obtained feature sets are then arranged in the decreasing order of discriminating capability using a criterion function, which is the ratio of the between class variance to the within class variance of the sample set, in the PCA domain. The results indicate high improvement in the classification performance compared to baseline algorithms.
Keywords: Discriminant analysis, intra-class probability distribution, principal component analysis, phase congruency.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1850273 Signing the First Packet in Amortization Scheme for Multicast Stream Authentication
Authors: Mohammed Shatnawi, Qusai Abuein, Susumu Shibusawa
Abstract:
Signature amortization schemes have been introduced for authenticating multicast streams, in which, a single signature is amortized over several packets. The hash value of each packet is computed, some hash values are appended to other packets, forming what is known as hash chain. These schemes divide the stream into blocks, each block is a number of packets, the signature packet in these schemes is either the first or the last packet of the block. Amortization schemes are efficient solutions in terms of computation and communication overhead, specially in real-time environment. The main effictive factor of amortization schemes is it-s hash chain construction. Some studies show that signing the first packet of each block reduces the receiver-s delay and prevents DoS attacks, other studies show that signing the last packet reduces the sender-s delay. To our knowledge, there is no studies that show which is better, to sign the first or the last packet in terms of authentication probability and resistance to packet loss. In th is paper we will introduce another scheme for authenticating multicast streams that is robust against packet loss, reduces the overhead, and prevents the DoS attacks experienced by the receiver in the same time. Our scheme-The Multiple Connected Chain signing the First packet (MCF) is to append the hash values of specific packets to other packets,then append some hashes to the signature packet which is sent as the first packet in the block. This scheme is aspecially efficient in terms of receiver-s delay. We discuss and evaluate the performance of our proposed scheme against those that sign the last packet of the block.Keywords: multicast stream authentication, hash chain construction, signature amortization, authentication probability.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1518272 Opponent Color and Curvelet Transform Based Image Retrieval System Using Genetic Algorithm
Authors: Yesubai Rubavathi Charles, Ravi Ramraj
Abstract:
In order to retrieve images efficiently from a large database, a unique method integrating color and texture features using genetic programming has been proposed. Opponent color histogram which gives shadow, shade, and light intensity invariant property is employed in the proposed framework for extracting color features. For texture feature extraction, fast discrete curvelet transform which captures more orientation information at different scales is incorporated to represent curved like edges. The recent scenario in the issues of image retrieval is to reduce the semantic gap between user’s preference and low level features. To address this concern, genetic algorithm combined with relevance feedback is embedded to reduce semantic gap and retrieve user’s preference images. Extensive and comparative experiments have been conducted to evaluate proposed framework for content based image retrieval on two databases, i.e., COIL-100 and Corel-1000. Experimental results clearly show that the proposed system surpassed other existing systems in terms of precision and recall. The proposed work achieves highest performance with average precision of 88.2% on COIL-100 and 76.3% on Corel, the average recall of 69.9% on COIL and 76.3% on Corel. Thus, the experimental results confirm that the proposed content based image retrieval system architecture attains better solution for image retrieval.Keywords: Content based image retrieval, Curvelet transform, Genetic algorithm, Opponent color histogram, Relevance feedback.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1822271 The Effect of Cow Reproductive Traits on Lifetime Productivity and Longevity
Authors: Lāsma Cielava, Daina Jonkus, Līga Paura
Abstract:
The age of first calving (AFC) is one of the most important factors that have a significant impact on cow productivity in different lactations and its whole life. A belated AFC leads to reduced reproductive performance and it is one of the main reasons for reduced longevity. Cows that calved in time period from 2001-2007 and in this time finished at least four lactations were included in the database. Data were obtained from 68841 crossbred Holstein Black and White (HM), crossbred Latvian Brown (LB), and Latvian Brown genetic resources (LBGR) cows. Cows were distributed in four groups depending on age at first calving. The longest lifespan was conducted for LBGR cows, but they were also characterized with lowest lifetime milk yield and life day milk yield. HM breed cows had the shortest lifespan, but in the lifespan of 2862.2 days was obtained in average 37916.4 kg milk accordingly 13.2 kg milk in one life day. HM breed cows were also characterized with longer calving intervals (CI) in first four lactations, but LBGR cows had the shortest CI in the study group. Age at first calving significantly affected the length of CI in different lactations (p<0.05). HM cows that first time calved >30 months old in the fourth lactation had the longest CI in all study groups (421.4 days). The LBGR cows were characterized with the shortest CI, but there was slight increase in second and third lactation. Age at first calving had a significant impact on cows’ age in each calving time. In the analysis, cow group was conducted that cows with age at first calving <24 months or in average 580.5 days at the time of fifth calving were 2156.7 days (5.9 years) old, but cows with age at first calving >30 months (932.6 days) at the time of fifth calving were 2560.9 days (7.3 years) old.
Keywords: Age at first calving, calving interval, longevity, milk yield.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1634270 Very Large Scale Integration Architecture of Finite Impulse Response Filter Implementation Using Retiming Technique
Authors: S. Jalaja, A. M. Vijaya Prakash
Abstract:
Recursive combination of an algorithm based on Karatsuba multiplication is exploited to design a generalized transpose and parallel Finite Impulse Response (FIR) Filter. Mid-range Karatsuba multiplication and Carry Save adder based on Karatsuba multiplication reduce time complexity for higher order multiplication implemented up to n-bit. As a result, we design modified N-tap Transpose and Parallel Symmetric FIR Filter Structure using Karatsuba algorithm. The mathematical formulation of the FFA Filter is derived. The proposed architecture involves significantly less area delay product (APD) then the existing block implementation. By adopting retiming technique, hardware cost is reduced further. The filter architecture is designed by using 90 nm technology library and is implemented by using cadence EDA Tool. The synthesized result shows better performance for different word length and block size. The design achieves switching activity reduction and low power consumption by applying with and without retiming for different combination of the circuit. The proposed structure achieves more than a half of the power reduction by adopting with and without retiming techniques compared to the earlier design structure. As a proof of the concept for block size 16 and filter length 64 for CKA method, it achieves a 51% as well as 70% less power by applying retiming technique, and for CSA method it achieves a 57% as well as 77% less power by applying retiming technique compared to the previously proposed design.Keywords: Carry save adder Karatsuba multiplication, mid-range Karatsuba multiplication, modified FFA, transposed filter, retiming.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 910269 An Autonomous Collaborative Forecasting System Implementation – The First Step towards Successful CPFR System
Authors: Chi-Fang Huang, Yun-Shiow Chen, Yun-Kung Chung
Abstract:
In the past decade, artificial neural networks (ANNs) have been regarded as an instrument for problem-solving and decision-making; indeed, they have already done with a substantial efficiency and effectiveness improvement in industries and businesses. In this paper, the Back-Propagation neural Networks (BPNs) will be modulated to demonstrate the performance of the collaborative forecasting (CF) function of a Collaborative Planning, Forecasting and Replenishment (CPFR®) system. CPFR functions the balance between the sufficient product supply and the necessary customer demand in a Supply and Demand Chain (SDC). Several classical standard BPN will be grouped, collaborated and exploited for the easy implementation of the proposed modular ANN framework based on the topology of a SDC. Each individual BPN is applied as a modular tool to perform the task of forecasting SKUs (Stock-Keeping Units) levels that are managed and supervised at a POS (point of sale), a wholesaler, and a manufacturer in an SDC. The proposed modular BPN-based CF system will be exemplified and experimentally verified using lots of datasets of the simulated SDC. The experimental results showed that a complex CF problem can be divided into a group of simpler sub-problems based on the single independent trading partners distributed over SDC, and its SKU forecasting accuracy was satisfied when the system forecasted values compared to the original simulated SDC data. The primary task of implementing an autonomous CF involves the study of supervised ANN learning methodology which aims at making “knowledgeable" decision for the best SKU sales plan and stocks management.Keywords: CPFR, artificial neural networks, global logistics, supply and demand chain.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1993268 Analysis of Supply Side Factors Affecting Bank Financing of Non-Oil Exports in Nigeria
Authors: Sama’ila Idi Ningi, Abubakar Yusuf Dutse
Abstract:
The banking sector poses a lot of problems in Nigeria in general and the non-oil export sector in particular. The banks' lack effectiveness in handling small, medium or long-term credit risk (lack of training of loan officers, lack of information on borrowers and absence of a reliable credit registry) results in non-oil exporters being burdened with high requirements, such as up to three years of financial statements, enough collateral to cover both the loan principal and interest (including a cash deposit that may be up to 30% of the loans' net present value), and to provide every detail of the international trade transaction in question. The stated problems triggered this research. Consequently, information on bank financing of non-oil exports was collected from 100 respondents from the 20 Deposit Money Banks (DMBs) in Nigeria. The data was analysed by the use of descriptive statistics correlation and regression. It is found that, Nigerian banks are participants in the financing of non-oil exports. Despite their participation, the rate of interest for credit extended to non-oil export is usually high, ranging between 15-20%. Small and medium sized non-oil export businesses lack the credit history for banks to judge them as reputable. Banks also consider the non-oil export sector very risky for investment. The banks actually do grant less credit than the exporters may require and therefore are not properly funded by banks. Banks grant very low volume of foreign currency loan in addition to, unfavorable exchange rate at which Naira is exchanged to the Dollar and other currencies in the country. This makes importation of inputs costly and negatively impacted on the non-oil export performance in Nigeria.
Keywords: Supply Side Factors, Bank Financing, Non-Oil Exports.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2711267 An Evaluation of the Feasibility of Several Industrial Wastes and Natural Materials as Precursors for the Production of Alkali Activated Materials
Authors: O. Alelweet, S. Pavia
Abstract:
In order to face current compelling environmental problems affecting the planet, the construction industry needs to adapt. It is widely acknowledged that there is a need for durable, high-performance, low-greenhouse gas emission binders that can be used as an alternative to Portland cement (PC) to lower the environmental impact of construction. Alkali activated materials (AAMs) are considered a more sustainable alternative to PC materials. The binders of AAMs result from the reaction of an alkali metal source and a silicate powder or precursor which can be a calcium silicate or an aluminosilicate-rich material. This paper evaluates the particle size, specific surface area, chemical and mineral composition and amorphousness of silicate materials (most industrial waste locally produced in Ireland and Saudi Arabia) to develop alkali-activated binders that can replace PC resources in specific applications. These include recycled ceramic brick, bauxite, illitic clay, fly ash and metallurgical slag. According to the results, the wastes are reactive and comply with building standards requirements. The study also evidenced that the reactivity of the Saudi bauxite (with significant kaolinite) can be enhanced on thermal activation; and high calcium in the slag will promote reaction; which should be possible with low alkalinity activators. The wastes evidenced variable water demands that will be taken into account for mixing with the activators. Finally, further research is proposed to further determine the reactive fraction of the clay-based precursors.
Keywords: Reactivity, water demand, alkali-activated materials, brick, bauxite, illitic clay, fly ash, slag.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 778