Search results for: full-potential KKR-green’s function method
20345 An Engineer-Oriented Life Cycle Assessment Tool for Building Carbon Footprint: The Building Carbon Footprint Evaluation System in Taiwan
Authors: Hsien-Te Lin
Abstract:
The purpose of this paper is to introduce the BCFES (building carbon footprint evaluation system), which is a LCA (life cycle assessment) tool developed by the Low Carbon Building Alliance (LCBA) in Taiwan. A qualified BCFES for the building industry should fulfill the function of evaluating carbon footprint throughout all stages in the life cycle of building projects, including the production, transportation and manufacturing of materials, construction, daily energy usage, renovation and demolition. However, many existing BCFESs are too complicated and not very designer-friendly, creating obstacles in the implementation of carbon reduction policies. One of the greatest obstacle is the misapplication of the carbon footprint inventory standards of PAS2050 or ISO14067, which are designed for mass-produced goods rather than building projects. When these product-oriented rules are applied to building projects, one must compute a tremendous amount of data for raw materials and the transportation of construction equipment throughout the construction period based on purchasing lists and construction logs. This verification method is very cumbersome by nature and unhelpful to the promotion of low carbon design. With a view to provide an engineer-oriented BCFE with pre-diagnosis functions, a component input/output (I/O) database system and a scenario simulation method for building energy are proposed herein. Most existing BCFESs base their calculations on a product-oriented carbon database for raw materials like cement, steel, glass, and wood. However, data on raw materials is meaningless for the purpose of encouraging carbon reduction design without a feedback mechanism, because an engineering project is not designed based on raw materials but rather on building components, such as flooring, walls, roofs, ceilings, roads or cabinets. The LCBA Database has been composited from existing carbon footprint databases for raw materials and architectural graphic standards. Project designers can now use the LCBA Database to conduct low carbon design in a much more simple and efficient way. Daily energy usage throughout a building's life cycle, including air conditioning, lighting, and electric equipment, is very difficult for the building designer to predict. A good BCFES should provide a simplified and designer-friendly method to overcome this obstacle in predicting energy consumption. In this paper, the author has developed a simplified tool, the dynamic Energy Use Intensity (EUI) method, to accurately predict energy usage with simple multiplications and additions using EUI data and the designed efficiency levels for the building envelope, AC, lighting and electrical equipment. Remarkably simple to use, it can help designers pre-diagnose hotspots in building carbon footprint and further enhance low carbon designs. The BCFES-LCBA offers the advantages of an engineer-friendly component I/O database, simplified energy prediction methods, pre-diagnosis of carbon hotspots and sensitivity to good low carbon designs, making it an increasingly popular carbon management tool in Taiwan. To date, about thirty projects have been awarded BCFES-LCBA certification and the assessment has become mandatory in some cities.Keywords: building carbon footprint, life cycle assessment, energy use intensity, building energy
Procedia PDF Downloads 13920344 Suitable Die Shaping for a Rectangular Shape Bottle by Application of FEM and AI Technique
Authors: N. Ploysook, R. Rugsaj, C. Suvanjumrat
Abstract:
The characteristic requirement for producing rectangular shape bottles was a uniform thickness of the plastic bottle wall. Die shaping was a good technique which controlled the wall thickness of bottles. An advance technology which was the finite element method (FEM) for blowing parison to be a rectangular shape bottle was conducted to reduce waste plastic from a trial and error method of a die shaping and parison control method. The artificial intelligent (AI) comprised of artificial neural network and genetic algorithm was selected to optimize the die gap shape from the FEM results. The application of AI technique could optimize the suitable die gap shape for the parison blow molding which did not depend on the parison control method to produce rectangular bottles with the uniform wall. Particularly, this application can be used with cheap blow molding machines without a parison controller therefore it will reduce cost of production in the bottle blow molding process.Keywords: AI, bottle, die shaping, FEM
Procedia PDF Downloads 23920343 Performance Analysis of New Types of Reference Targets Based on Spaceborne and Airborne SAR Data
Authors: Y. S. Zhou, C. R. Li, L. L. Tang, C. X. Gao, D. J. Wang, Y. Y. Guo
Abstract:
Triangular trihedral corner reflector (CR) has been widely used as point target for synthetic aperture radar (SAR) calibration and image quality assessment. The additional “tip” of the triangular plate does not contribute to the reflector’s theoretical RCS and if it interacts with a perfectly reflecting ground plane, it will yield an increase of RCS at the radar bore-sight and decrease the accuracy of SAR calibration and image quality assessment. Regarding this problem, two types of CRs were manufactured. One was the hexagonal trihedral CR. It is a self-illuminating CR with relatively small plate edge length, while large edge length usually introduces unexpected edge diffraction error. The other was the triangular trihedral CR with extended bottom plate which considers the effect of ‘tip’ into the total RCS. In order to assess the performance of the two types of new CRs, flight campaign over the National Calibration and Validation Site for High Resolution Remote Sensors was carried out. Six hexagonal trihedral CRs and two bottom-extended trihedral CRs, as well as several traditional triangular trihedral CRs, were deployed. KOMPSAT-5 X-band SAR image was acquired for the performance analysis of the hexagonal trihedral CRs. C-band airborne SAR images were acquired for the performance analysis of the bottom-extended trihedral CRs. The analysis results showed that the impulse response function of both the hexagonal trihedral CRs and bottom-extended trihedral CRs were much closer to the ideal sinc-function than the traditional triangular trihedral CRs. The flight campaign results validated the advantages of new types of CRs and they might be useful in the future SAR calibration mission.Keywords: synthetic aperture radar, calibration, corner reflector, KOMPSAT-5
Procedia PDF Downloads 27320342 Robotic Assisted vs Traditional Laparoscopic Partial Nephrectomy Peri-Operative Outcomes: A Comparative Single Surgeon Study
Authors: Gerard Bray, Derek Mao, Arya Bahadori, Sachinka Ranasinghe
Abstract:
The EAU currently recommends partial nephrectomy as the preferred management for localised cT1 renal tumours, irrespective of surgical approach. With the advent of robotic assisted partial nephrectomy, there is growing evidence that warm ischaemia time may be reduced compared to the traditional laparoscopic approach. There is still no clear differences between the two approaches with regards to other peri-operative and oncological outcomes. Current limitations in the field denote the lack of single surgeon series to compare the two approaches as other studies often include multiple operators of different experience levels. To the best of our knowledge, this study is the first single surgeon series comparing peri-operative outcomes of robotic assisted and laparoscopic PN. The current study aims to reduce intra-operator bias while maintaining an adequate sample size to assess the differences in outcomes between the two approaches. We retrospectively compared patient demographics, peri-operative outcomes, and renal function derangements of all partial nephrectomies undertaken by a single surgeon with experience in both laparoscopic and robotic surgery. Warm ischaemia time, length of stay, and acute renal function deterioration were all significantly reduced with robotic partial nephrectomy, compared to laparoscopic nephrectomy. This study highlights the benefits of robotic partial nephrectomy. Further prospective studies with larger sample sizes would be valuable additions to the current literature.Keywords: partial nephrectomy, robotic assisted partial nephrectomy, warm ischaemia time, peri-operative outcomes
Procedia PDF Downloads 14120341 Time Series Forecasting (TSF) Using Various Deep Learning Models
Authors: Jimeng Shi, Mahek Jain, Giri Narasimhan
Abstract:
Time Series Forecasting (TSF) is used to predict the target variables at a future time point based on the learning from previous time points. To keep the problem tractable, learning methods use data from a fixed-length window in the past as an explicit input. In this paper, we study how the performance of predictive models changes as a function of different look-back window sizes and different amounts of time to predict the future. We also consider the performance of the recent attention-based Transformer models, which have had good success in the image processing and natural language processing domains. In all, we compare four different deep learning methods (RNN, LSTM, GRU, and Transformer) along with a baseline method. The dataset (hourly) we used is the Beijing Air Quality Dataset from the UCI website, which includes a multivariate time series of many factors measured on an hourly basis for a period of 5 years (2010-14). For each model, we also report on the relationship between the performance and the look-back window sizes and the number of predicted time points into the future. Our experiments suggest that Transformer models have the best performance with the lowest Mean Average Errors (MAE = 14.599, 23.273) and Root Mean Square Errors (RSME = 23.573, 38.131) for most of our single-step and multi-steps predictions. The best size for the look-back window to predict 1 hour into the future appears to be one day, while 2 or 4 days perform the best to predict 3 hours into the future.Keywords: air quality prediction, deep learning algorithms, time series forecasting, look-back window
Procedia PDF Downloads 15420340 A Strategy of Direct Power Control for PWM Rectifier Reducing Ripple in Instantaneous Power
Authors: T. Mohammed Chikouche, K. Hartani
Abstract:
In order to solve the instantaneous power ripple and achieve better performance of direct power control (DPC) for a three-phase PWM rectifier, a control method is proposed in this paper. This control method is applied to overcome the instantaneous power ripple, to eliminate line current harmonics and therefore reduce the total harmonic distortion and to improve the power factor. A switching table is based on the analysis on the change of instantaneous active and reactive power, to select the optimum switching state of the three-phase PWM rectifier. The simulation result shows feasibility of this control method.Keywords: power quality, direct power control, power ripple, switching table, unity power factor
Procedia PDF Downloads 32220339 Environmental Controls on the Distribution of Intertidal Foraminifers in Sabkha Al-Kharrar, Saudi Arabia: Implications for Sea-Level Changes
Authors: Talha A. Al-Dubai, Rashad A. Bantan, Ramadan H. Abu-Zied, Brian G. Jones, Aaid G. Al-Zubieri
Abstract:
Contemporary foraminiferal samples sediments were collected from the intertidal sabkha of Al-Kharrar Lagoon, Saudi Arabia, to study the vertical distribution of Foraminifera and, based on a modern training set, their potential to develop a predictor of former sea-level changes in the area. Based on hierarchical cluster analysis, the intertidal sabkha is divided into three vertical zones (A, B & C) represented by three foraminiferal assemblages, where agglutinated species occupied Zone A and calcareous species occupied the other two zones. In Zone A (high intertidal), Agglutinella compressa, Clavulina angularis and C. multicamerata are dominant species with a minor presence of Peneroplis planatus, Coscinospira hemprichii, Sorites orbiculus, Quinqueloculina lamarckiana, Q. seminula, Ammonia convexa and A. tepida. In contrast, in Zone B (middle intertidal) the most abundant species are P. planatus, C. hemprichii, S. orbiculus, Q. lamarckiana, Q. seminula and Q. laevigata, while Zone C (low intertidal) is characterised by C. hemprichii, Q. costata, S. orbiculus, P. planatus, A. convexa, A. tepida, Spiroloculina communis and S. costigera. A transfer function for sea-level reconstruction was developed using a modern dataset of 75 contemporary sediment samples and 99 species collected from several transects across the sabkha. The model provided an error of 0.12m, suggesting that intertidal foraminifers are able to predict the past sea-level changes with high precision in Al-Kharrar Lagoon, and thus the future prediction of those changes in the area.Keywords: Lagoonal foraminifers, intertidal sabkha, vertical zonation, transfer function, sea level
Procedia PDF Downloads 16920338 Investigation of the Influences of Heat Sinks on the Thermal Efficiency of a Two Cylinder Stirling Engine Using the Taguchi Method
Authors: Chin-Kuei Lin, Wen-KO Liang, Yi-Xiang Wang, Shie-Chen Yang
Abstract:
In this study, experimental methods and finite element numerical simulation are used to investigate the influences of three different types of heat sinks on the output power of a two-cylinder Stirling engine. The Taguchi method is employed to optimize the geometric parameters of the air-cooled heat sink, aiming to optimize the output thermal efficiency. The parameters of the Taguchi method include the materials of the heat sink, and the shape, height, width, and spacing of the heat sink. The L18 orthogonal array is used for experimental design. The experimental results show that the thermal efficiency of the Stirling engine is significantly influenced by the material and geometric parameters of the heat sinks. The heat sinks effectively decrease the temperature at the cold end of the cylinder, increase the temperature difference between the hot and cold ends of the cylinder, and enhance the thermal efficiency of the Stirling engine. The experimental results are in agreement with the numerical simulation analysis. The results of this study will be more helpful for the application of Stirling engines.Keywords: two cylinder Stirling engine, Taguchi method, heat sink, finite element method
Procedia PDF Downloads 320337 Modification Encryption Time and Permutation in Advanced Encryption Standard Algorithm
Authors: Dalal N. Hammod, Ekhlas K. Gbashi
Abstract:
Today, cryptography is used in many applications to achieve high security in data transmission and in real-time communications. AES has long gained global acceptance and is used for securing sensitive data in various industries but has suffered from slow processing and take a large time to transfer data. This paper suggests a method to enhance Advance Encryption Standard (AES) Algorithm based on time and permutation. The suggested method (MAES) is based on modifying the SubByte and ShiftRrows in the encryption part and modification the InvSubByte and InvShiftRows in the decryption part. After the implementation of the proposal and testing the results, the Modified AES achieved good results in accomplishing the communication with high performance criteria in terms of randomness, encryption time, storage space, and avalanche effects. The proposed method has good randomness to ciphertext because this method passed NIST statistical tests against attacks; also, (MAES) reduced the encryption time by (10 %) than the time of the original AES; therefore, the modified AES is faster than the original AES. Also, the proposed method showed good results in memory utilization where the value is (54.36) for the MAES, but the value for the original AES is (66.23). Also, the avalanche effects used for calculating diffusion property are (52.08%) for the modified AES and (51.82%) percentage for the original AES.Keywords: modified AES, randomness test, encryption time, avalanche effects
Procedia PDF Downloads 24820336 Influence of Magnetic Field on Microstructure and Properties of Copper-Silver Composites
Authors: Engang Wang
Abstract:
The Cu-alloy composites are a kind of high-strength and high-conductivity Cu-based alloys, which have excellent mechanical and electrical properties and is widely used in electronic, electrical, machinery industrial fields. However, the solidification microstructure of the composites, such as the primary or second dendrite arm spacing, have important rule to its tensile strength and conductivity, and that is affected by its fabricating method. In this paper, two kinds of directional solidification methods; the exothermic powder method (EP method) and liquid metal cooling method (LMC method), were used to fabricate the Cu-alloy composites with applied different magnetic fields to investigate their influence on the solidifying microstructure of Cu-alloy, and further the fabricated Cu-alloy composites was drawn to wires to investigate the influence of fabricating method and magnetic fields on the drawing microstructure of fiber-reinforced Cu-alloy composites and its properties. The experiment of Cu-Ag alloy under directional solidification and horizontal magnetic fields with different processing parameters show that: 1) For the Cu-Ag alloy with EP method, the dendrite is directionally developed in the cooling copper mould and the solidifying microstructure is effectively refined by applying horizontal magnetic fields. 2) For the Cu-Ag alloy with LMC method, the primary dendrite arm spacing is decreased and the content of Ag in the dendrite increases as increasing the drawing velocity of solidification. 3) The dendrite is refined and the content of Ag in the dendrite increases as increasing the magnetic flux intensity; meanwhile, the growth direction of dendrite is also affected by magnetic field. The research results of Cu-Ag alloy in situ composites by drawing deforming process show that the micro-hardness of alloy is higher by decreasing dendrite arm spacing. When the dendrite growth orientation is consistent with the axial of the samples. the conductivity of the composites increases with the second dendrite arm spacing increases. However, its conductivity reduces with the applied magnetic fields owing to disrupting the dendrite growth orientation.Keywords: Cu-Ag composite, magnetic field, microstructure, solidification
Procedia PDF Downloads 21420335 A Study of Achievement and Attitude on Learning Science in English by Using Co – Teaching Method
Authors: Sakchai Rachniyom
Abstract:
Owing to the ASEAN community will formally take place in the few months; therefore, Thais should realize about the importance of English language. Since, it is regarded as a working language in the community. To promote Science students’ English proficiency, teacher should be able to teach in English language appropriately and effectively. The purposes of the quasi – experimental research are (1) to measure the learning achievement, (2) to evaluate students’ satisfaction on the teaching and learning and (3) to study the consequences of co – teaching method in order comprehend the learning achievement and improvement. The participants were 40 general science students teacher. Two types of research instruments were included; (1) an achievement test, and (2) a questionnaire. This research was conducted for 1 semester. The statistics used in this research were arithmetic mean and standard deviation. The findings of the study revealed that students’ achievement score was significantly increased at statistical level .05 and the students satisfied the teaching and learning at the highest level . The students’ involvement and teachers’ support were promoted. It was also reported students’ learning was improved by co – teaching method.Keywords: co – teaching method, learning science in english, teacher, education
Procedia PDF Downloads 48020334 Long Term Examination of the Profitability Estimation Focused on Benefits
Authors: Stephan Printz, Kristina Lahl, René Vossen, Sabina Jeschke
Abstract:
Strategic investment decisions are characterized by high innovation potential and long-term effects on the competitiveness of enterprises. Due to the uncertainty and risks involved in this complex decision making process, the need arises for well-structured support activities. A method that considers cost and the long-term added value is the cost-benefit effectiveness estimation. One of those methods is the “profitability estimation focused on benefits – PEFB”-method developed at the Institute of Management Cybernetics at RWTH Aachen University. The method copes with the challenges associated with strategic investment decisions by integrating long-term non-monetary aspects whilst also mapping the chronological sequence of an investment within the organization’s target system. Thus, this method is characterized as a holistic approach for the evaluation of costs and benefits of an investment. This participation-oriented method was applied to business environments in many workshops. The results of the workshops are a library of more than 96 cost aspects, as well as 122 benefit aspects. These aspects are preprocessed and comparatively analyzed with regards to their alignment to a series of risk levels. For the first time, an accumulation and a distribution of cost and benefit aspects regarding their impact and probability of occurrence are given. The results give evidence that the PEFB-method combines precise measures of financial accounting with the incorporation of benefits. Finally, the results constitute the basics for using information technology and data science for decision support when applying within the PEFB-method.Keywords: cost-benefit analysis, multi-criteria decision, profitability estimation focused on benefits, risk and uncertainty analysis
Procedia PDF Downloads 44520333 Acoustics Barrier Design to Reduce Railway Noise by Using Maekawa's Method
Authors: Malinda Sabrina, Khoerul Anwar
Abstract:
Railway noise generated by pass-by train has been described as a form of environmental pollutants especially for the residential area near the railway. Many studies have shown, that environmental noise particularly transportation noise has negative effects on people which resulting in annoyance and specific health problems such as cardiovascular disease, cognitive impairment and sleep disturbance. Therefore, various attempts are made to reduce the noise. One method of reducing such noise to acceptable noise levels is to build acoustically barrier walls. The objective of this study was to review the method of reducing railway noise and obtain the preliminary design of the acoustics barrier on the edge of railway tracks close to the residential area. The design of this barrier is using the Maekawa's method. Measurements have been performed in residential areas around the railroads in the Karawang - Indonesia with the absence of an acoustical barrier. From the observation, it was found that the railway was passed by five trains within thirty minutes. With the limited distance between the railway tracks and the location of the residential area as well as the street of residents, then it was obtained that a reduction in sound pressure level is 25 dBA. Maximum sound pressure level obtained is 86.9 dBA then by setting the barrier as high as 4 m at a distance, 2.5 m from the railway, the noise level received by residents in the settlement around the railway line becomes 61.9 dBA.Keywords: acoustics barrier, Maekawa's method, noise attenuation, railway noise
Procedia PDF Downloads 20020332 Maximum Power Point Tracking Based on Estimated Power for PV Energy Conversion System
Authors: Zainab Almukhtar, Adel Merabet
Abstract:
In this paper, a method for maximum power point tracking of a photovoltaic energy conversion system is presented. This method is based on using the difference between the power from the solar panel and an estimated power value to control the DC-DC converter of the photovoltaic system. The difference is continuously compared with a preset error permitted value. If the power difference is more than the error, the estimated power is multiplied by a factor and the operation is repeated until the difference is less or equal to the threshold error. The difference in power will be used to trigger a DC-DC boost converter in order to raise the voltage to where the maximum power point is achieved. The proposed method was experimentally verified through a PV energy conversion system driven by the OPAL-RT real time controller. The method was tested on varying radiation conditions and load requirements, and the Photovoltaic Panel was operated at its maximum power in different conditions of irradiation.Keywords: control system, error, solar panel, MPPT tracking
Procedia PDF Downloads 28420331 Clouds Influence on Atmospheric Ozone from GOME-2 Satellite Measurements
Authors: S. M. Samkeyat Shohan
Abstract:
This study is mainly focused on the determination and analysis of the photolysis rate of atmospheric, specifically tropospheric, ozone as function of cloud properties through-out the year 2007. The observational basis for ozone concentrations and cloud properties are the measurement data set of the Global Ozone Monitoring Experiment-2 (GOME-2) sensor on board the polar orbiting Metop-A satellite. Two different spectral ranges are used; ozone total column are calculated from the wavelength window 325 – 335 nm, while cloud properties, such as cloud top height (CTH) and cloud optical thick-ness (COT) are derived from the absorption band of molecular oxygen centered at 761 nm. Cloud fraction (CF) is derived from measurements in the ultraviolet, visible and near-infrared range of GOME-2. First, ozone concentrations above clouds are derived from ozone total columns, subtracting the contribution of stratospheric ozone and filtering those satellite measurements which have thin and low clouds. Then, the values of ozone photolysis derived from observations are compared with theoretical modeled results, in the latitudinal belt 5˚N-5˚S and 20˚N - 20˚S, as function of CF and COT. In general, good agreement is found between the data and the model, proving both the quality of the space-borne ozone and cloud properties as well as the modeling theory of ozone photolysis rate. The found discrepancies can, however, amount to approximately 15%. Latitudinal seasonal changes of photolysis rate of ozone are found to be negatively correlated to changes in upper-tropospheric ozone concentrations only in the autumn and summer months within the northern and southern tropical belts, respectively. This fact points to the entangled roles of temperature and nitrogen oxides in the ozone production, which are superimposed on its sole photolysis induced by thick and high clouds in the tropics.Keywords: cloud properties, photolysis rate, stratospheric ozone, tropospheric ozone
Procedia PDF Downloads 21120330 Effects of Turkish Classical Music on Cognitive Function, Depression and Quality of Life in Elderly
Authors: Rukiye Pinar Boluktas
Abstract:
According to 2015 statistics, in Turkey, 46% of older people live alone in their homes, 55% have poor health perceptions, 18% face poverty, and 43% are unhappy. Prevalence of depression is between 14% and 20%. In 2013, rate of suicide was 6.5. However, the most of older people prefer to live in their community although they are lonely, they face poverty, and face limitations as a result of chronic diseases and disabilities. Community based care for older people is also encouraged by Ministry of Health as it is more cost-effective. Music therapy is a simple, effective, safe, and nonpharmacologic intervention that may be used to decrease depression and to improve cognition, and health related quality of life (HRQOL). In Turkish culture, music is typically described as ‘food for soul’. This study aimed to investigate the effect of Turkish classical music songs in 32 community dwelling older people. Participants were received interventions two or three times per week, 50-60 min per session, for 8 weeks at a day health center. Each intervention session started listening music for 15-20 min to get remember songs, then followed singing songs as a group. Participants were assessed at baseline (week 0), and two follow-up at month 1 and month 2. Compared to baseline, at two follow-up, we observed that cognition improved, depression decreased, and SF-36 scores, including 8 domains and two summary scores increased. We conclude that an intervention comprising listening and singing Turkish classical music improve cognition, depression and HRQOL in older people.Keywords: cognitive function, depression, elderly, quality of life, Turkish classical music
Procedia PDF Downloads 16520329 Determining the Mode II Intra Ply Energy Release Rate of Composites Made of Prepreg
Authors: Philip Rose, Markus Linke, David Busquets
Abstract:
The distinction between interlaminar and intralaminar fracture toughness has already been investigated by several authors. For loading mode I, the double cantilever beam specimens were often used for the interlaminar fracture toughness and the compact tension specimen for the intralaminar fracture toughness. In order to minimize the influence of the different specimen geometries, a method was developed which allows the determination of both the interlaminar and the intralaminar fracture toughness on an almost identical specimen geometry. However, as this method is not applicable to prepreg semi-finished products, a further modification was developed, which is also suitable for prepreg laminates. After the successful application for the investigation of mode I with this method, the application of the method for loading mode II is presented in this paper. In addition to manufacturing differences, due to an additional fiber ply in which the controlled crack growth takes place, the adapted test procedure is also explained. By comparing the test results of standardized end-notched flexure (ENF) specimens with those of the modified ENF specimen, the difference between the interlaminar and intralaminar fracture toughness of the material Hexply 8552/IM7 is shown.Keywords: ENF, fracture toughness, interlaminar, mode II
Procedia PDF Downloads 13720328 Treating Voxels as Words: Word-to-Vector Methods for fMRI Meta-Analyses
Authors: Matthew Baucum
Abstract:
With the increasing popularity of fMRI as an experimental method, psychology and neuroscience can greatly benefit from advanced techniques for summarizing and synthesizing large amounts of data from brain imaging studies. One promising avenue is automated meta-analyses, in which natural language processing methods are used to identify the brain regions consistently associated with certain semantic concepts (e.g. “social”, “reward’) across large corpora of studies. This study builds on this approach by demonstrating how, in fMRI meta-analyses, individual voxels can be treated as vectors in a semantic space and evaluated for their “proximity” to terms of interest. In this technique, a low-dimensional semantic space is built from brain imaging study texts, allowing words in each text to be represented as vectors (where words that frequently appear together are near each other in the semantic space). Consequently, each voxel in a brain mask can be represented as a normalized vector sum of all of the words in the studies that showed activation in that voxel. The entire brain mask can then be visualized in terms of each voxel’s proximity to a given term of interest (e.g., “vision”, “decision making”) or collection of terms (e.g., “theory of mind”, “social”, “agent”), as measured by the cosine similarity between the voxel’s vector and the term vector (or the average of multiple term vectors). Analysis can also proceed in the opposite direction, allowing word cloud visualizations of the nearest semantic neighbors for a given brain region. This approach allows for continuous, fine-grained metrics of voxel-term associations, and relies on state-of-the-art “open vocabulary” methods that go beyond mere word-counts. An analysis of over 11,000 neuroimaging studies from an existing meta-analytic fMRI database demonstrates that this technique can be used to recover known neural bases for multiple psychological functions, suggesting this method’s utility for efficient, high-level meta-analyses of localized brain function. While automated text analytic methods are no replacement for deliberate, manual meta-analyses, they seem to show promise for the efficient aggregation of large bodies of scientific knowledge, at least on a relatively general level.Keywords: FMRI, machine learning, meta-analysis, text analysis
Procedia PDF Downloads 44920327 A Hybrid Derivative-Free Optimization Method for Pass Schedule Calculation in Cold Rolling Mill
Authors: Mohammadhadi Mirmohammadi, Reza Safian, Hossein Haddad
Abstract:
This paper presents an innovative solution for complex multi-objective optimization problem which is a part of efforts toward maximizing rolling mill throughput and minimizing processing costs in tandem cold rolling. This computational intelligence based optimization has been applied to the rolling schedules of tandem cold rolling mill. This method involves the combination of two derivative-free optimization procedures in the form of nested loops. The first optimization loop is based on Improving Hit and Run method which focus on balance of power, force and reduction distribution in rolling schedules. The second loop is a real-coded genetic algorithm based optimization procedure which optimizes energy consumption and productivity. An experimental result of application to five stand tandem cold rolling mill is presented.Keywords: derivative-free optimization, Improving Hit and Run method, real-coded genetic algorithm, rolling schedules of tandem cold rolling mill
Procedia PDF Downloads 69620326 Software Verification of Systematic Resampling for Optimization of Particle Filters
Authors: Osiris Terry, Kenneth Hopkinson, Laura Humphrey
Abstract:
Systematic resampling is the most popularly used resampling method in particle filters. This paper seeks to further the understanding of systematic resampling by defining a formula made up of variables from the sampling equation and the particle weights. The formula is then verified via SPARK, a software verification language. The verified systematic resampling formula states that the minimum/maximum number of possible samples taken of a particle is equal to the floor/ceiling value of particle weight divided by the sampling interval, respectively. This allows for the creation of a randomness spectrum that each resampling method can fall within. Methods on the lower end, e.g., systematic resampling, have less randomness and, thus, are quicker to reach an estimate. Although lower randomness allows for error by having a larger bias towards the size of the weight, having this bias creates vulnerabilities to the noise in the environment, e.g., jamming. Conclusively, this is the first step in characterizing each resampling method. This will allow target-tracking engineers to pick the best resampling method for their environment instead of choosing the most popularly used one.Keywords: SPARK, software verification, resampling, systematic resampling, particle filter, tracking
Procedia PDF Downloads 8420325 Melnikov Analysis for the Chaos of the Nonlocal Nanobeam Resting on Fractional-Order Softening Nonlinear Viscoelastic Foundations
Authors: Guy Joseph Eyebe, Gambo Betchewe, Alidou Mohamadou, Timoleon Crepin Kofane
Abstract:
In the present study, the dynamics of nanobeam resting on fractional order softening nonlinear viscoelastic pasternack foundations is studied. The Hamilton principle is used to derive the nonlinear equation of the motion. Approximate analytical solution is obtained by applying the standard averaging method. The Melnikov method is used to investigate the chaotic behaviors of device, the critical curve separating the chaotic and non-chaotic regions are found. It is shown that appearance of chaos in the system depends strongly on the fractional order parameter.Keywords: chaos, fractional-order, Melnikov method, nanobeam
Procedia PDF Downloads 15920324 Numerical Implementation and Testing of Fractioning Estimator Method for the Box-Counting Dimension of Fractal Objects
Authors: Abraham Terán Salcedo, Didier Samayoa Ochoa
Abstract:
This work presents a numerical implementation of a method for estimating the box-counting dimension of self-avoiding curves on a planar space, fractal objects captured on digital images; this method is named fractioning estimator. Classical methods of digital image processing, such as noise filtering, contrast manipulation, and thresholding, among others, are used in order to obtain binary images that are suitable for performing the necessary computations of the fractioning estimator. A user interface is developed for performing the image processing operations and testing the fractioning estimator on different captured images of real-life fractal objects. To analyze the results, the estimations obtained through the fractioning estimator are compared to the results obtained through other methods that are already implemented on different available software for computing and estimating the box-counting dimension.Keywords: box-counting, digital image processing, fractal dimension, numerical method
Procedia PDF Downloads 8320323 Comparative Performance of Artificial Bee Colony Based Algorithms for Wind-Thermal Unit Commitment
Authors: P. K. Singhal, R. Naresh, V. Sharma
Abstract:
This paper presents the three optimization models, namely New Binary Artificial Bee Colony (NBABC) algorithm, NBABC with Local Search (NBABC-LS), and NBABC with Genetic Crossover (NBABC-GC) for solving the Wind-Thermal Unit Commitment (WTUC) problem. The uncertain nature of the wind power is incorporated using the Weibull probability density function, which is used to calculate the overestimation and underestimation costs associated with the wind power fluctuation. The NBABC algorithm utilizes a mechanism based on the dissimilarity measure between binary strings for generating the binary solutions in WTUC problem. In NBABC algorithm, an intelligent scout bee phase is proposed that replaces the abandoned solution with the global best solution. The local search operator exploits the neighboring region of the current solutions, whereas the integration of genetic crossover with the NBABC algorithm increases the diversity in the search space and thus avoids the problem of local trappings encountered with the NBABC algorithm. These models are then used to decide the units on/off status, whereas the lambda iteration method is used to dispatch the hourly load demand among the committed units. The effectiveness of the proposed models is validated on an IEEE 10-unit thermal system combined with a wind farm over the planning period of 24 hours.Keywords: artificial bee colony algorithm, economic dispatch, unit commitment, wind power
Procedia PDF Downloads 37520322 A Relative Entropy Regularization Approach for Fuzzy C-Means Clustering Problem
Authors: Ouafa Amira, Jiangshe Zhang
Abstract:
Clustering is an unsupervised machine learning technique; its aim is to extract the data structures, in which similar data objects are grouped in the same cluster, whereas dissimilar objects are grouped in different clusters. Clustering methods are widely utilized in different fields, such as: image processing, computer vision , and pattern recognition, etc. Fuzzy c-means clustering (fcm) is one of the most well known fuzzy clustering methods. It is based on solving an optimization problem, in which a minimization of a given cost function has been studied. This minimization aims to decrease the dissimilarity inside clusters, where the dissimilarity here is measured by the distances between data objects and cluster centers. The degree of belonging of a data point in a cluster is measured by a membership function which is included in the interval [0, 1]. In fcm clustering, the membership degree is constrained with the condition that the sum of a data object’s memberships in all clusters must be equal to one. This constraint can cause several problems, specially when our data objects are included in a noisy space. Regularization approach took a part in fuzzy c-means clustering technique. This process introduces an additional information in order to solve an ill-posed optimization problem. In this study, we focus on regularization by relative entropy approach, where in our optimization problem we aim to minimize the dissimilarity inside clusters. Finding an appropriate membership degree to each data object is our objective, because an appropriate membership degree leads to an accurate clustering result. Our clustering results in synthetic data sets, gaussian based data sets, and real world data sets show that our proposed model achieves a good accuracy.Keywords: clustering, fuzzy c-means, regularization, relative entropy
Procedia PDF Downloads 25920321 Preparation of Li Ion Conductive Ceramics via Liquid Process
Authors: M. Kotobuki, M. Koishi
Abstract:
Li1.5Al0.5Ti1.5 (PO4)3(LATP) has received much attention as a solid electrolyte for lithium batteries. In this study, the LATP solid electrolyte is prepared by the co-precipitation method using Li3PO4 as a Li source. The LATP is successfully prepared and the Li ion conductivities of bulk (inner crystal) and total (inner crystal and grain boundary) are 1.1 × 10-3 and 1.1 × 10-4 S cm-1, respectively. These values are comparable to the reported values, in which Li2C2O4 is used as the Li source. It is conclude that the LATP solid electrolyte can be prepared by the co-precipitation method using Li3PO4 as the Li source and this procedure has an advantage in mass production over previous procedure using Li2C2O4 because Li3PO4 is lower price reagent compared with Li2C2O4.Keywords: co-precipitation method, lithium battery, NASICON-type electrolyte, solid electrolyte
Procedia PDF Downloads 35220320 Bio-Inspired Design Approach Analysis: A Case Study of Antoni Gaudi and Santiago Calatrava
Authors: Marzieh Imani
Abstract:
Antoni Gaudi and Santiago Calatrava have reputation for designing bio-inspired creative and technical buildings. Even though they have followed different independent approaches towards design, the source of bio-inspiration seems to be common. Taking a closer look at their projects reveals that Calatrava has been influenced by Gaudi in terms of interpreting nature and applying natural principles into the design process. This research firstly discusses the dialogue between Biomimicry and architecture. This review also explores human/nature discourse during the history by focusing on how nature revealed itself to the fine arts. This is explained by introducing naturalism and romantic style in architecture as the outcome of designers’ inclination towards nature. Reviewing the literature, theoretical background and practical illustration of nature have been included. The most dominant practical aspects of imitating nature are form and function. Nature has been reflected in architectural science resulted in shaping different architectural styles such as organic, green, sustainable, bionic, and biomorphic. By defining a set of common aspects of Gaudi and Calatrava‘s design approach and by considering biomimetic design categories (organism, ecosystem, and behaviour as the main division and form, function, process, material, and construction as subdivisions), Gaudi’s and Calatrava’s project have been analysed. This analysis explores if their design approaches are equivalent or different. Based on this analysis, Gaudi’s architecture can be recognised as biomorphic while Calatrava’s projects are literally biomimetic. Referring to these architects, this review suggests a new set of principles by which a bio-inspired project can be determined either biomorphic or biomimetic.Keywords: biomimicry, Calatrava, Gaudi, nature
Procedia PDF Downloads 28820319 An Application of Graph Theory to The Electrical Circuit Using Matrix Method
Authors: Samai'la Abdullahi
Abstract:
A graph is a pair of two set and so that a graph is a pictorial representation of a system using two basic element nodes and edges. A node is represented by a circle (either hallo shade) and edge is represented by a line segment connecting two nodes together. In this paper, we present a circuit network in the concept of graph theory application and also circuit models of graph are represented in logical connection method were we formulate matrix method of adjacency and incidence of matrix and application of truth table.Keywords: euler circuit and path, graph representation of circuit networks, representation of graph models, representation of circuit network using logical truth table
Procedia PDF Downloads 56120318 Numerical Study of 5kW Vertical Axis Wind Turbine Using DOE Method
Authors: Yan-Ting Lin, Wei-Nian Su
Abstract:
The purpose of this paper is to demonstrate the design of 5kW vertical axis wind turbine (VAWT) using DOE method. The NACA0015 airfoil was implemented for the design and 3D simulation. The critical design parameters are chord length, tip speed ratio (TSR), aspect ratio (AR) and pitch angle in this investigation. The RNG k-ε turbulent model and the sliding mesh method are adopted in the CFD simulation. The results show that the model with zero pitch, 0.3 m in chord length, TSR of 3, and AR of 10 demonstrated the optimum aerodynamic power under the uniform 10m/s inlet velocity. The aerodynamic power is 3.61kW and 3.89kW under TSR of 3 and 4 respectively. The aerodynamic power decreased dramatically while TSR increased to 5.Keywords: vertical axis wind turbine, CFD, DOE, VAWT
Procedia PDF Downloads 44120317 Numerical and Experimental Investigations of Cantilever Rectangular Plate Structure on Subsonic Flutter
Authors: Mevlüt Burak Dalmış, Kemal Yaman
Abstract:
In this study, flutter characteristics of cantilever rectangular plate structure under incompressible flow regime are investigated by comparing the results of commercial flutter analysis program ZAERO© with wind tunnel tests conducted in Ankara Wind Tunnel (ART). A rectangular polycarbonate (PC) plate, 5x125x1000 mm in dimensions, is used for both numerical and experimental investigations. Analysis and test results are very compatible with each other. A comparison between two different solution methods (g and k-method) of ZAERO© is also done. It is seen that, k-method gives closer result than the other one. However, g-method results are on conservative side and it is better to use conservative results namely g-method results. Even if the modal analysis results are used for the flutter analysis for this simple structure, a modal test should be conducted in order to validate the modal analysis results to have accurate flutter analysis results for more complicated structures.Keywords: flutter, plate, subsonic flow, wind tunnel
Procedia PDF Downloads 51820316 Settlement Prediction in Cape Flats Sands Using Shear Wave Velocity – Penetration Resistance Correlations
Authors: Nanine Fouche
Abstract:
The Cape Flats is a low-lying sand-covered expanse of approximately 460 square kilometres, situated to the southeast of the central business district of Cape Town in the Western Cape of South Africa. The aeolian sands masking this area are often loose and compressible in the upper 1m to 1.5m of the surface, and there is a general exceedance of the maximum allowable settlement in these sands. The settlement of shallow foundations on Cape Flats sands is commonly predicted using the results of in-situ tests such as the SPT or DPSH due to the difficulty of retrieving undisturbed samples for laboratory testing. Varying degrees of accuracy and reliability are associated with these methods. More recently, shear wave velocity (Vs) profiles obtained from seismic testing, such as continuous surface wave tests (CSW), are being used for settlement prediction. Such predictions have the advantage of considering non-linear stress-strain behaviour of soil and the degradation of stiffness with increasing strain. CSW tests are rarely executed in the Cape Flats, whereas SPT’s are commonly performed. For this reason, and to facilitate better settlement predictions in Cape Flats sand, equations representing shear wave velocity (Vs) as a function of SPT blow count (N60) and vertical effective stress (v’) were generated by statistical regression of site investigation data. To reveal the most appropriate method of overburden correction, analyses were performed with a separate overburden term (Pa/σ’v) as well as using stress corrected shear wave velocity and SPT blow counts (correcting Vs. and N60 to Vs1and (N1)60respectively). Shear wave velocity profiles and SPT blow count data from three sites masked by Cape Flats sands were utilised to generate 80 Vs-SPT N data pairs for analysis. Investigated terrains included sites in the suburbs of Athlone, Muizenburg, and Atlantis, all underlain by windblown deposits comprising fine and medium sand with varying fines contents. Elastic settlement analysis was also undertaken for the Cape Flats sands, using a non-linear stepwise method based on small-strain stiffness estimates, which was obtained from the best Vs-N60 model and compared to settlement estimates using the general elastic solution with stiffness profiles determined using Stroud’s (1989) and Webb’s (1969) SPT N60-E transformation models. Stroud’s method considers strain level indirectly whereasWebb’smethod does not take account of the variation in elastic modulus with strain. The expression of Vs. in terms of N60 and Pa/σv’ derived from the Atlantis data set revealed the best fit with R2 = 0.83 and a standard error of 83.5m/s. Less accurate Vs-SPT N relations associated with the combined data set is presumably the result of inversion routines used in the analysis of the CSW results showcasing significant variation in relative density and stiffness with depth. The regression analyses revealed that the inclusion of a separate overburden term in the regression of Vs and N60, produces improved fits, as opposed to the stress corrected equations in which the R2 of the regression is notably lower. It is the correction of Vs and N60 to Vs1 and (N1)60 with empirical constants ‘n’ and ‘m’ prior to regression, that introduces bias with respect to overburden pressure. When comparing settlement prediction methods, both Stroud’s method (considering strain level indirectly) and the small strain stiffness method predict higher stiffnesses for medium dense and dense profiles than Webb’s method, which takes no account of strain level in the determination of soil stiffness. Webb’s method appears to be suitable for loose sands only. The Versak software appears to underestimate differences in settlement between square and strip footings of similar width. In conclusion, settlement analysis using small-strain stiffness data from the proposed Vs-N60 model for Cape Flats sands provides a way to take account of the non-linear stress-strain behaviour of the sands when calculating settlement.Keywords: sands, settlement prediction, continuous surface wave test, small-strain stiffness, shear wave velocity, penetration resistance
Procedia PDF Downloads 175