Search results for: shape error
3701 Performance Evaluation and Dear Based Optimization on Machining Leather Specimens to Reduce Carbonization
Authors: Khaja Moiduddin, Tamer Khalaf, Muthuramalingam Thangaraj
Abstract:
Due to the variety of benefits over traditional cutting techniques, the usage of laser cutting technology has risen substantially in recent years. Hot wire machining can cut the leather in the required shape by controlling the wire by generating thermal energy. In the present study, an attempt has been made to investigate the effects of performance measures in the hot wire machining process on cutting leather specimens. Carbonization and material removal rates were considered as quality indicators. Burning leather during machining might cause carbon particles, reducing product quality. Minimizing the effect of carbon particles is crucial for assuring operator and environmental safety, health, and product quality. Hot wire machining can efficiently cut the specimens by controlling the current through it. Taguchi- DEAR-based optimization was also performed in the process, which resulted in a required Carbonization and material removal rate. Using the DEAR approach, the optimal parameters of the present study were found with 3.7% prediction error accuracy.Keywords: cabronization, leather, MRR, current
Procedia PDF Downloads 643700 Identification of Architectural Design Error Risk Factors in Construction Projects Using IDEF0 Technique
Authors: Sahar Tabarroki, Ahad Nazari
Abstract:
The design process is one of the most key project processes in the construction industry. Although architects have the responsibility to produce complete, accurate, and coordinated documents, architectural design is accompanied by many errors. A design error occurs when the constraints and requirements of the design are not satisfied. Errors are potentially costly and time-consuming to correct if not caught early during the design phase, and they become expensive in either construction documents or in the construction phase. The aim of this research is to identify the risk factors of architectural design errors, so identification of risks is necessary. First, a literature review in the design process was conducted and then a questionnaire was designed to identify the risks and risk factors. The questions in the form of the questionnaire were based on the “similar service description of study and supervision of architectural works” published by “Vice Presidency of Strategic Planning & Supervision of I.R. Iran” as the base of architects’ tasks. Second, the top 10 risks of architectural activities were identified. To determine the positions of possible causes of risks with respect to architectural activities, these activities were located in a design process modeled by the IDEF0 technique. The research was carried out by choosing a case study, checking the design drawings, interviewing its architect and client, and providing a checklist in order to identify the concrete examples of architectural design errors. The results revealed that activities such as “defining the current and future requirements of the project”, “studies and space planning,” and “time and cost estimation of suggested solution” has a higher error risk than others. Moreover, the most important causes include “unclear goals of a client”, “time force by a client”, and “lack of knowledge of architects about the requirements of end-users”. For error detecting in the case study, lack of criteria, standards and design criteria, and lack of coordination among them, was a barrier, anyway, “lack of coordination between architectural design and electrical and mechanical facility”, “violation of the standard dimensions and sizes in space designing”, “design omissions” were identified as the most important design errors.Keywords: architectural design, design error, risk management, risk factor
Procedia PDF Downloads 1303699 Maximum Initial Input Allowed to Iterative Learning Control Set-up Using Singular Values
Authors: Naser Alajmi, Ali Alobaidly, Mubarak Alhajri, Salem Salamah, Muhammad Alsubaie
Abstract:
Iterative Learning Control (ILC) known to be a controlling tool to overcome periodic disturbances for repetitive systems. This technique is required to let the error signal tends to zero as the number of operation increases. The learning process that lies within this context is strongly dependent on the initial input which if selected properly tends to let the learning process be more effective compared to the case where a system starts from blind. ILC uses previous recorded execution data to update the following execution/trial input such that a reference trajectory is followed to a high accuracy. Error convergence in ILC is generally highly dependent on the input applied to a plant for trial $1$, thus a good choice of initial starting input signal would make learning faster and as a consequence the error tends to zero faster as well. In the work presented within, an upper limit based on the Singular Values Principle (SV) is derived for the initial input signal applied at trial $1$ such that the system follow the reference in less number of trials without responding aggressively or exceeding the working envelope where a system is required to move within in a robot arm, for example. Simulation results presented illustrate the theory introduced within this paper.Keywords: initial input, iterative learning control, maximum input, singular values
Procedia PDF Downloads 2413698 The Non-Existence of Perfect 2-Error Correcting Lee Codes of Word Length 7 over Z
Authors: Catarina Cruz, Ana Breda
Abstract:
Tiling problems have been capturing the attention of many mathematicians due to their real-life applications. In this study, we deal with tilings of Zⁿ by Lee spheres, where n is a positive integer number, being these tilings related with error correcting codes on the transmission of information over a noisy channel. We focus our attention on the question ‘for what values of n and r does the n-dimensional Lee sphere of radius r tile Zⁿ?’. It seems that the n-dimensional Lee sphere of radius r does not tile Zⁿ for n ≥ 3 and r ≥ 2. Here, we prove that is not possible to tile Z⁷ with Lee spheres of radius 2 presenting a proof based on a combinatorial method and faithful to the geometric idea of the problem. The non-existence of such tilings has been studied by several authors being considered the most difficult cases those in which the radius of the Lee spheres is equal to 2. The relation between these tilings and error correcting codes is established considering the center of a Lee sphere as a codeword and the other elements of the sphere as words which are decoded by the central codeword. When the Lee spheres of radius r centered at elements of a set M ⊂ Zⁿ tile Zⁿ, M is a perfect r-error correcting Lee code of word length n over Z, denoted by PL(n, r). Our strategy to prove the non-existence of PL(7, 2) codes are based on the assumption of the existence of such code M. Without loss of generality, we suppose that O ∈ M, where O = (0, ..., 0). In this sense and taking into account that we are dealing with Lee spheres of radius 2, O covers all words which are distant two or fewer units from it. By the definition of PL(7, 2) code, each word which is distant three units from O must be covered by a unique codeword of M. These words have to be covered by codewords which dist five units from O. We prove the non-existence of PL(7, 2) codes showing that it is not possible to cover all the referred words without superposition of Lee spheres whose centers are distant five units from O, contradicting the definition of PL(7, 2) code. We achieve this contradiction by combining the cardinality of particular subsets of codewords which are distant five units from O. There exists an extensive literature on codes in the Lee metric. Here, we present a new approach to prove the non-existence of PL(7, 2) codes.Keywords: Golomb-Welch conjecture, Lee metric, perfect Lee codes, tilings
Procedia PDF Downloads 1603697 Assessment of Time-variant Work Stress for Human Error Prevention
Authors: Hyeon-Kyo Lim, Tong-Il Jang, Yong-Hee Lee
Abstract:
For an operator in a nuclear power plant, human error is one of the most dreaded factors that may result in unexpected accidents. The possibility of human errors may be low, but the risk of them would be unimaginably enormous. Thus, for accident prevention, it is quite indispensable to analyze the influence of any factors which may raise the possibility of human errors. During the past decades, not a few research results showed that performance of human operators may vary over time due to lots of factors. Among them, stress is known to be an indirect factor that may cause human errors and result in mental illness. Until now, not a few assessment tools have been developed to assess stress level of human workers. However, it still is questionable to utilize them for human performance anticipation which is related with human error possibility, because they were mainly developed from the viewpoint of mental health rather than industrial safety. Stress level of a person may go up or down with work time. In that sense, if they would be applicable in the safety aspect, they should be able to assess the variation resulted from work time at least. Therefore, this study aimed to compare their applicability for safety purpose. More than 10 kinds of work stress tools were analyzed with reference to assessment items, assessment and analysis methods, and follow-up measures which are known to close related factors with work stress. The results showed that most tools mainly focused their weights on some common organizational factors such as demands, supports, and relationships, in sequence. Their weights were broadly similar. However, they failed to recommend practical solutions. Instead, they merely advised to set up overall counterplans in PDCA cycle or risk management activities which would be far from practical human error prevention. Thus, it was concluded that application of stress assessment tools mainly developed for mental health seemed to be impractical for safety purpose with respect to human performance anticipation, and that development of a new assessment tools would be inevitable if anyone wants to assess stress level in the aspect of human performance variation and accident prevention. As a consequence, as practical counterplans, this study proposed a new scheme for assessment of work stress level of a human operator that may vary over work time which is closely related with the possibility of human errors.Keywords: human error, human performance, work stress, assessment tool, time-variant, accident prevention
Procedia PDF Downloads 6723696 Banking Sector Development and Economic Growth: Evidence from the State of Qatar
Authors: Fekri Shawtari
Abstract:
The banking sector plays a very crucial role in the economic development of the country. As a financial intermediary, it has assigned a great role in the economic growth and stability. This paper aims to examine the empirically the relationship between banking industry and economic growth in state of Qatar. We adopt the VAR vector error correction model (VECM) along with Granger causality to address the issue over the long-run and short-run between the banking sector and economic growth. It is expected that the results will give policy directions to the policymakers to make strategies that are conducive toward boosting development to achieve the targeted economic growth in current situation.Keywords: economic growth, banking sector, Qatar, vector error correction model, VECM
Procedia PDF Downloads 1703695 Virtual Assessment of Measurement Error in the Fractional Flow Reserve
Authors: Keltoum Chahour, Mickael Binois
Abstract:
Due to a lack of standardization during the invasive fractional flow reserve (FFR) procedure, the index is subject to many sources of uncertainties. In this paper, we investigate -through simulation- the effect of the (FFR) device position and configuration on the obtained value of the (FFR) fraction. For this purpose, we use computational fluid dynamics (CFD) in a 3D domain corresponding to a diseased arterial portion. The (FFR) pressure captor is introduced inside it with a given length and coefficient of bending to capture the (FFR) value. To get over the computational limitations, basically, the time of the simulation is about 2h 15min for one (FFR) value; we generate a Gaussian Process (GP) model for (FFR) prediction. The (GP) model indicates good accuracy and demonstrates the effective error in the measurement created by the random configuration of the pressure captor.Keywords: fractional flow reserve, Gaussian processes, computational fluid dynamics, drift
Procedia PDF Downloads 1343694 Advanced Approach to Analysis the Thin Strip Profile in Cold Rolling of Pair Roll Crossing and Shifting Mill Using an Arbitrary Lagrangian-Eulerian Technique
Authors: Abdulrahman Aljabri, Essam R. I. Mahmoud, Hamad Almohamedi, Zhengyi Jiang
Abstract:
Cold rolled thin strip has received intensive attention through technological and theoretical progress in the rolling process, as well as researchers have focused on its control during rolling as an essential parameter for producing thinner strip with good shape and profile. An advanced approach has been proposed to analysis the thin strip profile in cold rolling of pair roll crossing and shifting mill using Finite Element Analysis (FEA) with an ALE technique. The ALE (Arbitrary Lagrangian-Eulerian) techniques to enable more flexibility of the ALE technique in the adjustment of the finite element mesh, which provides a significant tool for simulating the thin strip under realistic rolling process constraint and provide accurate model results. The FEA can provide theoretical basis for the 3D model of controlling the strip shape and profile in thin strip rolling, and deliver an optimal rolling process parameter, and suggest corrective changes during cold rolling of thin strip.Keywords: pair roll crossing, work roll shifting, strip shape and profile, finite element modeling
Procedia PDF Downloads 963693 Modeling Visual Memorability Assessment with Autoencoders Reveals Characteristics of Memorable Images
Authors: Elham Bagheri, Yalda Mohsenzadeh
Abstract:
Image memorability refers to the phenomenon where certain images are more likely to be remembered by humans than others. It is a quantifiable and intrinsic attribute of an image. Understanding how visual perception and memory interact is important in both cognitive science and artificial intelligence. It reveals the complex processes that support human cognition and helps to improve machine learning algorithms by mimicking the brain's efficient data processing and storage mechanisms. To explore the computational underpinnings of image memorability, this study examines the relationship between an image's reconstruction error, distinctiveness in latent space, and its memorability score. A trained autoencoder is used to replicate human-like memorability assessment inspired by the visual memory game employed in memorability estimations. This study leverages a VGG-based autoencoder that is pre-trained on the vast ImageNet dataset, enabling it to recognize patterns and features that are common to a wide and diverse range of images. An empirical analysis is conducted using the MemCat dataset, which includes 10,000 images from five broad categories: animals, sports, food, landscapes, and vehicles, along with their corresponding memorability scores. The memorability score assigned to each image represents the probability of that image being remembered by participants after a single exposure. The autoencoder is finetuned for one epoch with a batch size of one, attempting to create a scenario similar to human memorability experiments where memorability is quantified by the likelihood of an image being remembered after being seen only once. The reconstruction error, which is quantified as the difference between the original and reconstructed images, serves as a measure of how well the autoencoder has learned to represent the data. The reconstruction error of each image, the error reduction, and its distinctiveness in latent space are calculated and correlated with the memorability score. Distinctiveness is measured as the Euclidean distance between each image's latent representation and its nearest neighbor within the autoencoder's latent space. Different structural and perceptual loss functions are considered to quantify the reconstruction error. The results indicate that there is a strong correlation between the reconstruction error and the distinctiveness of images and their memorability scores. This suggests that images with more unique distinct features that challenge the autoencoder's compressive capacities are inherently more memorable. There is also a negative correlation between the reduction in reconstruction error compared to the autoencoder pre-trained on ImageNet, which suggests that highly memorable images are harder to reconstruct, probably due to having features that are more difficult to learn by the autoencoder. These insights suggest a new pathway for evaluating image memorability, which could potentially impact industries reliant on visual content and mark a step forward in merging the fields of artificial intelligence and cognitive science. The current research opens avenues for utilizing neural representations as instruments for understanding and predicting visual memory.Keywords: autoencoder, computational vision, image memorability, image reconstruction, memory retention, reconstruction error, visual perception
Procedia PDF Downloads 903692 Cellular Components of the Hemal Node of Egyptian Cattle
Authors: Amira E. Derbalah, Doaa M. Zaghloul
Abstract:
10 clinically healthy hemal nodes were collected from male bulls aged 2-3 years. Light microscopy revealed a capsule of connective tissue consisted mainly of collagen fiber surrounding hemal node, numerous erythrocytes were found in wide subcapsular sinus under the capsule. The parenchyma of the hemal node was divided into cortex and medulla. Diffused lymphocytes, and lymphoid follicles, having germinal centers were the main components of the cortex, while in the medulla there was wide medullary sinus, diffused lymphocytes and few lymphoid nodules. The area occupied with lymph nodules was larger than that occupied with non-nodular structure of lymphoid cords and blood sinusoids. Electron microscopy revealed the cellular components of hemal node including elements of circulating erythrocytes intermingled with lymphocytes, plasma cells, mast cells, reticular cells, macrophages, megakaryocytes and endothelial cells lining the blood sinuses. The lymphocytes were somewhat triangular in shape with cytoplasmic processes extending between adjacent erythrocytes. Nuclei were triangular to oval in shape, lightly stained with clear nuclear membrane indentation and clear nucleoli. The reticular cells were elongated in shape with cytoplasmic processes extending between adjacent lymphocytes, rough endoplasmic reticulum, ribosomes and few lysosomes were seen in their cytoplasm. Nucleus was elongated in shape with less condensed chromatin. Plasma cells were oval to irregular in shape with numerous dilated rough endoplasmic reticulum containing electron lucent material occupying the whole cytoplasm and few mitochondria were found. Nuclei were centrally located and oval in shape with heterochromatin emarginated and often clumped near the nuclear membrane. Occasionally megakaryocytes and mast cells were seen among lymphocytes. Megakaryocytes had multilobulated nucleus and free ribosomes often appearing as small aggregates in their cytoplasm, while mast cell had their characteristic electron dense granule in the cytoplasm, few electron lucent granules were found also, we conclude that, the main function of the hemal node of cattle is proliferation of lymphocytes. No role for plasma cell in erythrophagocytosis could be suggested.Keywords: cattle, electron microscopy, hemal node, histology, immune system
Procedia PDF Downloads 4023691 Shape Memory Alloy Structural Damper Manufactured by Selective Laser Melting
Authors: Tiziana Biasutti, Daniela Rigamonti, Lorenzo Palmiotti, Adelaide Nespoli, Paolo Bettini
Abstract:
Aerospace industry is based on the continuous development of new technologies and solutions that allows constant improvement of the systems. Shape Memory Alloys are smart materials that can be used as dampers due to their pseudoelastic effect. The purpose of the research was to design a passive damper in Nitinol, manufactured by Selective Laser Melting, for space applications to reduce vibration between different structural parts in space structures. The powder is NiTi (50.2 at.% of Ni). The structure manufactured by additive technology allows us to eliminate the presence of joint and moving parts and to have a compact solution with high structural strength. The designed dampers had single or double cell structures with three different internal angles (30°, 45° and 60°). This particular shape has damping properties also without the pseudoelastic effect. For this reason, the geometries were reproduced in different materials, SS316L and Ti6Al4V, to test the geometry loss factor. The mechanical performances of these specimens were compared to the ones of NiTi structures, pointing out good damping properties of the designed structure and the highest performances of the NiTi pseudoelastic effect. The NiTi damper was mechanically characterized by static and dynamic tests and with DSC and microscope observations. The experimental results were verified with numerical models and with some scaled steel specimens in which optical fibers were embedded. The realized structure presented good mechanical and damping properties. It was observed that the loss factor and the dissipated energy increased with the angles of the cells.Keywords: additive manufacturing, damper, nitinol, pseudo elastic effect, selective laser melting, shape memory alloys
Procedia PDF Downloads 1073690 Financial Inclusion for Inclusive Growth in an Emerging Economy
Authors: Godwin Chigozie Okpara, William Chimee Nwaoha
Abstract:
The paper set out to stress on how financial inclusion index could be calculated and also investigated the impact of inclusive finance on inclusive growth in an emerging economy. In the light of these objectives, chi-wins method was used to calculate indexes of financial inclusion while co-integration and error correction model were used for evaluation of the impact of financial inclusion on inclusive growth. The result of the analysis revealed that financial inclusion while having a long-run relationship with GDP growth is an insignificant function of the growth of the economy. The speed of adjustment is correctly signed and significant. On the basis of these results, the researchers called for tireless efforts of government and banking sector in promoting financial inclusion in developing countries.Keywords: chi-wins index, co-integration, error correction model, financial inclusion
Procedia PDF Downloads 6533689 The Underestimate of the Annual Maximum Rainfall Depths Due to Coarse Time Resolution Data
Authors: Renato Morbidelli, Carla Saltalippi, Alessia Flammini, Tommaso Picciafuoco, Corrado Corradini
Abstract:
A considerable part of rainfall data to be used in the hydrological practice is available in aggregated form within constant time intervals. This can produce undesirable effects, like the underestimate of the annual maximum rainfall depth, Hd, associated with a given duration, d, that is the basic quantity in the development of rainfall depth-duration-frequency relationships and in determining if climate change is producing effects on extreme event intensities and frequencies. The errors in the evaluation of Hd from data characterized by a coarse temporal aggregation, ta, and a procedure to reduce the non-homogeneity of the Hd series are here investigated. Our results indicate that: 1) in the worst conditions, for d=ta, the estimation of a single Hd value can be affected by an underestimation error up to 50%, while the average underestimation error for a series with at least 15-20 Hd values, is less than or equal to 16.7%; 2) the underestimation error values follow an exponential probability density function; 3) each very long time series of Hd contains many underestimated values; 4) relationships between the non-dimensional ratio ta/d and the average underestimate of Hd, derived from continuous rainfall data observed in many stations of Central Italy, may overcome this issue; 5) these equations should allow to improve the Hd estimates and the associated depth-duration-frequency curves at least in areas with similar climatic conditions.Keywords: central Italy, extreme events, rainfall data, underestimation errors
Procedia PDF Downloads 1913688 MCERTL: Mutation-Based Correction Engine for Register-Transfer Level Designs
Authors: Khaled Salah
Abstract:
In this paper, we present MCERTL (mutation-based correction engine for RTL designs) as an automatic error correction technique based on mutation analysis. A mutation-based correction methodology is proposed to automatically fix the erroneous RTL designs. The proposed strategy combines the processes of mutation and assertion-based localization. The erroneous statements are mutated to produce possible fixes for the failed RTL code. A concurrent mutation engine is proposed to mitigate the computational cost of running sequential mutants operators. The proposed methodology is evaluated against some benchmarks. The experimental results demonstrate that our proposed method enables us to automatically locate and correct multiple bugs at reasonable time.Keywords: bug localization, error correction, mutation, mutants
Procedia PDF Downloads 2803687 An Application of Modified M-out-of-N Bootstrap Method to Heavy-Tailed Distributions
Authors: Hannah F. Opayinka, Adedayo A. Adepoju
Abstract:
This study is an extension of a prior study on the modification of the existing m-out-of-n (moon) bootstrap method for heavy-tailed distributions in which modified m-out-of-n (mmoon) was proposed as an alternative method to the existing moon technique. In this study, both moon and mmoon techniques were applied to two real income datasets which followed Lognormal and Pareto distributions respectively with finite variances. The performances of these two techniques were compared using Standard Error (SE) and Root Mean Square Error (RMSE). The findings showed that mmoon outperformed moon bootstrap in terms of smaller SEs and RMSEs for all the sample sizes considered in the two datasets.Keywords: Bootstrap, income data, lognormal distribution, Pareto distribution
Procedia PDF Downloads 1863686 Comparison between Separable and Irreducible Goppa Code in McEliece Cryptosystem
Authors: Newroz Nooralddin Abdulrazaq, Thuraya Mahmood Qaradaghi
Abstract:
The McEliece cryptosystem is an asymmetric type of cryptography based on error correction code. The classical McEliece used irreducible binary Goppa code which considered unbreakable until now especially with parameter [1024, 524, and 101], but it is suffering from large public key matrix which leads to be difficult to be used practically. In this work Irreducible and Separable Goppa codes have been introduced. The Irreducible and Separable Goppa codes used are with flexible parameters and dynamic error vectors. A Comparison between Separable and Irreducible Goppa code in McEliece Cryptosystem has been done. For encryption stage, to get better result for comparison, two types of testing have been chosen; in the first one the random message is constant while the parameters of Goppa code have been changed. But for the second test, the parameters of Goppa code are constant (m=8 and t=10) while the random message have been changed. The results show that the time needed to calculate parity check matrix in separable are higher than the one for irreducible McEliece cryptosystem, which is considered expected results due to calculate extra parity check matrix in decryption process for g2(z) in separable type, and the time needed to execute error locator in decryption stage in separable type is better than the time needed to calculate it in irreducible type. The proposed implementation has been done by Visual studio C#.Keywords: McEliece cryptosystem, Goppa code, separable, irreducible
Procedia PDF Downloads 2663685 Shape Sensing and Damage Detection of Thin-Walled Cylinders Using an Inverse Finite Element Method
Authors: Ionel D. Craiu, Mihai Nedelcu
Abstract:
Thin-walled cylinders are often used by the offshore industry as columns of floating installations. Based on observed strains, the inverse Finite Element Method (iFEM) may rebuild the deformation of structures. Structural Health Monitoring uses this approach extensively. However, the number of in-situ strain gauges is what determines how accurate it is, and for shell structures with complicated deformation, this number can easily become too high for practical use. Any thin-walled beam member's complicated deformation can be modeled by the Generalized Beam Theory (GBT) as a linear combination of pre-specified cross-section deformation modes. GBT uses bar finite elements as opposed to shell finite elements. This paper proposes an iFEM/GBT formulation for the shape sensing of thin-walled cylinders based on these benefits. This method significantly reduces the number of strain gauges compared to using the traditional inverse-shell finite elements. Using numerical simulations, dent damage detection is achieved by comparing the strain distributions of the undamaged and damaged members. The effect of noise on strain measurements is also investigated.Keywords: damage detection, generalized beam theory, inverse finite element method, shape sensing
Procedia PDF Downloads 1133684 Effect of Plastic Deformation on the Carbide-Free Bainite Transformation in Medium C-Si Steel
Authors: Mufath Zorgani, Carlos Garcia-Mateo, Mohammad Jahazi
Abstract:
In this study, the influence of pre-strained austenite on the extent of isothermal bainite transformation in medium-carbon, high-silicon steel was investigated. Different amounts of deformations were applied at 600°C on the austenite right before quenching to the region, where isothermal bainitic transformation is activated. Four different temperatures of 325, 350, 375, and 400°C considering similar holding time 1800s at each temperature, were selected to investigate the extent of isothermal bainitic transformation. The results showed that the deformation-free austenite transforms to the higher volume fraction of CFB bainite when the isothermal transformation temperature reduced from 400 to 325°C, the introduction of plastic deformation in austenite prior to the formation of bainite invariably involves a delay of the same or identical isothermal treatment. On the other side, when the isothermal transformation temperature and deformation increases, the volume fraction and the plate thickness of bainite decreases and the amount of retained austenite increases. The shape of retained austenite is mostly representing blocky-shape one due to the less amount of transformed bainite. Moreover, the plate-like shape bainite cannot be resolved when the deformation amount reached 30%, and the isothermal transformation temperatures are of 375 and 400°C. The amount of retained austenite and the percentage of its transformation to martensite during the final cooling stage play a significant role in the variation of hardness level for different thermomechanical regimes.Keywords: ausforming, carbide free bainite, dilatometry, microstructure
Procedia PDF Downloads 1283683 An Analysis of Transition in Building Form from Abolition of Diagonal Plane Control by Street Width: Focusing on Site Plan and Urban Block
Authors: Joohyun Park, Jin Baek
Abstract:
The purpose of this study is to Analyze the role and effect arise from Diagonal Plane Control by Street Width (DPCSW) in Architecture in Seoul, and to predict the aspect of transition about the relationship among buildings and Urban morphology After the abolition. To find the tendency of building shape regulation, This study review Building Acts concerned with form making (the building to land Ratio, building designated line, wall designated line, building height limit (DPCSW) and etc.) and simulate the shape of urban blocks made by Acts in drawings. The review results show DPCSW is not only limitation about height, but also making the building setback from road and make the Road broader. And it makes the typical shape of the urban block that buildings are moving away from surrounding road After the Abolition of DPCSW; it is expected by the legislature that domestic real estate’s market would be promoted by increased total floor areas in each building. Some substitution from the legislature is announced, but it just deals with Building Maximum unit by Block unit except the regulation about arrangement in urban Figure and Ground. In conclusion, refrain from the uncontrolled development of city, It is important to make regulation about not only height factors but limitation line in land. Furthermore, through revising District Unit Plan, It is positively necessary to reset the relationship between buildings for the making the city space better.Keywords: diagonal plane control by street width, building maximum height, district unit plan, building acts, urban block type, morphology, building shape
Procedia PDF Downloads 3123682 Selection of Rayleigh Damping Coefficients for Seismic Response Analysis of Soil Layers
Authors: Huai-Feng Wang, Meng-Lin Lou, Ru-Lin Zhang
Abstract:
One good analysis method in seismic response analysis is direct time integration, which widely adopts Rayleigh damping. An approach is presented for selection of Rayleigh damping coefficients to be used in seismic analyses to produce a response that is consistent with Modal damping response. In the presented approach, the expression of the error of peak response, acquired through complete quadratic combination method, and Rayleigh damping coefficients was set up and then the coefficients were produced by minimizing the error. Two finite element modes of soil layers, excited by 28 seismic waves, were used to demonstrate the feasibility and validity.Keywords: Rayleigh damping, modal damping, damping coefficients, seismic response analysis
Procedia PDF Downloads 4383681 Design of a Standard Weather Data Acquisition Device for the Federal University of Technology, Akure Nigeria
Authors: Isaac Kayode Ogunlade
Abstract:
Data acquisition (DAQ) is the process by which physical phenomena from the real world are transformed into an electrical signal(s) that are measured and converted into a digital format for processing, analysis, and storage by a computer. The DAQ is designed using PIC18F4550 microcontroller, communicating with Personal Computer (PC) through USB (Universal Serial Bus). The research deployed initial knowledge of data acquisition system and embedded system to develop a weather data acquisition device using LM35 sensor to measure weather parameters and the use of Artificial Intelligence(Artificial Neural Network - ANN)and statistical approach(Autoregressive Integrated Moving Average – ARIMA) to predict precipitation (rainfall). The device is placed by a standard device in the Department of Meteorology, Federal University of Technology, Akure (FUTA) to know the performance evaluation of the device. Both devices (standard and designed) were subjected to 180 days with the same atmospheric condition for data mining (temperature, relative humidity, and pressure). The acquired data is trained in MATLAB R2012b environment using ANN, and ARIMAto predict precipitation (rainfall). Root Mean Square Error (RMSE), Mean Absolute Error (MAE), Correction Square (R2), and Mean Percentage Error (MPE) was deplored as standardize evaluation to know the performance of the models in the prediction of precipitation. The results from the working of the developed device show that the device has an efficiency of 96% and is also compatible with Personal Computer (PC) and laptops. The simulation result for acquired data shows that ANN models precipitation (rainfall) prediction for two months (May and June 2017) revealed a disparity error of 1.59%; while ARIMA is 2.63%, respectively. The device will be useful in research, practical laboratories, and industrial environments.Keywords: data acquisition system, design device, weather development, predict precipitation and (FUTA) standard device
Procedia PDF Downloads 913680 Design of H-Shape X-band Application Electrically Small Antenna
Authors: Riki H. Patel, Arpan H. Desai, Trushit Upadhyaya
Abstract:
This paper presents a new small electrically antenna rectangular X- band micro-strip patch antenna loaded with material Rogers RT/duroid 5870 (tm). The present discussion focuses on small Electrically antenna which are electrically small compared to wave length the performance of electrically small antenna are closely related to their electrical size, the gain can be increased to maintain the efficiency of the radiator. Basically micro-strip Patch antennas have been used in satellite communications and for their good characteristics such as lightness, low cost, and so on. Here in the design H- shape folded dipole, which increase the band width of the antenna.Keywords: electrically small antennas, X-band application, antenna, micro-strip patch, frequency antenna, feed, gain
Procedia PDF Downloads 4663679 Aerodynamic Analysis of Multiple Winglets for Aircrafts
Authors: S. Pooja Pragati, B. Sudarsan, S. Raj Kumar
Abstract:
This paper provides a practical design of a new concept of massive Induced Drag reductions of stream vise staggered multiple winglets. It is designed to provide an optimum performance of a winglet from conventional designs. In preparing for a mechanical design, aspects such as shape, dimensions are analyzed to yield a huge amount of reduction in fuel consumption and increased performance. Owing to its simplicity of application and effectiveness we believe that it will enable us to consider its enhanced version for the grid effect of the staggered multiple winglets on the deflected mass flow of the wing system. The objective of the analysis were to compare the aerodynamic characteristics of two winglet configuration and to investigate the performance of two winglets shape simulated at selected cant angle of 0,45,60 degree.Keywords: multiple winglets, induced drag, aerodynamics analysis, low speed aircrafts
Procedia PDF Downloads 4803678 Hybrid Localization Schemes for Wireless Sensor Networks
Authors: Fatima Babar, Majid I. Khan, Malik Najmus Saqib, Muhammad Tahir
Abstract:
This article provides range based improvements over a well-known single-hop range free localization scheme, Approximate Point in Triangulation (APIT) by proposing an energy efficient Barycentric coordinate based Point-In-Triangulation (PIT) test along with PIT based trilateration. These improvements result in energy efficiency, reduced localization error and improved localization coverage compared to APIT and its variants. Moreover, we propose to embed Received signal strength indication (RSSI) based distance estimation in DV-Hop which is a multi-hop localization scheme. The proposed localization algorithm achieves energy efficiency and reduced localization error compared to DV-Hop and its available improvements. Furthermore, a hybrid multi-hop localization scheme is also proposed that utilize Barycentric coordinate based PIT test and both range based (Received signal strength indicator) and range free (hop count) techniques for distance estimation. Our experimental results provide evidence that proposed hybrid multi-hop localization scheme results in two to five times reduction in the localization error compare to DV-Hop and its variants, at reduced energy requirements.Keywords: Localization, Trilateration, Triangulation, Wireless Sensor Networks
Procedia PDF Downloads 4673677 New HCI Design Process Education
Authors: Jongwan Kim
Abstract:
Human Computer Interaction (HCI) is a subject covering the study, plan, and design of interactions between humans and computers. The prevalent use of digital mobile devices is increasing the need for education and research on HCI. This work is focused on a new education method geared towards reducing errors while developing application programs that incorporate role-changing brainstorming techniques during HCI design process. The proposed method has been applied to a capstone design course in the last spring semester. Students discovered some examples about UI design improvement and their error discovering and reducing capability was promoted. An UI design improvement, PC voice control for people with disabilities as an assistive technology examplar, will be presented. The improvement of these students' design ability will be helpful to the real field work.Keywords: HCI, design process, error reducing education, role-changing brainstorming, assistive technology
Procedia PDF Downloads 4903676 Income-Consumption Relationships in Pakistan (1980-2011): A Cointegration Approach
Authors: Himayatullah Khan, Alena Fedorova
Abstract:
The present paper analyses the income-consumption relationships in Pakistan using annual time series data from 1980-81 to 2010-1. The paper uses the Augmented Dickey-Fuller test to check the unit root and stationarity in these two time series. The paper finds that the two time series are nonstationary but stationary at their first difference levels. The Augmented Engle-Granger test and the Cointegrating Regression Durbin-Watson test imply that the two time series of consumption and income are cointegrated and that long-run marginal propensity to consume is 0.88 which is given by the estimated (static) equilibrium relation. The paper also used the error correction mechanism to find out to model dynamic relationship. The purpose of the ECM is to indicate the speed of adjustment from the short-run equilibrium to the long-run equilibrium state. The results show that MPC is equal to 0.93 and is highly significant. The coefficient of Engle-Granger residuals is negative but insignificant. Statistically, the equilibrium error term is zero, which suggests that consumption adjusts to changes in GDP in the same period. The short-run changes in GDP have a positive impact on short-run changes in consumption. The paper concludes that we may interpret 0.93 as the short-run MPC. The pair-wise Granger Causality test shows that both GDP and consumption Granger cause each other.Keywords: cointegrating regression, Augmented Dickey Fuller test, Augmented Engle-Granger test, Granger causality, error correction mechanism
Procedia PDF Downloads 4143675 EMI Radiation Prediction and Final Measurement Process Optimization by Neural Network
Authors: Hussam Elias, Ninovic Perez, Holger Hirsch
Abstract:
The completion of the EMC regulations worldwide is growing steadily as the usage of electronics in our daily lives is increasing more than ever. In this paper, we introduce a novel method to perform the final phase of Electromagnetic compatibility (EMC) measurement and to reduce the required test time according to the norm EN 55032 by using a developed tool and the conventional neural network(CNN). The neural network was trained using real EMC measurements, which were performed in the Semi Anechoic Chamber (SAC) by CETECOM GmbH in Essen, Germany. To implement our proposed method, we wrote software to perform the radiated electromagnetic interference (EMI) measurements and use the CNN to predict and determine the position of the turntable that meets the maximum radiation value.Keywords: conventional neural network, electromagnetic compatibility measurement, mean absolute error, position error
Procedia PDF Downloads 2003674 Optimal Sensing Technique for Estimating Stress Distribution of 2-D Steel Frame Structure Using Genetic Algorithm
Authors: Jun Su Park, Byung Kwan Oh, Jin Woo Hwang, Yousok Kim, Hyo Seon Park
Abstract:
For the structural safety, the maximum stress calculated from the stress distribution of a structure is widely used. The stress distribution can be estimated by deformed shape of the structure obtained from measurement. Although the estimation of stress is strongly affected by the location and number of sensing points, most studies have conducted the stress estimation without reasonable basis on sensing plan such as the location and number of sensors. In this paper, an optimal sensing technique for estimating the stress distribution is proposed. This technique proposes the optimal location and number of sensing points for a 2-D frame structure while minimizing the error of stress distribution between analytical model and estimation by cubic smoothing splines using genetic algorithm. To verify the proposed method, the optimal sensor measurement technique is applied to simulation tests on 2-D steel frame structure. The simulation tests are performed under various loading scenarios. Through those tests, the optimal sensing plan for the structure is suggested and verified.Keywords: genetic algorithm, optimal sensing, optimizing sensor placements, steel frame structure
Procedia PDF Downloads 5313673 A Study on Computational Fluid Dynamics (CFD)-Based Design Optimization Techniques Using Multi-Objective Evolutionary Algorithms (MOEA)
Authors: Ahmed E. Hodaib, Mohamed A. Hashem
Abstract:
In engineering applications, a design has to be as fully perfect as possible in some defined case. The designer has to overcome many challenges in order to reach the optimal solution to a specific problem. This process is called optimization. Generally, there is always a function called “objective function” that is required to be maximized or minimized by choosing input parameters called “degrees of freedom” within an allowed domain called “search space” and computing the values of the objective function for these input values. It becomes more complex when we have more than one objective for our design. As an example for Multi-Objective Optimization Problem (MOP): A structural design that aims to minimize weight and maximize strength. In such case, the Pareto Optimal Frontier (POF) is used, which is a curve plotting two objective functions for the best cases. At this point, a designer should make a decision to choose the point on the curve. Engineers use algorithms or iterative methods for optimization. In this paper, we will discuss the Evolutionary Algorithms (EA) which are widely used with Multi-objective Optimization Problems due to their robustness, simplicity, suitability to be coupled and to be parallelized. Evolutionary algorithms are developed to guarantee the convergence to an optimal solution. An EA uses mechanisms inspired by Darwinian evolution principles. Technically, they belong to the family of trial and error problem solvers and can be considered global optimization methods with a stochastic optimization character. The optimization is initialized by picking random solutions from the search space and then the solution progresses towards the optimal point by using operators such as Selection, Combination, Cross-over and/or Mutation. These operators are applied to the old solutions “parents” so that new sets of design variables called “children” appear. The process is repeated until the optimal solution to the problem is reached. Reliable and robust computational fluid dynamics solvers are nowadays commonly utilized in the design and analyses of various engineering systems, such as aircraft, turbo-machinery, and auto-motives. Coupling of Computational Fluid Dynamics “CFD” and Multi-Objective Evolutionary Algorithms “MOEA” has become substantial in aerospace engineering applications, such as in aerodynamic shape optimization and advanced turbo-machinery design.Keywords: mathematical optimization, multi-objective evolutionary algorithms "MOEA", computational fluid dynamics "CFD", aerodynamic shape optimization
Procedia PDF Downloads 2563672 Semiparametric Regression Of Truncated Spline Biresponse On Farmer Loyalty And Attachment Modeling
Authors: Adji Achmad Rinaldo Fernandes
Abstract:
Regression analysis is a statistical method that is able to describe and predict causal relationships between individuals. Not all relationships have a known curve shape; often, there are relationship patterns that cannot be known in the shape of the curve; besides that, a cause can have an impact on more than one effect, so that between effects can also have a close relationship in it. Regression analysis that can be done to find out the relationship can be brought closer to the semiparametric regression of truncated spline biresponse. The purpose of this study is to examine the function estimator and determine the best model of truncated spline biresponse semiparametric regression. The results of the secondary data study showed that the best model with the highest order of quadratic and a maximum of two knots with a Goodness of fit value in the form of Adjusted R2 of 88.5%.Keywords: biresponse, farmer attachment, farmer loyalty, truncated spline
Procedia PDF Downloads 36