Search results for: planar shape models
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 3362

Search results for: planar shape models

1502 Identifying Knowledge Gaps in Incorporating Toxicity of Particulate Matter Constituents for Developing Regulatory Limits on Particulate Matter

Authors: Ananya Das, Arun Kumar, Gazala Habib, Vivekanandan Perumal

Abstract:

Regulatory bodies has proposed limits on Particulate Matter (PM) concentration in air; however, it does not explicitly indicate the incorporation of effects of toxicities of constituents of PM in developing regulatory limits. This study aimed to provide a structured approach to incorporate toxic effects of components in developing regulatory limits on PM. A four-step human health risk assessment framework consists of - (1) hazard identification (parameters: PM and its constituents and their associated toxic effects on health), (2) exposure assessment (parameters: concentrations of PM and constituents, information on size and shape of PM; fate and transport of PM and constituents in respiratory system), (3) dose-response assessment (parameters: reference dose or target toxicity dose of PM and its constituents), and (4) risk estimation (metric: hazard quotient and/or lifetime incremental risk of cancer as applicable). Then parameters required at every step were obtained from literature. Using this information, an attempt has been made to determine limits on PM using component-specific information. An example calculation was conducted for exposures of PM2.5 and its metal constituents from Indian ambient environment to determine limit on PM values. Identified data gaps were: (1) concentrations of PM and its constituents and their relationship with sampling regions, (2) relationship of toxicity of PM with its components.

Keywords: Air, component-specific toxicity, human health risks, particulate matter.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1188
1501 Numerical Investigation of Non Fourier Heat Conduction in a Semi-infinite Body due to a Moving Concentrated Heat Source Composed with Radiational Boundary Condition

Authors: M. Akbari, S. Sadodin

Abstract:

In this paper, the melting of a semi-infinite body as a result of a moving laser beam has been studied. Because the Fourier heat transfer equation at short times and large dimensions does not have sufficient accuracy; a non-Fourier form of heat transfer equation has been used. Due to the fact that the beam is moving in x direction, the temperature distribution and the melting pool shape are not asymmetric. As a result, the problem is a transient threedimensional problem. Therefore, thermophysical properties such as heat conductivity coefficient, density and heat capacity are functions of temperature and material states. The enthalpy technique, used for the solution of phase change problems, has been used in an explicit finite volume form for the hyperbolic heat transfer equation. This technique has been used to calculate the transient temperature distribution in the semi-infinite body and the growth rate of the melt pool. In order to validate the numerical results, comparisons were made with experimental data. Finally, the results of this paper were compared with similar problem that has used the Fourier theory. The comparison shows the influence of infinite speed of heat propagation in Fourier theory on the temperature distribution and the melt pool size.

Keywords: Non-Fourier, Enthalpy technique, Melt pool, Radiational boundary condition

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1980
1500 Flow Discharge Determination in Straight Compound Channels Using ANNs

Authors: A. Zahiri, A. A. Dehghani

Abstract:

Although many researchers have studied the flow hydraulics in compound channels, there are still many complicated problems in determination of their flow rating curves. Many different methods have been presented for these channels but extending them for all types of compound channels with different geometrical and hydraulic conditions is certainly difficult. In this study, by aid of nearly 400 laboratory and field data sets of geometry and flow rating curves from 30 different straight compound sections and using artificial neural networks (ANNs), flow discharge in compound channels was estimated. 13 dimensionless input variables including relative depth, relative roughness, relative width, aspect ratio, bed slope, main channel side slopes, flood plains side slopes and berm inclination and one output variable (flow discharge), have been used in ANNs. Comparison of ANNs model and traditional method (divided channel method-DCM) shows high accuracy of ANNs model results. The results of Sensitivity analysis showed that the relative depth with 47.6 percent contribution, is the most effective input parameter for flow discharge prediction. Relative width and relative roughness have 19.3 and 12.2 percent of importance, respectively. On the other hand, shape parameter, main channel and flood plains side slopes with 2.1, 3.8 and 3.8 percent of contribution, have the least importance.

Keywords: ANN model, compound channels, divided channel method (DCM), flow rating curve

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2558
1499 Experimental Inspection of Damage and Performance Evaluation after Repair and Strengthening of Jiamusi Highway Prestressed Concrete Bridge in China

Authors: Ali Fadhil Naser, Wang Zonglin

Abstract:

The main objectives of this study are to inspect and identify any damage of jaimusi highway prestressed concrete bridge after repair and strengthening of damaged structural members and to evaluate the performance of the bridge structural members by adopting static load test. Inspection program after repair and strengthening includes identifying and evaluating the structural members of bridge such as T-shape cantilever structure, hanging beams, corbels, external tendons, anchor beams, sticking steel plate, and piers. The results of inspection show that the overall state of the bridge structural member after repair and strengthening is good. The results of rebound test of concrete strength show that the average strength of concrete is 46.31Mpa. Whereas, the average value of concrete strength of anchor beam is 49.82Mpa. According to the results of static load test, the experimental values are less than theoretical values of internal forces, deflection, and strain, indicating that the stiffness of the experimental structure, overall deformation and integrity satisfy the designed standard and the working performance is good, and the undertaking capacity has a certain surplus. There is not visible change in the length and width of cracks and there are not new cracks under experimental load.

Keywords: Jiamusi Bridge, Damage inspection, deflection, strain.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1815
1498 Transportation Mode Choice Analysis for Accessibility of the Mehrabad International Airport by Statistical Models

Authors: N. Mirzaei Varzeghani, M. Saffarzadeh, A. Naderan, A. Taheri

Abstract:

Countries are progressing, and the world's busiest airports see year-on-year increases in travel demand. Passenger acceptability of an airport depends on the airport's appeals, which may include one of these routes between the city and the airport, as well as the facilities to reach them. One of the critical roles of transportation planners is to predict future transportation demand so that an integrated, multi-purpose system can be provided and diverse modes of transportation (rail, air, and land) can be delivered to a destination like an airport. In this study, 356 questionnaires were filled out in person over six days. First, the attraction of business and non-business trips was studied using data and a linear regression model. Lower travel costs, more passengers aged 55 and older using this airport, and other factors are essential for business trips. Non-business travelers, on the other hand, have prioritized using personal vehicles to get to the airport and ensuring convenient access to the airport. Business travelers are also less price-sensitive than non-business travelers regarding airport travel. Furthermore, carrying additional luggage (for example, more than one suitcase per person) undoubtedly decreases the attractiveness of public transit. Afterward, based on the manner and purpose of the trip, the locations with the highest trip generation to the airport were identified. The most famous district in Tehran was District 2, with 23 visits, while the most popular mode of transportation was an online taxi, with 12 trips from that location. Then, significant variables in separation and behavior of travel methods to access the airport were investigated for all systems. In this scenario, the most crucial factor is the time it takes to get to the airport, followed by the method's user-friendliness as a component of passenger preference. It has also been demonstrated that enhancing public transportation trip times reduces private transportation's market share, including taxicabs. Based on the responses of personal and semi-public vehicles, the desire of passengers to approach the airport via public transportation systems was explored to enhance present techniques and develop new strategies for providing the most efficient modes of transportation. Using the binary model, it was clear that business travelers and people who had already driven to the airport were the least likely to change.

Keywords: Multimodal transportation, travel behavior, demand modeling, statistical models.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 532
1497 Cross-Cultural Strategies for Web Design

Authors: Armando Calabrese, Guendalina Capece, Michele Corbò, Nathan Levialdi Ghiron, Matteo M. Marucchi

Abstract:

People from different cultures favor web pages characterized by the values of their culture and, therefore, tend to prefer different characteristics of a website according to their cultural values in terms of navigation, security, product information, customer service, shopping and design tools. For a company aiming to globalize its market it is useful to implement country specific cultural interfaces and different web sites for countries with different cultures. This paper, following the conclusions proposed by two models of Hall and Hofstede, and the studies of Marcus and Gould, defines, through an empirical analysis, the guidelines of web design for both the Scandinavian countries and Malaysia.

Keywords: Cultural dimensions, cultural markers, Hofstede, web design, web marketing.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3484
1496 Effect of Transverse Reinforcement on the Behavior of Tension Lap splice in High-Strength Reinforced Concrete Beams

Authors: Ahmed H. Abdel-Kareem, Hala. Abousafa, Omia S. El-Hadidi

Abstract:

The results of an experimental program conducted on seventeen simply supported concrete beams to study the effect of transverse reinforcement on the behavior of lap splice of steel reinforcement in tension zones in high strength concrete beams, are presented. The parameters included in the experimental program were the concrete compressive strength, the lap splice length, the amount of transverse reinforcement provided within the splice region, and the shape of transverse reinforcement around spliced bars. The experimental results showed that the displacement ductility increased and the mode of failure changed from splitting bond failure to flexural failure when the amount of transverse reinforcement in splice region increased, and the compressive strength increased up to 100 MPa. The presence of transverse reinforcement around spliced bars had pronounced effect on increasing the ultimate load, the ultimate deflection, and the displacement ductility. The prediction of maximum steel stresses for spliced bars using ACI 318-05 building code was compared with the experimental results. The comparison showed that the effect of transverse reinforcement around spliced bars has to be considered into the design equations for lap splice length in high strength concrete beams.

Keywords: Ductility, high strength concrete, tension lap splice, transverse reinforcement, steel stresses.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 4712
1495 Development and Validation of an Instrument Measuring the Coping Strategies in Situations of Stress

Authors: Lucie Côté, Martin Lauzier, Guy Beauchamp, France Guertin

Abstract:

Stress causes deleterious effects to the physical, psychological and organizational levels, which highlight the need to use effective coping strategies to deal with it. Several coping models exist, but they don’t integrate the different strategies in a coherent way nor do they take into account the new research on the emotional coping and acceptance of the stressful situation. To fill these gaps, an integrative model incorporating the main coping strategies was developed. This model arises from the review of the scientific literature on coping and from a qualitative study carried out among workers with low or high levels of stress, as well as from an analysis of clinical cases. The model allows one to understand under what circumstances the strategies are effective or ineffective and to learn how one might use them more wisely. It includes Specific Strategies in controllable situations (the Modification of the Situation and the Resignation-Disempowerment), Specific Strategies in non-controllable situations (Acceptance and Stubborn Relentlessness) as well as so-called General Strategies (Wellbeing and Avoidance). This study is intended to undertake and present the process of development and validation of an instrument to measure coping strategies based on this model. An initial pool of items has been generated from the conceptual definitions and three expert judges have validated the content. Of these, 18 items have been selected for a short form questionnaire. A sample of 300 students and employees from a Quebec university was used for the validation of the questionnaire. Concerning the reliability of the instrument, the indices observed following the inter-rater agreement (Krippendorff’s alpha) and the calculation of the coefficients for internal consistency (Cronbach's alpha) are satisfactory. To evaluate the construct validity, a confirmatory factor analysis using MPlus supports the existence of a model with six factors. The results of this analysis suggest also that this configuration is superior to other alternative models. The correlations show that the factors are only loosely related to each other. Overall, the analyses carried out suggest that the instrument has good psychometric qualities and demonstrates the relevance of further work to establish predictive validity and reconfirm its structure. This instrument will help researchers and clinicians better understand and assess coping strategies to cope with stress and thus prevent mental health issues.

Keywords: Acceptance, coping strategies, measurement instrument, questionnaire, stress, validation process.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 922
1494 A new Cellular Automata Model of Cardiac Action Potential Propagation based on Summation of Excited Neighbors

Authors: F. Pourhasanzade, S. H. Sabzpoushan

Abstract:

The heart tissue is an excitable media. A Cellular Automata is a type of model that can be used to model cardiac action potential propagation. One of the advantages of this approach against the methods based on differential equations is its high speed in large scale simulations. Recent cellular automata models are not able to avoid flat edges in the result patterns or have large neighborhoods. In this paper, we present a new model to eliminate flat edges by minimum number of neighbors.

Keywords: Cellular Automata, Action Potential Simulation, Isotropic Pattern.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1949
1493 An Introduction to Giulia Annalinda Neglia Viewpoint on Morphology of the Islamic City Using Written Content Analysis Approach

Authors: Mohammad Saber Eslamlou

Abstract:

Morphology of Islamic cities has been extensively studied by researchers. In this regard, there exist much difference in method of analysis, classification, recognition, confrontation and comparative method of urban morphology. The present paper aims to examine the previous methods, approaches and insights and how Dr. Giulia Annalinda Neglia dealt with the analysis of morphology of Islamic cities. Neglia is assistant professor in University of Bari, Italy (UNIBA) who has published numerous papers and books on Islamic cities. We introduce the works in the field of morphology of Islamic cities and then, her thoughts, insights and research methodologies are presented and analyzed in critical perspective. This is a qualitative research on her written works, which have been classified in three major categories. The present paper focuses mainly on her works regarding morphology and physical shape of Islamic cities. The results of her works’ review suggest that she has used Moratoria typology in investigating morphology of Islamic cities. Moreover, overall structure of the cities under investigation is often described linear; however, she is against to define a single framework for the recognition of morphology in Islamic cities. She believes that fabric of each region in the city follows from the principles of a specific period or urban pattern, in particular, Hellenistic and Roman structures. Furthermore, she believes that it is impossible to understand the morphology of a city without taking into account the obvious and hidden developments associated with it, because form of building and their surrounding open spaces are written history of the city.

Keywords: City, Islamic city, morphology of city, Giulia Annalinda Neglia.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 340
1492 Extended Least Squares LS–SVM

Authors: József Valyon, Gábor Horváth

Abstract:

Among neural models the Support Vector Machine (SVM) solutions are attracting increasing attention, mostly because they eliminate certain crucial questions involved by neural network construction. The main drawback of standard SVM is its high computational complexity, therefore recently a new technique, the Least Squares SVM (LS–SVM) has been introduced. In this paper we present an extended view of the Least Squares Support Vector Regression (LS–SVR), which enables us to develop new formulations and algorithms to this regression technique. Based on manipulating the linear equation set -which embodies all information about the regression in the learning process- some new methods are introduced to simplify the formulations, speed up the calculations and/or provide better results.

Keywords: Function estimation, Least–Squares Support VectorMachines, Regression, System Modeling

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2009
1491 The Sign in the Communication Process

Authors: S. Pesina, T. Solonchak

Abstract:

In the process of information transmission (concept verbalization) we deal mostly with the substance (contents), and then pay attention to the form. Recalling events from the remote past, often we cannot exactly reproduce specific heard or pronounced words, as well as the syntactic structures. We remember events, feelings, images; we recall the general contents of the discourse. The thought gets a specific language form only during the concept verbalization phase. With minimum time for pondering, depending on the language competence level, the grammar and syntactic shaping often occurs automatically with the use of famous models and stereotypes. This means that the language form adapts itself to the consciousness, and not vice versa.

Keywords: Lexical eidos, phenomenology, noema, polysemantic word, semantic core.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1942
1490 Quantification of Peptides based on Isotope Dilution Surface Enhanced Raman Scattering

Authors: F. Yaghobian, R. Stosch, B. Güttler

Abstract:

This study aims to demonstrate the quantification of peptides based on isotope dilution surface enhanced Raman scattering (IDSERS). SERS spectra of phenylalanine (Phe), leucine (Leu) and two peptide sequences TGQIFK (T13) and YSFLQNPQTSLCFSESIPTPSNR (T6) as part of the 22-kDa human growth hormone (hGH) were obtained on Ag-nanoparticle covered substrates. On the basis of the dominant Phe and Leu vibrational modes, precise partial least squares (PLS) prediction models were built enabling the determination of unknown T13 and T6 concentrations. Detection of hGH in its physiological concentration in order to investigate the possibility of protein quantification has been achieved.

Keywords: Surface Enhanced Raman Scattering, Quantification, Peptides.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1672
1489 A Contribution to 3D Modeling of Manufacturing Tolerance Optimization

Authors: F. Sebaa, A. Cheikh, M. Rahou

Abstract:

The study of the generated defects on manufactured parts shows the difficulty to maintain parts in their positions during the machining process and to estimate them during the pre-process plan. This work presents a contribution to the development of 3D models for the optimization of the manufacturing tolerances. An experimental study allows the measurement of the defects of part positioning for the determination of ε and the choice of an optimal setup of the part. An approach of 3D tolerance based on the small displacements method permits the determination of the manufacturing errors upstream. A developed tool, allows an automatic generation of the tolerance intervals along the three axes.

Keywords: Manufacturing tolerances, 3D modeling, optimization, errors.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1573
1488 A New Time Dependent, High Temperature Analytical Model for the Single-electron Box in Digital Applications

Authors: M.J. Sharifi

Abstract:

Several models have been introduced so far for single electron box, SEB, which all of them were restricted to DC response and or low temperature limit. In this paper we introduce a new time dependent, high temperature analytical model for SEB for the first time. DC behavior of the introduced model will be verified against SIMON software and its time behavior will be verified against a newly published paper regarding step response of SEB.

Keywords: Single electron box, SPICE, SIMON, Timedependent, Circuit model.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1236
1487 Removal of Basic Blue 3 from Aqueous Solution by Adsorption Onto Durio Ziberthinus

Authors: Siew-Teng

Abstract:

Durian husk (DH), a fruit waste, was studied for its ability to remove Basic blue 3 (BB3) from aqueous solutions. Batch kinetic studies were carried out to study the sorption characteristics under various experimental conditions. The optimum pH for the dye removal occurred in the pH range of 3-10. Sorption was found to be concentration and agitation dependent. The kinetics of dye sorption fitted a pseudo-second order rate expression. Both Langmuir and Freundlich models appeared to provide reasonable fittings for the sorption data of BB3 on durian husk. Maximum sorption capacity calculated from the Langmuir model is 49.50 mg g-1.

Keywords: Durian husk, Batch study, Sorption, Basic Blue 3

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1280
1486 Retrieval Augmented Generation against the Machine: Merging Human Cyber Security Expertise with Generative AI

Authors: Brennan Lodge

Abstract:

Amidst a complex regulatory landscape, Retrieval Augmented Generation (RAG) emerges as a transformative tool for Governance Risk and Compliance (GRC) officers. This paper details the application of RAG in synthesizing Large Language Models (LLMs) with external knowledge bases, offering GRC professionals an advanced means to adapt to rapid changes in compliance requirements. While the development for standalone LLMs is exciting, such models do have their downsides. LLMs cannot easily expand or revise their memory, and they cannot straightforwardly provide insight into their predictions, and may produce “hallucinations.” Leveraging a pre-trained seq2seq transformer and a dense vector index of domain-specific data, this approach integrates real-time data retrieval into the generative process, enabling gap analysis and the dynamic generation of compliance and risk management content. We delve into the mechanics of RAG, focusing on its dual structure that pairs parametric knowledge contained within the transformer model with non-parametric data extracted from an updatable corpus. This hybrid model enhances decision-making through context-rich insights, drawing from the most current and relevant information, thereby enabling GRC officers to maintain a proactive compliance stance. Our methodology aligns with the latest advances in neural network fine-tuning, providing a granular, token-level application of retrieved information to inform and generate compliance narratives. By employing RAG, we exhibit a scalable solution that can adapt to novel regulatory challenges and cybersecurity threats, offering GRC officers a robust, predictive tool that augments their expertise. The granular application of RAG’s dual structure not only improves compliance and risk management protocols but also informs the development of compliance narratives with pinpoint accuracy. It underscores AI’s emerging role in strategic risk mitigation and proactive policy formation, positioning GRC officers to anticipate and navigate the complexities of regulatory evolution confidently.

Keywords: Retrieval Augmented Generation, Governance Risk and Compliance, Cybersecurity, AI-driven Compliance, Risk Management, Generative AI.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 124
1485 Decomposition Method for Neural Multiclass Classification Problem

Authors: H. El Ayech, A. Trabelsi

Abstract:

In this article we are going to discuss the improvement of the multi classes- classification problem using multi layer Perceptron. The considered approach consists in breaking down the n-class problem into two-classes- subproblems. The training of each two-class subproblem is made independently; as for the phase of test, we are going to confront a vector that we want to classify to all two classes- models, the elected class will be the strongest one that won-t lose any competition with the other classes. Rates of recognition gotten with the multi class-s approach by two-class-s decomposition are clearly better that those gotten by the simple multi class-s approach.

Keywords: Artificial neural network, letter-recognition, Multi class Classification, Multi Layer Perceptron.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1572
1484 Forecasting the Istanbul Stock Exchange National 100 Index Using an Artificial Neural Network

Authors: Birol Yildiz, Abdullah Yalama, Metin Coskun

Abstract:

Many studies have shown that Artificial Neural Networks (ANN) have been widely used for forecasting financial markets, because of many financial and economic variables are nonlinear, and an ANN can model flexible linear or non-linear relationship among variables. The purpose of the study was to employ an ANN models to predict the direction of the Istanbul Stock Exchange National 100 Indices (ISE National-100). As a result of this study, the model forecast the direction of the ISE National-100 to an accuracy of 74, 51%.

Keywords: Artificial Neural Networks, Istanbul StockExchange, Non-linear Modeling.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2239
1483 Identification of an Unstable Nonlinear System: Quadrotor

Authors: Mauricio Pe˜na, Adriana Luna, Carol Rodr´ıguez

Abstract:

In the following article we begin from a multi-parameter unstable nonlinear model of a Quadrotor. We design a control to stabilize and assure the attitude of the device, starting off a linearized system at the equilibrium point of the null angles of Euler (hover), which provides us a control with limited capacities at small angles of rotation of the vehicle in three dimensions. In order to clear this obstacle, we propose the identification of models in different angles by means of simulations and the design of a controller specifically implemented for the identification task, that in future works will allow the development of controllers according to fast and agile angles of Euler for Quadrotor.

Keywords: Quadrotor, model, control, identification.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2737
1482 Calculation of Wave Function at the Origin (WFO) for Heavy Mesons by Numerical Solving of the Schrodinger Equation

Authors: M. Momeni Feyli

Abstract:

Many recent high energy physics calculations involving charm and beauty invoke wave function at the origin (WFO) for the meson bound state. Uncertainties of charm and beauty quark masses and different models for potentials governing these bound states require a simple numerical algorithm for evaluation of the WFO's for these bound states. We present a simple algorithm for this propose which provides WFO's with high precision compared with similar ones already obtained in the literature.

Keywords: Mesons, Bound states, Schrodinger equation, Nonrelativistic quark model.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1503
1481 A Strategy to Optimize the SPC Scheme for Mass Production of HDD Arm with ClusteringTechnique and Three-Way Control Chart

Authors: W. Chattinnawat

Abstract:

Consider a mass production of HDD arms where hundreds of CNC machines are used to manufacturer the HDD arms. According to an overwhelming number of machines and models of arm, construction of separate control chart for monitoring each HDD arm model by each machine is not feasible. This research proposed a strategy to optimize the SPC management on shop floor. The procedure started from identifying the clusters of the machine with similar manufacturing performance using clustering technique. The three way control chart ( I - MR - R ) is then applied to each clustered group of machine. This proposed research has advantageous to the manufacturer in terms of not only better performance of the SPC but also the quality management paradigm.

Keywords: Three way control chart. I - MR - R , between/within variation, HDD arm.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1635
1480 Probability of Globality

Authors: Eva Eggeling, Dieter W. Fellner, Torsten Ullrich

Abstract:

The objective of global optimization is to find the globally best solution of a model. Nonlinear models are ubiquitous in many applications and their solution often requires a global search approach; i.e. for a function f from a set A ⊂ Rn to the real numbers, an element x0 ∈ A is sought-after, such that ∀ x ∈ A : f(x0) ≤ f(x). Depending on the field of application, the question whether a found solution x0 is not only a local minimum but a global one is very important. This article presents a probabilistic approach to determine the probability of a solution being a global minimum. The approach is independent of the used global search method and only requires a limited, convex parameter domain A as well as a Lipschitz continuous function f whose Lipschitz constant is not needed to be known.

Keywords: global optimization, probability theory, probability of globality

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1582
1479 Matrix Completion with Heterogeneous Observation Cost Using Sparsity-Number of Column-Space

Authors: Ilqar Ramazanli

Abstract:

The matrix completion problem has been studied broadly under many underlying conditions. In many real-life scenarios, we could expect elements from distinct columns or distinct positions to have a different cost. In this paper, we explore this generalization under adaptive conditions. We approach the problem under two different cost models. The first one is that entries from different columns have different observation costs, but, within the same column, each entry has a uniform cost. The second one is any two entry has different observation cost, despite being the same or different columns. We provide complexity analysis of our algorithms and provide tightness guarantees.

Keywords: Matrix completion, adaptive learning, heterogeneous cost, Matroid optimization.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 495
1478 Sperm Identification Using Elliptic Model and Tail Detection

Authors: Vahid Reza Nafisi, Mohammad Hasan Moradi, Mohammad Hosain Nasr-Esfahani

Abstract:

The conventional assessment of human semen is a highly subjective assessment, with considerable intra- and interlaboratory variability. Computer-Assisted Sperm Analysis (CASA) systems provide a rapid and automated assessment of the sperm characteristics, together with improved standardization and quality control. However, the outcome of CASA systems is sensitive to the method of experimentation. While conventional CASA systems use digital microscopes with phase-contrast accessories, producing higher contrast images, we have used raw semen samples (no staining materials) and a regular light microscope, with a digital camera directly attached to its eyepiece, to insure cost benefits and simple assembling of the system. However, since the accurate finding of sperms in the semen image is the first step in the examination and analysis of the semen, any error in this step can affect the outcome of the analysis. This article introduces and explains an algorithm for finding sperms in low contrast images: First, an image enhancement algorithm is applied to remove extra particles from the image. Then, the foreground particles (including sperms and round cells) are segmented form the background. Finally, based on certain features and criteria, sperms are separated from other cells.

Keywords: Computer-Assisted Sperm Analysis (CASA), Sperm identification, Tail detection, Elliptic shape model.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1928
1477 Investigating the Capacity of Ultimate Torsion of Concrete Prismatic Beams with Transverse Spiral Bars

Authors: Hadi Barghlame, M. A. Lotfollahi-Yaghin

Abstract:

In this paper, the torsion capacity of ultimate point on rectangular beams with spiral reinforcements in the torsion direction and its anti-direction are investigated. Therefore, models of above-mentioned beams have been numerically analyzed under various loads using ANSYS software. It was observed that, spirallyreinforced prismatic beam and beam with spiral links, show lower torsion capacity than beam with normal links also in anti-direction. The result is that the concrete regulations are violated in this case.

Keywords: RC beam, ultimate torsion, finite element, prismatic beams, spirally tie.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1905
1476 Air Handling Units Power Consumption Using Generalized Additive Model for Anomaly Detection: A Case Study in a Singapore Campus

Authors: Ju Peng Poh, Jun Yu Charles Lee, Jonathan Chew Hoe Khoo

Abstract:

The emergence of digital twin technology, a digital replica of physical world, has improved the real-time access to data from sensors about the performance of buildings. This digital transformation has opened up many opportunities to improve the management of the building by using the data collected to help monitor consumption patterns and energy leakages. One example is the integration of predictive models for anomaly detection. In this paper, we use the GAM (Generalised Additive Model) for the anomaly detection of Air Handling Units (AHU) power consumption pattern. There is ample research work on the use of GAM for the prediction of power consumption at the office building and nation-wide level. However, there is limited illustration of its anomaly detection capabilities, prescriptive analytics case study, and its integration with the latest development of digital twin technology. In this paper, we applied the general GAM modelling framework on the historical data of the AHU power consumption and cooling load of the building between Jan 2018 to Aug 2019 from an education campus in Singapore to train prediction models that, in turn, yield predicted values and ranges. The historical data are seamlessly extracted from the digital twin for modelling purposes. We enhanced the utility of the GAM model by using it to power a real-time anomaly detection system based on the forward predicted ranges. The magnitude of deviation from the upper and lower bounds of the uncertainty intervals is used to inform and identify anomalous data points, all based on historical data, without explicit intervention from domain experts. Notwithstanding, the domain expert fits in through an optional feedback loop through which iterative data cleansing is performed. After an anomalously high or low level of power consumption detected, a set of rule-based conditions are evaluated in real-time to help determine the next course of action for the facilities manager. The performance of GAM is then compared with other approaches to evaluate its effectiveness. Lastly, we discuss the successfully deployment of this approach for the detection of anomalous power consumption pattern and illustrated with real-world use cases.

Keywords: Anomaly detection, digital twin, Generalised Additive Model, Power Consumption Model.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 501
1475 A Study of the Built Environment Design Elements Embedded into the Multiple Criteria Strategic Planning Model for an Urban Renewal

Authors: Wann-Ming Wey

Abstract:

The link between urban planning and design principles and the built environment of an urban renewal area is of interest to the field of urban studies. During the past decade, there has also been increasing interest in urban planning and design; this interest is motivated by the possibility that design policies associated with the built environment can be used to control, manage, and shape individual activity and behavior. However, direct assessments and design techniques of the links between how urban planning design policies influence individuals are still rare in the field. Recent research efforts in urban design have focused on the idea that land use and design policies can be used to increase the quality of design projects for an urban renewal area-s built environment. The development of appropriate design techniques for the built environment is an essential element of this research. Quality function deployment (QFD) is a powerful tool for improving alternative urban design and quality for urban renewal areas, and for procuring a citizen-driven quality system. In this research, we propose an integrated framework based on QFD and an Analytic Network Process (ANP) approach to determine the Alternative Technical Requirements (ATRs) to be considered in designing an urban renewal planning and design alternative. We also identify the research designs and methodologies that can be used to evaluate the performance of urban built environment projects. An application in an urban renewal built environment planning and design project evaluation is presented to illustrate the proposed framework.

Keywords: Analytic Network Process, Built Environment, Quality Function Deployment, Urban Design, Urban Renewal.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2089
1474 Economic Evaluation of Degradation by Corrosion of an on-Grid Battery Energy Storage System: A Case Study in Algeria Territory

Authors: Fouzia Brihmat

Abstract:

Economic planning models, which are used to build microgrids and Distributed Energy Resources (DER), are the current norm for expressing such confidence. These models often decide both short-term DER dispatch and long-term DER investments. This research investigates the most cost-effective hybrid (photovoltaic-diesel) renewable energy system (HRES) based on Total Net Present Cost (TNPC) in an Algerian Saharan area, which has a high potential for solar irradiation and has a production capacity of 1 GW/h. Lead-acid batteries have been around much longer and are easier to understand, but have limited storage capacity. Lithium-ion batteries last longer, are lighter, but generally more expensive. By combining the advantages of each chemistry, we produce cost-effective high-capacity battery banks that operate solely on AC coupling. The financial implications of this research describe the corrosion process that occurs at the interface between the active material and grid material of the positive plate of a lead-acid battery. The best cost study for the HRES is completed with the assistance of the HOMER Pro MATLAB Link. Additionally, during the course of the project's 20 years, the system is simulated for each time step. In this model, which takes into consideration decline in solar efficiency, changes in battery storage levels over time, and rises in fuel prices above the rate of inflation, the trade-off is that the model is more accurate, but the computation takes longer. We initially utilized the optimizer to run the model without multi-year in order to discover the best system architecture. The optimal system for the single-year scenario is the Danvest generator, which has 760 kW, 200 kWh of the necessary quantity of lead-acid storage, and a somewhat lower Cost Of Energy (COE) of $0.309/kWh. Different scenarios that account for fluctuations in the gasified biomass generator's production of electricity have been simulated, and various strategies to guarantee the balance between generation and consumption have been investigated.

Keywords: Battery, Corrosion, Diesel, Economic planning optimization, Hybrid energy system, HES, Lead-acid battery, Li-ion battery, multi-year planning, microgrid, price forecast, total net present cost, wind.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 165
1473 Modelling Extreme Temperature in Malaysia Using Generalized Extreme Value Distribution

Authors: Husna Hasan, Norfatin Salam, Mohd Bakri Adam

Abstract:

Extreme temperature of several stations in Malaysia is modelled by fitting the monthly maximum to the Generalized Extreme Value (GEV) distribution. The Mann-Kendall (MK) test suggests a non-stationary model. Two models are considered for stations with trend and the Likelihood Ratio test is used to determine the best-fitting model. Results show that half of the stations favour a model which is linear for the location parameters. The return level is the level of events (maximum temperature) which is expected to be exceeded once, on average, in a given number of years, is obtained.

Keywords: Extreme temperature, extreme value, return level.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2835