Search results for: equivalent circuit parameters
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 9854

Search results for: equivalent circuit parameters

9374 The Impact of Modeling Method of Moisture Emission from the Swimming Pool on the Accuracy of Numerical Calculations of Air Parameters in Ventilated Natatorium

Authors: Piotr Ciuman, Barbara Lipska

Abstract:

The aim of presented research was to improve numerical predictions of air parameters distribution in the actual natatorium by the selection of calculation formula of mass flux of moisture emitted from the pool. Selected correlation should ensure the best compliance of numerical results with the measurements' results of these parameters in the facility. The numerical model of the natatorium was developed, for which boundary conditions were prepared on the basis of measurements' results carried out in the actual facility. Numerical calculations were carried out with the use of ANSYS CFX software, with six formulas being implemented, which in various ways made the moisture emission dependent on water surface temperature and air parameters in the natatorium. The results of calculations with the use of these formulas were compared for air parameters' distributions: Specific humidity, velocity and temperature in the facility. For the selection of the best formula, numerical results of these parameters in occupied zone were validated by comparison with the measurements' results carried out at selected points of this zone.

Keywords: experimental validation, indoor swimming pool, moisture emission, natatorium, numerical calculations CFD, thermal and humidity conditions, ventilation

Procedia PDF Downloads 375
9373 When Conducting an Analysis of Workplace Incidents, It Is Imperative to Meticulously Calculate Both the Frequency and Severity of Injuries Sustain

Authors: Arash Yousefi

Abstract:

Experts suggest that relying exclusively on parameters to convey a situation or establish a condition may not be adequate. Assessing and appraising incidents in a system based on accident parameters, such as accident frequency, lost workdays, or fatalities, may not always be precise and occasionally erroneous. The frequency rate of accidents is a metric that assesses the correlation between the number of accidents causing work-time loss due to injuries and the total working hours of personnel over a year. Traditionally, this has been calculated based on one million working hours, but the American Occupational Safety and Health Organization has updated its standards. The new coefficient of 200/000 working hours is now used to compute the frequency rate of accidents. It's crucial to ensure that the total working hours of employees are equally represented when calculating individual event and incident numbers. The accident severity rate is a metric used to determine the amount of time lost or wasted during a given period, often a year, in relation to the total number of working hours. It measures the percentage of work hours lost or wasted compared to the total number of useful working hours, which provides valuable insight into the number of days lost or wasted due to work-related incidents for each working hour. Calculating the severity of an incident can be difficult if a worker suffers permanent disability or death. To determine lost days, coefficients specified in the "tables of days equivalent to OSHA or ANSI standards" for disabling injuries are used. The accident frequency coefficient denotes the rate at which accidents occur, while the accident severity coefficient specifies the extent of damage and injury caused by these accidents. These coefficients are crucial in accurately assessing the magnitude and impact of accidents.

Keywords: incidents, safety, analysis, frequency, severity, injuries, determine

Procedia PDF Downloads 59
9372 An Examination of Changes on Natural Vegetation due to Charcoal Production Using Multi Temporal Land SAT Data

Authors: T. Garba, Y. Y. Babanyara, M. Isah, A. K. Muktari, R. Y. Abdullahi

Abstract:

The increased in demand of fuel wood for heating, cooking and sometimes bakery has continued to exert appreciable impact on natural vegetation. This study focus on the use of multi-temporal data from land sat TM of 1986, land sat EMT of 1999 and lands sat ETM of 2006 to investigate the changes of Natural Vegetation resulting from charcoal production activities. The three images were classified based on bare soil, built up areas, cultivated land, and natural vegetation, Rock out crop and water bodies. From the classified images Land sat TM of 1986 it shows natural vegetation of the study area to be 308,941.48 hectares equivalent to 50% of the area it then reduces to 278,061.21 which is 42.92% in 1999 it again depreciated to 199,647.81 in 2006 equivalent to 30.83% of the area. Consequently cultivated continue increasing from 259,346.80 hectares (42%) in 1986 to 312,966.27 hectares (48.3%) in 1999 and then to 341.719.92 hectares (52.78%). These show that within the span of 20 years (1986 to 2006) the natural vegetation is depreciated by 119,293.81 hectares. This implies that if the menace is not control the natural might likely be lost in another twenty years. This is because forest cleared for charcoal production is normally converted to farmland. The study therefore concluded that there is the need for alternatives source of domestic energy such as the use of biomass which can easily be accessible and affordable to people. In addition, the study recommended that there should be strong policies enforcement for the protection forest reserved.

Keywords: charcoal, classification, data, images, land use, natural vegetation

Procedia PDF Downloads 346
9371 Optimization of Loudspeaker Part Design Parameters by Air Viscosity Damping Effect

Authors: Yue Hu, Xilu Zhao, Takao Yamaguchi, Manabu Sasajima, Yoshio Koike, Akira Hara

Abstract:

This study optimized the design parameters of a cone loudspeaker as an example of high flexibility of the product design. We developed an acoustic analysis software program that considers the impact of damping caused by air viscosity. In sound reproduction, it is difficult to optimize each parameter of the loudspeaker design. To overcome the limitation of the design problem in practice, this study presents an acoustic analysis algorithm to optimize the design parameters of the loudspeaker. The material character of cone paper and the loudspeaker edge were the design parameters, and the vibration displacement of the cone paper was the objective function. The results of the analysis showed that the design had high accuracy as compared to the predicted value. These results suggested that although the parameter design is difficult, with experience and intuition, the design can be performed easily using the optimized design found with the acoustic analysis software.

Keywords: air viscosity, design parameters, loudspeaker, optimization

Procedia PDF Downloads 488
9370 Risk Assessment of Radiation Hazard for a Typical WWER1000: Cancer Risk Analysis during a Hypothetical Accident

Authors: R. Gharari, N. Kojouri, R. Hosseini Aghdam, E. Alibeigi, B. Salmasian

Abstract:

In this research, the WWER1000/V446 (a PWR Russian type reactor) is chosen as the case study. It is assumed that radioactive materials that release into the environment are more than allowable limit due to a complete failure of the ventilation system (reactor stack). In the following, the HOTSPOT and the RASCAL computational codes have been used and coupled with a developed program using MATLAB software to evaluate Total effective dose equivalent (TEDE) and cancer risk according to the BEIR equations for various human organs. In addition, effects of the containment spray system and climate conditions on the TEDE have been investigated. According to the obtained results, there is an inverse correlation between the received dose and the wind speed; the amount of the TEDE for wind speed 2 m/s and is more than wind speed for 14 m/s during the class A of the climate (2.168 and 0.444 mSv, respectively). Also, containment spray system can effect and reduce the amount of the fission products and TEDE. Furthermore, the probability of the cancer risk for women is more than men, and for children is more than adults. In addition, a specific emergency zonal planning is proposed. Results are promising in which the site selection of the WWER1000/V446 were considered safe for the public in this situation.

Keywords: TEDE, total effective dose equivalent, RASCAL and HOTSPOT codes, BEIR equations, cancer risk

Procedia PDF Downloads 143
9369 Experimental Investigation on the Effects of Electroless Nickel Phosphorus Deposition, pH and Temperature with the Varying Coating Bath Parameters on Impact Energy by Taguchi Method

Authors: D. Kari Basavaraja, M. G. Skanda, C. Soumya, V. Ramesh

Abstract:

This paper discusses the effects of sodium hypophosphite concentration, pH, and temperature on deposition rate. This paper also discusses the evaluation of coating strength, surface, and subsurface by varying the bath parameters, percentage of phosphate, plating temperature, and pH of the plating solution. Taguchi technique has been used for the analysis. In the experiment, nickel chloride which is a source of nickel when mixed with sodium hypophosphite has been used as the reducing agent and the source of phosphate and sodium hydroxide has been used to vary the pH of the coating bath. The coated samples are tested for impact energy by conducting impact test. Finally, the effects of coating bath parameters on the impact energy absorbed have been plotted, and analysis has been carried out. Further, percentage contribution of coating bath parameters using Design of Experiments approach (DOE) has been analysed. Finally, it can be concluded that the bath parameters of the Ni-P coating will certainly influence on the strength of the specimen.

Keywords: bath parameters, coatings, design of experiment, fracture toughness, impact strength

Procedia PDF Downloads 331
9368 Phytochemical Study and Biological Activity of Sage (Salvia officinalis L.)

Authors: Mekhaldi Abdelkader, Bouzned Ahcen, Djibaoui Rachid, Hamoum Hakim

Abstract:

This study presents an attempt to evaluate the antioxidant and antimicrobial activity of methanolic extract and essential oils prepared from the leaves of sage (Salvia officinalis L.). The content of polyphenols in the methanolic extract of the leaves from Salvia officinalis extract was determined by spectrophoto- metrically, calculated as gallic acid and catechin equivalent. Antioxidant activity was evaluated by free radical scavenging activity using 2,2-diphenylpicryl-1-picrylhydrazyl (DPPH) assay. The plant essential oil and methanol extract were also subjected to screenings for the evaluation of their antioxidant activities using 2, 2-diphenyl-1-picrylhydrazyl (DPPH) test. While the plant essential oil showed only weak antioxidant activities, its methanol extract was considerably active in DPPH (IC50= 37.29µg/ml) test. Appreciable total phenolic content (31.25mg/g) was also detected for the plant methanol extract as gallic acid equivalent in the Folin–Ciocalteu test. The plant was also screened for its antimicrobial activity and good to moderate inhibitions were recorded for its essential oil and methanol extract against most of the tested microorganisms. The present investigation revealed that this plant has rich source of antioxidant properties. It is for this reason that sage has found increasing application in food formulations.

Keywords: antibacterial activity, antioxidant activity, flavonoid, polyphenol, salvia officinalis

Procedia PDF Downloads 387
9367 Measuring the Height of a Person in Closed Circuit Television Video Footage Using 3D Human Body Model

Authors: Dojoon Jung, Kiwoong Moon, Joong Lee

Abstract:

The height of criminals is one of the important clues that can determine the scope of the suspect's search or exclude the suspect from the search target. Although measuring the height of criminals by video alone is limited by various reasons, the 3D data of the scene and the Closed Circuit Television (CCTV) footage are matched, the height of the criminal can be measured. However, it is still difficult to measure the height of CCTV footage in the non-contact type measurement method because of variables such as position, posture, and head shape of criminals. In this paper, we propose a method of matching the CCTV footage with the 3D data on the crime scene and measuring the height of the person using the 3D human body model in the matched data. In the proposed method, the height is measured by using 3D human model in various scenes of the person in the CCTV footage, and the measurement value of the target person is corrected by the measurement error of the replay CCTV footage of the reference person. We tested for 20 people's walking CCTV footage captured from an indoor and an outdoor and corrected the measurement values with 5 reference persons. Experimental results show that the measurement error (true value-measured value) average is 0.45 cm, and this method is effective for the measurement of the person's height in CCTV footage.

Keywords: human height, CCTV footage, 2D/3D matching, 3D human body model

Procedia PDF Downloads 226
9366 Study of Landslide Behavior with Topographic Monitoring and Numerical Modeling

Authors: ZerarkaHizia, Akchiche Mustapha, Prunier Florent

Abstract:

Landslide of Ain El Hammam (AEH) has been an old slip since 1969; it was reactivated after an intense rainfall period in 2008 where it presents a complex shape and affects broad areas. The schist of AEH is more or less altered; the alteration is facilitated by the fracturing of the rock in its upper part, the presence of flowing water as well as physical and chemical mechanisms of desegregation in joint of altered schist. The factors following these instabilities are mostly related to the geological formation, the hydro-climatic conditions and the topography of the region. The city of AEH is located on the top of a steep slope at 50 km from the city of TiziOuzou (Algeria). AEH’s topographic monitoring of unstable slope allows analyzing the structure and the different deformation mechanism and the gradual change in the geometry, the direction of change of slip. It also allows us to delimit the area affected by the movement. This work aims to study the behavior of AEH landslide with topographic monitoring and to validate the results with numerical modeling of the slip site, when the hydraulic factors are identified as the most important factors for the reactivation of this landslide. With the help of the numerical code PLAXIS 2D and PlaxFlow, the precipitations and the steady state flow are modeled. To identify the mechanism of deformation and to predict the spread of the AEH landslide numerically, we used the equivalent deviatory strain, and these results were visualized by MATLAB software.

Keywords: equivalent deviatory strain, landslide, numerical modeling, topographic monitoring

Procedia PDF Downloads 268
9365 Development of a French to Yorùbá Machine Translation System

Authors: Benjamen Nathaniel, Eludiora Safiriyu Ijiyemi, Egume Oneme Lucky

Abstract:

A review on machine translation systems shows that a lot of computational artefacts has been carried out to translate written or spoken texts from a source language to Yorùbá language through Machine Translation systems. However, there are no work on French to Yorùbá language machine translation system; hence, the study investigated the process involved in the translation of French-to-Yorùbá language equivalent with the view to adopting a rule- based MT approach to build a Machine Translation framework from simple sentences administered through questionnaire. Articles and relevant textbooks were reviewed with key speakers of both languages interviewed to find out the processes involved in the translation of French language and their equivalent in Yorùbálanguage simple sentences using home domain terminologies. Achieving this, a model was formulated using phrase grammar structure, re-write rule, parse tree, automata theory- based techniques, designed and implemented respectively with unified modeling language (UML) and python programming language. Analysing the result, it was observed when carrying out the result that, the Machine Translation system performed 18.45% above Experimental Subject Respondent and 2.7% below Linguistics Expert when analysed with word orthography, sentence syntax and semantic correctness of the sentences. And, when compared with Google Machine Translation system, it was noticed that the developed system performed better on lexicons of the target language.

Keywords: machine translation (MT), rule-based, French language, Yoru`ba´ language

Procedia PDF Downloads 30
9364 Impact of Machining Parameters on the Surface Roughness of Machined PU Block

Authors: Louis Denis Kevin Catherine, Raja Aziz Raja Ma’arof, Azrina Arshad, Sangeeth Suresh

Abstract:

Machining parameters are very important in determining the surface quality of any material. In the past decade, some new engineering materials were developed for the manufacturing industry which created a need to conduct an investigation on the impact of the said parameters on their surface roughness. The polyurethane (PU) block is widely used in the automotive industry to manufacture parts such as checking fixtures that are used to verify the dimensional accuracy of automotive parts. In this paper, the design of experiment (DOE) was used to investigate the effect of the milling parameters on the PU block. Furthermore, an analysis of the machined surface chemical composition was done using scanning electron microscope (SEM). It was found that the surface roughness of the PU block is severely affected when PU undergoes a flood machining process instead of a dry condition. In addition, the step over and the silicon content were found to be the most significant parameters that influence the surface quality of the PU block.

Keywords: polyurethane (PU), design of experiment (DOE), scanning electron microscope (SEM), surface roughness

Procedia PDF Downloads 493
9363 The Influence of Microscopic Features on the Self-Cleaning Ability of Developed 3D Printed Fabric-Like Structures Using Different Printing Parameters

Authors: Ayat Adnan Atwah, Muhammad A. Khan

Abstract:

Self-cleaning surfaces are getting significant attention in industrial fields. Especially for textile fabrics, it is observed that self-cleaning textile fabric surfaces are created by manipulating the surface features with the help of coatings and nanoparticles, which are considered costly and far more complicated. However, controlling the fabrication parameters of textile fabrics at the microscopic level by exploring the potential for self-cleaning has not been addressed. This study aimed to establish the context of self-cleaning textile fabrics by controlling the fabrication parameters of the textile fabric at the microscopic level. Therefore, 3D-printed textile fabrics were fabricated using the low-cost fused filament fabrication (FFF) technique. The printing parameters, such as orientation angle (O), layer height (LH), and extruder width (EW), were used to control the microscopic features of the printed fabrics. The combination of three printing parameters was created to provide the best self-cleaning textile fabric surface: (LH) (0.15, 0.13, 0.10 mm) and (EW) (0.5, 0.4, 0.3 mm) along with two different (O) of (45º and 90º). Three different thermoplastic flexible filament materials were used: (TPU 98A), (TPE felaflex), and (TPC flex45). The printing parameters were optimised to get the optimum self-cleaning ability of the printed specimens. Furthermore, the impact of these characteristics on mechanical strength at the fabric-woven structure level was investigated. The study revealed that the printing parameters significantly affect the self-cleaning properties after adjusting the selected combination of layer height, extruder width, and printing orientation. A linear regression model was effectively developed to demonstrate the association between 3D printing parameters (layer height, extruder width, and orientation). According to the experimental results, (TPE felaflex) has a better self-cleaning ability than the other two materials.

Keywords: 3D printing, self-cleaning fabric, microscopic features, printing parameters, fabrication

Procedia PDF Downloads 56
9362 Transformers in Gene Expression-Based Classification

Authors: Babak Forouraghi

Abstract:

A genetic circuit is a collection of interacting genes and proteins that enable individual cells to implement and perform vital biological functions such as cell division, growth, death, and signaling. In cell engineering, synthetic gene circuits are engineered networks of genes specifically designed to implement functionalities that are not evolved by nature. These engineered networks enable scientists to tackle complex problems such as engineering cells to produce therapeutics within the patient's body, altering T cells to target cancer-related antigens for treatment, improving antibody production using engineered cells, tissue engineering, and production of genetically modified plants and livestock. Construction of computational models to realize genetic circuits is an especially challenging task since it requires the discovery of flow of genetic information in complex biological systems. Building synthetic biological models is also a time-consuming process with relatively low prediction accuracy for highly complex genetic circuits. The primary goal of this study was to investigate the utility of a pre-trained bidirectional encoder transformer that can accurately predict gene expressions in genetic circuit designs. The main reason behind using transformers is their innate ability (attention mechanism) to take account of the semantic context present in long DNA chains that are heavily dependent on spatial representation of their constituent genes. Previous approaches to gene circuit design, such as CNN and RNN architectures, are unable to capture semantic dependencies in long contexts as required in most real-world applications of synthetic biology. For instance, RNN models (LSTM, GRU), although able to learn long-term dependencies, greatly suffer from vanishing gradient and low-efficiency problem when they sequentially process past states and compresses contextual information into a bottleneck with long input sequences. In other words, these architectures are not equipped with the necessary attention mechanisms to follow a long chain of genes with thousands of tokens. To address the above-mentioned limitations of previous approaches, a transformer model was built in this work as a variation to the existing DNA Bidirectional Encoder Representations from Transformers (DNABERT) model. It is shown that the proposed transformer is capable of capturing contextual information from long input sequences with attention mechanism. In a previous work on genetic circuit design, the traditional approaches to classification and regression, such as Random Forrest, Support Vector Machine, and Artificial Neural Networks, were able to achieve reasonably high R2 accuracy levels of 0.95 to 0.97. However, the transformer model utilized in this work with its attention-based mechanism, was able to achieve a perfect accuracy level of 100%. Further, it is demonstrated that the efficiency of the transformer-based gene expression classifier is not dependent on presence of large amounts of training examples, which may be difficult to compile in many real-world gene circuit designs.

Keywords: transformers, generative ai, gene expression design, classification

Procedia PDF Downloads 32
9361 Enhancing Transfer Path Analysis with In-Situ Component Transfer Path Analysis for Interface Forces Identification

Authors: Raef Cherif, Houssine Bakkali, Wafaa El Khatiri, Yacine Yaddaden

Abstract:

The analysis of how vibrations are transmitted between components is required in many engineering applications. Transfer path analysis (TPA) has been a valuable engineering tool for solving Noise, Vibration, and Harshness (NVH problems using sub-structuring applications. The most challenging part of a TPA analysis is estimating the equivalent forces at the contact points between the active and the passive side. Component TPA in situ Method calculates these forces by inverting the frequency response functions (FRFs) measured at the passive subsystem, relating the motion at indicator points to forces at the interface. However, matrix inversion could pose problems due to the ill-conditioning of the matrices leading to inaccurate results. This paper establishes a TPA model for an academic system consisting of two plates linked by four springs. A numerical study has been performed to improve the interface forces identification. Several parameters are studied and discussed, such as the singular value rejection and the number and position of indicator points chosen and used in the inversion matrix.

Keywords: transfer path analysis, matrix inverse method, indicator points, SVD decomposition

Procedia PDF Downloads 59
9360 Optical Imaging Based Detection of Solder Paste in Printed Circuit Board Jet-Printing Inspection

Authors: D. Heinemann, S. Schramm, S. Knabner, D. Baumgarten

Abstract:

Purpose: Applying solder paste to printed circuit boards (PCB) with stencils has been the method of choice over the past years. A new method uses a jet printer to deposit tiny droplets of solder paste through an ejector mechanism onto the board. This allows for more flexible PCB layouts with smaller components. Due to the viscosity of the solder paste, air blisters can be trapped in the cartridge. This can lead to missing solder joints or deviations in the applied solder volume. Therefore, a built-in and real-time inspection of the printing process is needed to minimize uncertainties and increase the efficiency of the process by immediate correction. The objective of the current study is the design of an optimal imaging system and the development of an automatic algorithm for the detection of applied solder joints from optical from the captured images. Methods: In a first approach, a camera module connected to a microcomputer and LED strips are employed to capture images of the printed circuit board under four different illuminations (white, red, green and blue). Subsequently, an improved system including a ring light, an objective lens, and a monochromatic camera was set up to acquire higher quality images. The obtained images can be divided into three main components: the PCB itself (i.e., the background), the reflections induced by unsoldered positions or screw holes and the solder joints. Non-uniform illumination is corrected by estimating the background using a morphological opening and subtraction from the input image. Image sharpening is applied in order to prevent error pixels in the subsequent segmentation. The intensity thresholds which divide the main components are obtained from the multimodal histogram using three probability density functions. Determining the intersections delivers proper thresholds for the segmentation. Remaining edge gradients produces small error areas which are removed by another morphological opening. For quantitative analysis of the segmentation results, the dice coefficient is used. Results: The obtained PCB images show a significant gradient in all RGB channels, resulting from ambient light. Using different lightings and color channels 12 images of a single PCB are available. A visual inspection and the investigation of 27 specific points show the best differentiation between those points using a red lighting and a green color channel. Estimating two thresholds from analyzing the multimodal histogram of the corrected images and using them for segmentation precisely extracts the solder joints. The comparison of the results to manually segmented images yield high sensitivity and specificity values. Analyzing the overall result delivers a Dice coefficient of 0.89 which varies for single object segmentations between 0.96 for a good segmented solder joints and 0.25 for single negative outliers. Conclusion: Our results demonstrate that the presented optical imaging system and the developed algorithm can robustly detect solder joints on printed circuit boards. Future work will comprise a modified lighting system which allows for more precise segmentation results using structure analysis.

Keywords: printed circuit board jet-printing, inspection, segmentation, solder paste detection

Procedia PDF Downloads 310
9359 Influence of Solenoid Configuration on Electromagnetic Acceleration of Plunger

Authors: Shreyansh Bharadwaj, Raghavendra Kollipara, Sijoy C. D., R. K. Mittal

Abstract:

Utilizing the Lorentz force to propel an electrically conductive plunger through a solenoid represents a fundamental application in electromagnetism. The parameters of the solenoid significantly influence the force exerted on the plunger, impacting its response. A parametric study has been done to understand the effect of these parameters on the force acting on the plunger. This study is done to determine the most optimal combination of parameters to obtain the fast response. Analysis has been carried out using an algorithm capable of simulating the scenario of a plunger undergoing acceleration within a solenoid. Authors have conducted an analysis focusing on several key configuration parameters of the solenoid. These parameters include the inter-layer gap (in the case of a multi-turn solenoid), different conductor diameters, varying numbers of turns, and diverse numbers of layers. Primary objective of this paper is to discern how alterations in these parameters affect the force applied to the plunger. Through extensive numerical simulations, a dataset has been generated and utilized to construct informative plots. These plots provide visual representations of the relationships between the solenoid configuration parameters and the resulting force exerted on the plunger, which can further be used to deduce scaling laws. This research endeavors to offer valuable insights into optimizing solenoid configurations for enhanced electromagnetic acceleration, thereby contributing to advancements in electromagnetic propulsion technology.

Keywords: Lorentz force, solenoid configuration, electromagnetic acceleration, parametric analysis, simulation

Procedia PDF Downloads 19
9358 Constructing the Joint Mean-Variance Regions for Univariate and Bivariate Normal Distributions: Approach Based on the Measure of Cumulative Distribution Functions

Authors: Valerii Dashuk

Abstract:

The usage of the confidence intervals in economics and econometrics is widespread. To be able to investigate a random variable more thoroughly, joint tests are applied. One of such examples is joint mean-variance test. A new approach for testing such hypotheses and constructing confidence sets is introduced. Exploring both the value of the random variable and its deviation with the help of this technique allows checking simultaneously the shift and the probability of that shift (i.e., portfolio risks). Another application is based on the normal distribution, which is fully defined by mean and variance, therefore could be tested using the introduced approach. This method is based on the difference of probability density functions. The starting point is two sets of normal distribution parameters that should be compared (whether they may be considered as identical with given significance level). Then the absolute difference in probabilities at each 'point' of the domain of these distributions is calculated. This measure is transformed to a function of cumulative distribution functions and compared to the critical values. Critical values table was designed from the simulations. The approach was compared with the other techniques for the univariate case. It differs qualitatively and quantitatively in easiness of implementation, computation speed, accuracy of the critical region (theoretical vs. real significance level). Stable results when working with outliers and non-normal distributions, as well as scaling possibilities, are also strong sides of the method. The main advantage of this approach is the possibility to extend it to infinite-dimension case, which was not possible in the most of the previous works. At the moment expansion to 2-dimensional state is done and it allows to test jointly up to 5 parameters. Therefore the derived technique is equivalent to classic tests in standard situations but gives more efficient alternatives in nonstandard problems and on big amounts of data.

Keywords: confidence set, cumulative distribution function, hypotheses testing, normal distribution, probability density function

Procedia PDF Downloads 148
9357 Human 3D Metastatic Melanoma Models for in vitro Evaluation of Targeted Therapy Efficiency

Authors: Delphine Morales, Florian Lombart, Agathe Truchot, Pauline Maire, Pascale Vigneron, Antoine Galmiche, Catherine Lok, Muriel Vayssade

Abstract:

Targeted therapy molecules are used as a first-line treatment for metastatic melanoma with B-Raf mutation. Nevertheless, these molecules can cause side effects to patients and are efficient on 50 to 60 % of them. Indeed, melanoma cell sensitivity to targeted therapy molecules is dependent on tumor microenvironment (cell-cell and cell-extracellular matrix interactions). To better unravel factors modulating cell sensitivity to B-Raf inhibitor, we have developed and compared several melanoma models: from metastatic melanoma cells cultured as monolayer (2D) to a co-culture in a 3D dermal equivalent. Cell response was studied in different melanoma cell lines such as SK-MEL-28 (mutant B-Raf (V600E), sensitive to Vemurafenib), SK-MEL-3 (mutant B-Raf (V600E), resistant to Vemurafenib) and a primary culture of dermal human fibroblasts (HDFn). Assays have initially been performed in a monolayer cell culture (2D), then a second time on a 3D dermal equivalent (dermal human fibroblasts embedded in a collagen gel). All cell lines were treated with Vemurafenib (a B-Raf inhibitor) for 48 hours at various concentrations. Cell sensitivity to treatment was assessed under various aspects: Cell proliferation (cell counting, EdU incorporation, MTS assay), MAPK signaling pathway analysis (Western-Blotting), Apoptosis (TUNEL), Cytokine release (IL-6, IL-1α, HGF, TGF-β, TNF-α) upon Vemurafenib treatment (ELISA) and histology for 3D models. In 2D configuration, the inhibitory effect of Vemurafenib on cell proliferation was confirmed on SK-MEL-28 cells (IC50=0.5 µM), and not on the SK-MEL-3 cell line. No apoptotic signal was detected in SK-MEL-28-treated cells, suggesting a cytostatic effect of the Vemurafenib rather than a cytotoxic one. The inhibition of SK-MEL-28 cell proliferation upon treatment was correlated with a strong expression decrease of phosphorylated proteins involved in the MAPK pathway (ERK, MEK, and AKT/PKB). Vemurafenib (from 5 µM to 10 µM) also slowed down HDFn proliferation, whatever cell culture configuration (monolayer or 3D dermal equivalent). SK-MEL-28 cells cultured in the dermal equivalent were still sensitive to high Vemurafenib concentrations. To better characterize all cell population impacts (melanoma cells, dermal fibroblasts) on Vemurafenib efficacy, cytokine release is being studied in 2D and 3D models. We have successfully developed and validated a relevant 3D model, mimicking cutaneous metastatic melanoma and tumor microenvironment. This 3D melanoma model will become more complex by adding a third cell population, keratinocytes, allowing us to characterize the epidermis influence on the melanoma cell sensitivity to Vemurafenib. In the long run, the establishment of more relevant 3D melanoma models with patients’ cells might be useful for personalized therapy development. The authors would like to thank the Picardie region and the European Regional Development Fund (ERDF) 2014/2020 for the funding of this work and Oise committee of "La ligue contre le cancer".

Keywords: 3D human skin model, melanoma, tissue engineering, vemurafenib efficiency

Procedia PDF Downloads 282
9356 Pushing the Boundary of Parallel Tractability for Ontology Materialization via Boolean Circuits

Authors: Zhangquan Zhou, Guilin Qi

Abstract:

Materialization is an important reasoning service for applications built on the Web Ontology Language (OWL). To make materialization efficient in practice, current research focuses on deciding tractability of an ontology language and designing parallel reasoning algorithms. However, some well-known large-scale ontologies, such as YAGO, have been shown to have good performance for parallel reasoning, but they are expressed in ontology languages that are not parallelly tractable, i.e., the reasoning is inherently sequential in the worst case. This motivates us to study the problem of parallel tractability of ontology materialization from a theoretical perspective. That is we aim to identify the ontologies for which materialization is parallelly tractable, i.e., in the NC complexity. Since the NC complexity is defined based on Boolean circuit that is widely used to investigate parallel computing problems, we first transform the problem of materialization to evaluation of Boolean circuits, and then study the problem of parallel tractability based on circuits. In this work, we focus on datalog rewritable ontology languages. We use Boolean circuits to identify two classes of datalog rewritable ontologies (called parallelly tractable classes) such that materialization over them is parallelly tractable. We further investigate the parallel tractability of materialization of a datalog rewritable OWL fragment DHL (Description Horn Logic). Based on the above results, we analyze real-world datasets and show that many ontologies expressed in DHL belong to the parallelly tractable classes.

Keywords: ontology materialization, parallel reasoning, datalog, Boolean circuit

Procedia PDF Downloads 245
9355 Investigation of Wind Farm Interaction with Ethiopian Electric Power’s Grid: A Case Study at Ashegoda Wind Farm

Authors: Fikremariam Beyene, Getachew Bekele

Abstract:

Ethiopia is currently on the move with various projects to raise the amount of power generated in the country. The progress observed in recent years indicates this fact clearly and indisputably. The rural electrification program, the modernization of the power transmission system, the development of wind farm is some of the main accomplishments worth mentioning. As it is well known, currently, wind power is globally embraced as one of the most important sources of energy mainly for its environmentally friendly characteristics, and also that once it is installed, it is a source available free of charge. However, integration of wind power plant with an existing network has many challenges that need to be given serious attention. In Ethiopia, a number of wind farms are either installed or are under construction. A series of wind farm is planned to be installed in the near future. Ashegoda Wind farm (13.2°, 39.6°), which is the subject of this study, is the first large scale wind farm under construction with the capacity of 120 MW. The first phase of 120 MW (30 MW) has been completed and is expected to be connected to the grid soon. This paper is concerned with the investigation of the wind farm interaction with the national grid under transient operating condition. The main concern is the fault ride through (FRT) capability of the system when the grid voltage drops to exceedingly low values because of short circuit fault and also the active and reactive power behavior of wind turbines after the fault is cleared. On the wind turbine side, a detailed dynamic modelling of variable speed wind turbine of a 1 MW capacity running with a squirrel cage induction generator and full-scale power electronics converters is done and analyzed using simulation software DIgSILENT PowerFactory. On the Ethiopian electric power corporation side, after having collected sufficient data for the analysis, the grid network is modeled. In the model, a fault ride-through (FRT) capability of the plant is studied by applying 3-phase short circuit on the grid terminal near the wind farm. The results show that the Ashegoda wind farm can ride from voltage deep within a short time and the active and reactive power performance of the wind farm is also promising.

Keywords: squirrel cage induction generator, active and reactive power, DIgSILENT PowerFactory, fault ride-through capability, 3-phase short circuit

Procedia PDF Downloads 143
9354 A Parametric Study on Effects of Internal Factors on Carbonation of Reinforced Concrete

Authors: Kunal Tongaria, Abhishek Mangal, S. Mandal, Devendra Mohan

Abstract:

The carbonation of concrete is a phenomenon which is a function of various interdependent parameters. Therefore, in spite of numerous literature and database, the useful generalization is not an easy task. These interdependent parameters can be grouped under the category of internal and external factors. This paper focuses on the internal parameters which govern and increase the probability of the ingress of deleterious substances into concrete. The mechanism of effects of internal parameters such as microstructure for with and without supplementary cementing materials (SCM), water/binder ratio, the age of concrete etc. has been discussed. This is followed by the comparison of various proposed mathematical models for the deterioration of concrete. Based on existing laboratory experiments as well as field results, this paper concludes the present understanding of mechanism, modeling and future research needs in this field.

Keywords: carbonation, diffusion coefficient, microstructure of concrete, reinforced concrete

Procedia PDF Downloads 381
9353 Wear Measurement of Thermomechanical Parameters of the Metal Carbide

Authors: Riad Harouz, Brahim Mahfoud

Abstract:

The threads and the circles on reinforced concrete are obtained by process of hot rolling with pebbles finishers in metal carbide which present a way of rolling around the outside diameter. Our observation is that this throat presents geometrical wear after the end of its cycle determined in tonnage. In our study, we have determined, in a first step, experimentally measurements of the wear in terms of thermo-mechanical parameters (Speed, Load, and Temperature) and the influence of these parameters on the wear. In the second stage, we have developed a mathematical model of lifetime useful for the prognostic of the wear and their changes.

Keywords: lifetime, metal carbides, modeling, thermo-mechanical, wear

Procedia PDF Downloads 281
9352 An Amended Method for Assessment of Hypertrophic Scars Viscoelastic Parameters

Authors: Iveta Bryjova

Abstract:

Recording of viscoelastic strain-vs-time curves with the aid of the suction method and a follow-up analysis, resulting into evaluation of standard viscoelastic parameters, is a significant technique for non-invasive contact diagnostics of mechanical properties of skin and assessment of its conditions, particularly in acute burns, hypertrophic scarring (the most common complication of burn trauma) and reconstructive surgery. For elimination of the skin thickness contribution, usable viscoelastic parameters deduced from the strain-vs-time curves are restricted to the relative ones (i.e. those expressed as a ratio of two dimensional parameters), like grosselasticity, net-elasticity, biological elasticity or Qu’s area parameters, in literature and practice conventionally referred to as R2, R5, R6, R7, Q1, Q2, and Q3. With the exception of parameters R2 and Q1, the remaining ones substantially depend on the position of inflection point separating the elastic linear and viscoelastic segments of the strain-vs-time curve. The standard algorithm implemented in commercially available devices relies heavily on the experimental fact that the inflection time comes about 0.1 sec after the suction switch-on/off, which depreciates credibility of parameters thus obtained. Although the Qu’s US 7,556,605 patent suggests a method of improving the precision of the inflection determination, there is still room for nonnegligible improving. In this contribution, a novel method of inflection point determination utilizing the advantageous properties of the Savitzky–Golay filtering is presented. The method allows computation of derivatives of smoothed strain-vs-time curve, more exact location of inflection and consequently more reliable values of aforementioned viscoelastic parameters. An improved applicability of the five inflection-dependent relative viscoelastic parameters is demonstrated by recasting a former study under the new method, and by comparing its results with those provided by the methods that have been used so far.

Keywords: Savitzky–Golay filter, scarring, skin, viscoelasticity

Procedia PDF Downloads 275
9351 Effect of Silt Presence on Shear Strength Parameters of Unsaturated Sandy Soils

Authors: R. Ziaie Moayed, E. Khavaninzadeh, M. Ghorbani Tochaee

Abstract:

Direct shear test is widely used in soil mechanics experiment to determine the shear strength parameters of granular soils. For analysis of soil stability problems such as bearing capacity, slope stability and lateral pressure on soil retaining structures, the shear strength parameters must be known well. In the present study, shear strength parameters are determined in silty-sand mixtures. Direct shear tests are performed on 161 Firoozkooh sand with different silt content at a relative density of 70% in three vertical stress of 100, 150, and 200 kPa. Wet tamping method is used for soil sample preparation, and the results include diagrams of shear stress versus shear deformation and sample height changes against shear deformation. Accordingly, in different silt percent, the shear strength parameters of the soil such as internal friction angle and dilation angle are calculated and compared. According to the results, when the sample contains up to 10% silt, peak shear strength and internal friction angle have an upward trend. However, if the sample contains 10% to 50% of silt a downward trend is seen in peak shear strength and internal friction angle.

Keywords: shear strength parameters, direct shear test, silty sand, shear stress, shear deformation

Procedia PDF Downloads 137
9350 Modelling and Optimization Analysis of Silicon/MgZnO-CBTSSe Tandem Solar Cells

Authors: Vallisree Sivathanu, Kumaraswamidhas Lakshmi Annamalai, Trupti Ranjan Lenka

Abstract:

We report a tandem solar cell model with Silicon as the bottom cell absorber material and Cu₂BaSn(S, Se)₄(CBTSSe) as absorber material for the top cell. As a first step, the top and bottom cells were modelled and validated by comparison with the experiment. Once the individual cells are validated, then the tandem structure is modelled with Indium Tin Oxide(ITO) as conducting layer between the top and bottom cells. The tandem structure yielded better open circuit voltage and fill factor; however, the efficiency obtained is 7.01%. The top cell and the bottom cells are investigated with the help of electron-hole current density, photogeneration rate, and external quantum efficiency profiles. In order to minimize the various loss mechanisms in the tandem solar cell, the material parameters are optimized within experimentally achievable limits. Initially, the top cell optimization was carried out; then, the bottom cell is optimized for maximizing the light absorption, and upon minimizing the current and photon losses in the tandem structure, the maximum achievable efficiency is predicted to be 19.52%.

Keywords: CBTSSe, silicon, tandem, solar cell, device modeling, current losses, photon losses

Procedia PDF Downloads 84
9349 Practical Challenges of Tunable Parameters in Matlab/Simulink Code Generation

Authors: Ebrahim Shayesteh, Nikolaos Styliaras, Alin George Raducu, Ozan Sahin, Daniel Pombo VáZquez, Jonas Funkquist, Sotirios Thanopoulos

Abstract:

One of the important requirements in many code generation projects is defining some of the model parameters tunable. This helps to update the model parameters without performing the code generation again. This paper studies the concept of embedded code generation by MATLAB/Simulink coder targeting the TwinCAT Simulink system. The generated runtime modules are then tested and deployed to the TwinCAT 3 engineering environment. However, defining the parameters tunable in MATLAB/Simulink code generation targeting TwinCAT is not very straightforward. This paper focuses on this subject and reviews some of the techniques tested here to make the parameters tunable in generated runtime modules. Three techniques are proposed for this purpose, including normal tunable parameters, callback functions, and mask subsystems. Moreover, some test Simulink models are developed and used to evaluate the results of proposed approaches. A brief summary of the study results is presented in the following. First of all, the parameters defined tunable and used in defining the values of other Simulink elements (e.g., gain value of a gain block) could be changed after the code generation and this value updating will affect the values of all elements defined based on the values of the tunable parameter. For instance, if parameter K=1 is defined as a tunable parameter in the code generation process and this parameter is used to gain a gain block in Simulink, the gain value for the gain block is equal to 1 in the gain block TwinCAT environment after the code generation. But, the value of K can be changed to a new value (e.g., K=2) in TwinCAT (without doing any new code generation in MATLAB). Then, the gain value of the gain block will change to 2. Secondly, adding a callback function in the form of “pre-load function,” “post-load function,” “start function,” and will not help to make the parameters tunable without performing a new code generation. This means that any MATLAB files should be run before performing the code generation. The parameters defined/calculated in this file will be used as fixed values in the generated code. Thus, adding these files as callback functions to the Simulink model will not make these parameters flexible since the MATLAB files will not be attached to the generated code. Therefore, to change the parameters defined/calculated in these files, the code generation should be done again. However, adding these files as callback functions forces MATLAB to run them before the code generation, and there is no need to define the parameters mentioned in these files separately. Finally, using a tunable parameter in defining/calculating the values of other parameters through the mask is an efficient method to change the value of the latter parameters after the code generation. For instance, if tunable parameter K is used in calculating the value of two other parameters K1 and K2 and, after the code generation, the value of K is updated in TwinCAT environment, the value of parameters K1 and K2 will also be updated (without any new code generation).

Keywords: code generation, MATLAB, tunable parameters, TwinCAT

Procedia PDF Downloads 203
9348 Parametric Influence and Optimization of Wire-EDM on Oil Hardened Non-Shrinking Steel

Authors: Nixon Kuruvila, H. V. Ravindra

Abstract:

Wire-cut Electro Discharge Machining (WEDM) is a special form of conventional EDM process in which electrode is a continuously moving conductive wire. The present study aims at determining parametric influence and optimum process parameters of Wire-EDM using Taguchi’s Technique and Genetic algorithm. The variation of the performance parameters with machining parameters was mathematically modeled by Regression analysis method. The objective functions are Dimensional Accuracy (DA) and Material Removal Rate (MRR). Experiments were designed as per Taguchi’s L16 Orthogonal Array (OA) where in Pulse-on duration, Pulse-off duration, Current, Bed-speed and Flushing rate have been considered as the important input parameters. The matrix experiments were conducted for the material Oil Hardened Non Shrinking Steel (OHNS) having the thickness of 40 mm. The results of the study reveals that among the machining parameters it is preferable to go in for lower pulse-off duration for achieving over all good performance. Regarding MRR, OHNS is to be eroded with medium pulse-off duration and higher flush rate. Finally, the validation exercise performed with the optimum levels of the process parameters. The results confirm the efficiency of the approach employed for optimization of process parameters in this study.

Keywords: dimensional accuracy (DA), regression analysis (RA), Taguchi method (TM), volumetric material removal rate (VMRR)

Procedia PDF Downloads 387
9347 Microfluidic Device for Real-Time Electrical Impedance Measurements of Biological Cells

Authors: Anil Koklu, Amin Mansoorifar, Ali Beskok

Abstract:

Dielectric spectroscopy (DS) is a noninvasive, label free technique for a long term real-time measurements of the impedance spectra of biological cells. DS enables characterization of cellular dielectric properties such as membrane capacitance and cytoplasmic conductivity. We have developed a lab-on-a-chip device that uses an electro-activated microwells array for loading, DS measurements, and unloading of biological cells. We utilized from dielectrophoresis (DEP) to capture target cells inside the wells and release them after DS measurement. DEP is a label-free technique that exploits differences among dielectric properties of the particles. In detail, DEP is the motion of polarizable particles suspended in an ionic solution and subjected to a spatially non-uniform external electric field. To the best of our knowledge, this is the first microfluidic chip that combines DEP and DS to analyze biological cells using electro-activated wells. Device performance is tested using two different cell lines of prostate cancer cells (RV122, PC-3). Impedance measurements were conducted at 0.2 V in the 10 kHz to 40 MHz range with 6 s time resolution. An equivalent circuit model was developed to extract the cell membrane capacitance and cell cytoplasmic conductivity from the impedance spectra. We report the time course of the variations in dielectric properties of PC-3 and RV122 cells suspended in low conductivity medium (LCB), which enhances dielectrophoretic and impedance responses, and their response to sudden pH change from a pH of 7.3 to a pH of 5.8. It is shown that microfluidic chip allowed online measurements of dielectric properties of prostate cancer cells and the assessment of the cellular level variations under external stimuli such as different buffer conductivity and pH. Based on these data, we intend to deploy the current device for single cell measurements by fabricating separately addressable N × N electrode platforms. Such a device will allow time-dependent dielectric response measurements for individual cells with the ability of selectively releasing them using negative-DEP and pressure driven flow.

Keywords: microfluidic, microfabrication, lab on a chip, AC electrokinetics, dielectric spectroscopy

Procedia PDF Downloads 126
9346 Genetic Algorithm Based Deep Learning Parameters Tuning for Robot Object Recognition and Grasping

Authors: Delowar Hossain, Genci Capi

Abstract:

This paper concerns with the problem of deep learning parameters tuning using a genetic algorithm (GA) in order to improve the performance of deep learning (DL) method. We present a GA based DL method for robot object recognition and grasping. GA is used to optimize the DL parameters in learning procedure in term of the fitness function that is good enough. After finishing the evolution process, we receive the optimal number of DL parameters. To evaluate the performance of our method, we consider the object recognition and robot grasping tasks. Experimental results show that our method is efficient for robot object recognition and grasping.

Keywords: deep learning, genetic algorithm, object recognition, robot grasping

Procedia PDF Downloads 328
9345 Investigation on Machine Tools Energy Consumptions

Authors: Shiva Abdoli, Daniel T.Semere

Abstract:

Several researches have been conducted to study consumption of energy in cutting process. Most of these researches are focusing to measure the consumption and propose consumption reduction methods. In this work, the relation between the cutting parameters and the consumption is investigated in order to establish a generalized energy consumption model that can be used for process and production planning in real production lines. Using the generalized model, the process planning will be carried out by taking into account the energy as a function of the selected process parameters. Similarly, the generalized model can be used in production planning to select the right operational parameters like batch sizes, routing, buffer size, etc. in a production line. The description and derivation of the model as well as a case study are given in this paper to illustrate the applicability and validity of the model.

Keywords: process parameters, cutting process, energy efficiency, Material Removal Rate (MRR)

Procedia PDF Downloads 475