Search results for: p-hub median problem
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 7550

Search results for: p-hub median problem

6350 A Study of Evolutional Control Systems

Authors: Ti-Jun Xiao, Zhe Xu

Abstract:

Controllability is one of the fundamental issues in control systems. In this paper, we study the controllability of second order evolutional control systems in Hilbert spaces with memory and boundary controls, which model dynamic behaviors of some viscoelastic materials. Transferring the control problem into a moment problem and showing the Riesz property of a family of functions related to Cauchy problems for some integrodifferential equations, we obtain a general boundary controllability theorem for these second order evolutional control systems. This controllability theorem is applicable to various concrete 1D viscoelastic systems and recovers some previous related results. It is worth noting that Riesz sequences can be used for numerical computations of the control functions and the identification of new Riesz sequence is of independent interest for the basis-function theory. Moreover, using the Riesz sequences, we obtain the existence and uniqueness of (weak) solutions to these second order evolutional control systems in Hilbert spaces. Finally, we derive the exact boundary controllability of a viscoelastic beam equation, as an application of our abstract theorem.

Keywords: evolutional control system, controllability, boundary control, existence and uniqueness

Procedia PDF Downloads 203
6349 Comparison of Classical Computer Vision vs. Convolutional Neural Networks Approaches for Weed Mapping in Aerial Images

Authors: Paulo Cesar Pereira Junior, Alexandre Monteiro, Rafael da Luz Ribeiro, Antonio Carlos Sobieranski, Aldo von Wangenheim

Abstract:

In this paper, we present a comparison between convolutional neural networks and classical computer vision approaches, for the specific precision agriculture problem of weed mapping on sugarcane fields aerial images. A systematic literature review was conducted to find which computer vision methods are being used on this specific problem. The most cited methods were implemented, as well as four models of convolutional neural networks. All implemented approaches were tested using the same dataset, and their results were quantitatively and qualitatively analyzed. The obtained results were compared to a human expert made ground truth for validation. The results indicate that the convolutional neural networks present better precision and generalize better than the classical models.

Keywords: convolutional neural networks, deep learning, digital image processing, precision agriculture, semantic segmentation, unmanned aerial vehicles

Procedia PDF Downloads 226
6348 Plotting of an Ideal Logic versus Resource Outflow Graph through Response Analysis on a Strategic Management Case Study Based Questionnaire

Authors: Vinay A. Sharma, Shiva Prasad H. C.

Abstract:

The initial stages of any project are often observed to be in a mixed set of conditions. Setting up the project is a tough task, but taking the initial decisions is rather not complex, as some of the critical factors are yet to be introduced into the scenario. These simple initial decisions potentially shape the timeline and subsequent events that might later be plotted on it. Proceeding towards the solution for a problem is the primary objective in the initial stages. The optimization in the solutions can come later, and hence, the resources deployed towards attaining the solution are higher than what they would have been in the optimized versions. A ‘logic’ that counters the problem is essentially the core of the desired solution. Thus, if the problem is solved, the deployment of resources has led to the required logic being attained. As the project proceeds along, the individuals working on the project face fresh challenges as a team and are better accustomed to their surroundings. The developed, optimized solutions are then considered for implementation, as the individuals are now experienced, and know better of the consequences and causes of possible failure, and thus integrate the adequate tolerances wherever required. Furthermore, as the team graduates in terms of strength, acquires prodigious knowledge, and begins its efficient transfer, the individuals in charge of the project along with the managers focus more on the optimized solutions rather than the traditional ones to minimize the required resources. Hence, as time progresses, the authorities prioritize attainment of the required logic, at a lower amount of dedicated resources. For empirical analysis of the stated theory, leaders and key figures in organizations are surveyed for their ideas on appropriate logic required for tackling a problem. Key-pointers spotted in successfully implemented solutions are noted from the analysis of the responses and a metric for measuring logic is developed. A graph is plotted with the quantifiable logic on the Y-axis, and the dedicated resources for the solutions to various problems on the X-axis. The dedicated resources are plotted over time, and hence the X-axis is also a measure of time. In the initial stages of the project, the graph is rather linear, as the required logic will be attained, but the consumed resources are also high. With time, the authorities begin focusing on optimized solutions, since the logic attained through them is higher, but the resources deployed are comparatively lower. Hence, the difference between consecutive plotted ‘resources’ reduces and as a result, the slope of the graph gradually increases. On an overview, the graph takes a parabolic shape (beginning on the origin), as with each resource investment, ideally, the difference keeps on decreasing, and the logic attained through the solution keeps increasing. Even if the resource investment is higher, the managers and authorities, ideally make sure that the investment is being made on a proportionally high logic for a larger problem, that is, ideally the slope of the graph increases with the plotting of each point.

Keywords: decision-making, leadership, logic, strategic management

Procedia PDF Downloads 91
6347 Visualization of Energy Waves via Airy Functions in Time-Domain

Authors: E. Sener, O. Isik, E. Eroglu, U. Sahin

Abstract:

The main idea is to solve the system of Maxwell’s equations in accordance with the causality principle to get the energy quantities via Airy functions in a hollow rectangular waveguide. We used the evolutionary approach to electromagnetics that is an analytical time-domain method. The boundary-value problem for the system of Maxwell’s equations is reformulated in transverse and longitudinal coordinates. A self-adjoint operator is obtained and the complete set of Eigen vectors of the operator initiates an orthonormal basis of the solution space. Hence, the sought electromagnetic field can be presented in terms of this basis. Within the presentation, the scalar coefficients are governed by Klein-Gordon equation. Ultimately, in this study, time-domain waveguide problem is solved analytically in accordance with the causality principle. Moreover, the graphical results are visualized for the case when the energy and surplus of the energy for the time-domain waveguide modes are represented via airy functions.

Keywords: airy functions, Klein-Gordon Equation, Maxwell’s equations, Surplus of energy, wave boundary operators

Procedia PDF Downloads 345
6346 Serum Granulocyte Colony Stimulating Factor is a Potent Stimulator of Hematopoeitic Progenitor Cells Mobilization in Trauma Hemorrhagic Shock

Authors: Manoj Kumar, Sujata Mohanty, D. N. Rao, Arul Selvi, Sanjeev K. Bhoi

Abstract:

Background: Hematopoietic progenitor cells (HPC) mobilized from bone marrow to peripheral blood has been observed in severe trauma and hemorrhagic shock patients. Granulocyte-colony stimulating factor (G-CSF) is a potent stimulator that mobilized HPC from bone marrow to peripheral blood. Objective: Our aim of the study was to investigate the serum G-CSF levels and correlate with HPC and outcome. Methods: Peripheral blood sample from 50 hemorrhagic shock patients was collected on arrival for determination of G-CSF and peripheral blood HPC (PBHPC) and compared with healthy control (n=15). Determination of serum levels of G-CSF by sandwich ELISA and PBHPC by Sysmex XE-2100. Data were categorized by age, sex, Injury Severity Score (ISS), and laboratory data was prospectively collected. Data are expressed as mean±SD and median (min, max). Results: Significantly increased the serum level of G-CSF (264.8 vs. 79.1 pg/ml) and peripheral blood HPC (0.1 vs. 0.01 %) in the T/HS patients when compared with control group. Conclusions: Our studies suggest serum G-CSF elevated in T/HS patients. The elevated in G-CSF was also associated with mobilization of HPC from BM to peripheral blood HPC. Increased the levels of G-CSF in T/HS may play a significant role in the alteration of the hematopoietic compartment.

Keywords: granulocyte colony stimulating factor, G-CSF, hematopoietic progenitor cells, HPC, trauma hemorrhagic shock, T/HS, outcome

Procedia PDF Downloads 312
6345 The Relevance of the Generalist Judge’s Discretionary Limits in the Institutional Debate

Authors: Antonio Sepúlveda, Camila Marques, Carlos Bolonha, Igor De Lazari, Henrique Rangel

Abstract:

The judicial practice faces a tension between normative discretion and institutional capacities. There are clarity graduations of the statutory text that might induce different specialization levels of the judges. A major problem stemming from that tension is a greater discretion without a proportional specialization. The normative clarity, although its absence can be overcome through specialization, avoids problems related to disproportionate discretion and judicial dissonance. When judicial interpretation deals with the lack of legal clarity, a significant juridical insecurity frame is verified. Decisional uniformity mechanisms are created in order to surpass these problems. Brazil brings great examples, such as the súmulas, the enunciados, and the súmulas vinculantes. Despite of the resistance presented to the latter, mainly based on judges’ independence, even countries of the Common Law tradition develop such mechanisms. The British Guidelines face the lack of legal clarity problem and promote a decisional consonance system.

Keywords: generalist judges, institutional capacities, normative clarity, normative discretion

Procedia PDF Downloads 452
6344 Polypropylene Fibres Dyeable with Acid Dyes

Authors: H. M. Wang, C. J. Chang

Abstract:

As the threat of global climate change is more seriously, "net zero emissions by 2050" has become a common global goal. In order to reduce the consumption of petrochemical raw materials and reduce carbon emissions, low-carbon fiber materials have become key materials in the future global textile supply chain. This project uses polyolefin raw materials to modify through synthesis and amination to develop low-temperature dyeable polypropylene fibers, endow them with low-temperature dyeability and high color fastness that can be combined with acid dyes, and improve the problem of low coloring strength. The color fastness to washing can reach the requirement of commerce with 3.5 level or more. Therefore, we realize the entry of polypropylene fiber into the clothing textile supply chain, replace existing fiber raw materials, solve the problem of domestic chemical fiber, textile, and clothing industry's plight of no low-carbon alternative new material sources, and provide the textile industry with a solution to achieve the goal of net zero emissions in 2050.

Keywords: acid dyes, dyeing, low-temperature, polypropylene fiber

Procedia PDF Downloads 72
6343 Estimation of Synchronous Machine Synchronizing and Damping Torque Coefficients

Authors: Khaled M. EL-Naggar

Abstract:

Synchronizing and damping torque coefficients of a synchronous machine can give a quite clear picture for machine behavior during transients. These coefficients are used as a power system transient stability measurement. In this paper, a crow search optimization algorithm is presented and implemented to study the power system stability during transients. The algorithm makes use of the machine responses to perform the stability study in time domain. The problem is formulated as a dynamic estimation problem. An objective function that minimizes the error square in the estimated coefficients is designed. The method is tested using practical system with different study cases. Results are reported and a thorough discussion is presented. The study illustrates that the proposed method can estimate the stability coefficients for the critical stable cases where other methods may fail. The tests proved that the proposed tool is an accurate and reliable tool for estimating the machine coefficients for assessment of power system stability.

Keywords: optimization, estimation, synchronous, machine, crow search

Procedia PDF Downloads 119
6342 Optimal Placement of Phasor Measurement Units (PMU) Using Mixed Integer Programming (MIP) for Complete Observability in Power System Network

Authors: Harshith Gowda K. S, Tejaskumar N, Shubhanga R. B, Gowtham N, Deekshith Gowda H. S

Abstract:

Phasor measurement units (PMU) are playing an important role in the current power system for state estimation. It is necessary to have complete observability of the power system while minimizing the cost. For this purpose, the optimal location of the phasor measurement units in the power system is essential. In a bus system, zero injection buses need to be evaluated to minimize the number of PMUs. In this paper, the optimization problem is formulated using mixed integer programming to obtain the optimal location of the PMUs with increased observability. The formulation consists of with and without zero injection bus as constraints. The formulated problem is simulated using a CPLEX solver in the GAMS software package. The proposed method is tested on IEEE 30, IEEE 39, IEEE 57, and IEEE 118 bus systems. The results obtained show that the number of PMUs required is minimal with increased observability.

Keywords: PMU, observability, mixed integer programming (MIP), zero injection buses (ZIB)

Procedia PDF Downloads 150
6341 Towards a Quantification of the Wind Erosion of the Gharb Shoreline Soils in Morocco by the Application of a Mathematical Model

Authors: Mohammed Kachtali, Imad Fenjiro, Jamal Alkarkouri

Abstract:

Wind erosion is a serious environmental problem in arid and semi-arid regions. Indeed, wind erosion easily removes the finest particles of the soil surface, which also contribute to losing soil fertility. The siltation of infrastructures and cultivated areas and the negative impact on health are additional consequences of wind erosion. In Morocco, wind erosion constitutes the main factor of silting up in coast and Sahara. The aim of our study is to use an equation of wind erosion in order to estimate the soil loses by wind erosion in the coast of Gharb (North of Morocco). The used equation in our model includes the geographic data, climatic data of 30 years and edaphic data collected from area study which contained 11 crossing of 4 stations. Our results have shown that the values of wind erosion are higher and very different between some crossings (p < 0.001). This difference is explained by topography, soil texture, and climate. In conclusion, wind erosion is higher in Gharb coast and varies from station to another; this problem required several methods of control and mitigation.

Keywords: Gharb coast, modeling, silting, wind erosion

Procedia PDF Downloads 120
6340 Objective Evaluation on Medical Image Compression Using Wavelet Transformation

Authors: Amhimmid Mohammed Saffour, Mustafa Mohamed Abdullah

Abstract:

The use of computers for handling image data in the healthcare is growing. However, the amount of data produced by modern image generating techniques is vast. This data might be a problem from a storage point of view or when the data is sent over a network. This paper using wavelet transform technique for medical images compression. MATLAB program, are designed to evaluate medical images storage and transmission time problem at Sebha Medical Center Libya. In this paper, three different Computed Tomography images which are abdomen, brain and chest have been selected and compressed using wavelet transform. Objective evaluation has been performed to measure the quality of the compressed images. For this evaluation, the results show that the Peak Signal to Noise Ratio (PSNR) which indicates the quality of the compressed image is ranging from (25.89db to 34.35db for abdomen images, 23.26db to 33.3db for brain images and 25.5db to 36.11db for chest images. These values shows that the compression ratio is nearly to 30:1 is acceptable.

Keywords: medical image, Matlab, image compression, wavelet's, objective evaluation

Procedia PDF Downloads 270
6339 Effect of Correlation of Random Variables on Structural Reliability Index

Authors: Agnieszka Dudzik

Abstract:

The problem of correlation between random variables in the structural reliability analysis has been extensively discussed in literature on the subject. The cases taken under consideration were usually related to correlation between random variables from one side of ultimate limit state: correlation between particular loads applied on structure or correlation between resistance of particular members of a structure as a system. It has been proved that positive correlation between these random variables reduces the reliability of structure and increases the probability of failure. In the paper, the problem of correlation between random variables from both side of the limit state equation will be taken under consideration. The simplest case where these random variables are of the normal distributions will be concerned. The case when a degree of that correlation is described by the covariance or the coefficient of correlation will be used. Special attention will be paid on questions: how much that correlation changes the reliability level and can it be ignored. In reliability analysis will be used well-known methods for assessment of the failure probability: based on the Hasofer-Lind reliability index and Monte Carlo method adapted to the problem of correlation. The main purpose of this work will be a presentation how correlation of random variables influence on reliability index of steel bar structures. Structural design parameters will be defined as deterministic values and random variables. The latter will be correlated. The criterion of structural failure will be expressed by limit functions related to the ultimate and serviceability limit state. In the description of random variables will be used only for the normal distribution. Sensitivity of reliability index to the random variables will be defined. If the reliability index sensitivity due to the random variable X will be low when compared with other variables, it can be stated that the impact of this variable on failure probability is small. Therefore, in successive computations, it can be treated as a deterministic parameter. Sensitivity analysis leads to simplify the description of the mathematical model, determine the new limit functions and values of the Hasofer-Lind reliability index. In the examples, the NUMPRESS software will be used in the reliability analysis.

Keywords: correlation of random variables, reliability index, sensitivity of reliability index, steel structure

Procedia PDF Downloads 218
6338 Triangular Hesitant Fuzzy TOPSIS Approach in Investment Projects Management

Authors: Irina Khutsishvili

Abstract:

The presented study develops a decision support methodology for multi-criteria group decision-making problem. The proposed methodology is based on the TOPSIS (Technique for Order Performance by Similarity to Ideal Solution) approach in the hesitant fuzzy environment. The main idea of decision-making problem is a selection of one best alternative or several ranking alternatives among a set of feasible alternatives. Typically, the process of decision-making is based on an evaluation of certain criteria. In many MCDM problems (such as medical diagnosis, project management, business and financial management, etc.), the process of decision-making involves experts' assessments. These assessments frequently are expressed in fuzzy numbers, confidence intervals, intuitionistic fuzzy values, hesitant fuzzy elements and so on. However, a more realistic approach is using linguistic expert assessments (linguistic variables). In the proposed methodology both the values and weights of the criteria take the form of linguistic variables, given by all decision makers. Then, these assessments are expressed in triangular fuzzy numbers. Consequently, proposed approach is based on triangular hesitant fuzzy TOPSIS decision-making model. Following the TOPSIS algorithm, first, the fuzzy positive ideal solution (FPIS) and the fuzzy negative-ideal solution (FNIS) are defined. Then the ranking of alternatives is performed in accordance with the proximity of their distances to the both FPIS and FNIS. Based on proposed approach the software package has been developed, which was used to rank investment projects in the real investment decision-making problem. The application and testing of the software were carried out based on the data provided by the ‘Bank of Georgia’.

Keywords: fuzzy TOPSIS approach, investment project, linguistic variable, multi-criteria decision making, triangular hesitant fuzzy set

Procedia PDF Downloads 405
6337 Vibration and Parametric Instability Analysis of Delaminated Composite Beams

Authors: A. Szekrényes

Abstract:

This paper revisits the free vibration problem of delaminated composite beams. It is shown that during the vibration of composite beams the delaminated parts are subjected to the parametric excitation. This can lead to the dynamic buckling during the motion of the structure. The equation of motion includes time-dependent stiffness and so it leads to a system of Mathieu-Hill differential equations. The free vibration analysis of beams is carried out in the usual way by using beam finite elements. The dynamic buckling problem is investigated locally, and the critical buckling forces are determined by the modified harmonic balance method by using an imposed time function of the motion. The stability diagrams are created, and the numerical predictions are compared to experimental results. The most important findings are the critical amplitudes at which delamination buckling takes place, the stability diagrams representing the instability of the system, and the realistic mode shape prediction in contrast with the unrealistic results of models available in the literature.

Keywords: delamination, free vibration, parametric excitation, sweep excitation

Procedia PDF Downloads 326
6336 An Approach to the Assembly Line Balancing Problem with Uncertain Operation Time

Authors: Zhongmin Wang, Lin Wei, Hengshan Zhang, Tianhua Chen, Yimin Zhou

Abstract:

The assembly line balancing problems are signficant in mass production systems. In order to deal with the uncertainties that practically exist but barely mentioned in the literature, this paper develops a mathematic model with an optimisation algorithm to solve the assembly line balancing problem with uncertainty operation time. The developed model is able to work with a variable number of workstations under the uncertain environment, aiming to obtain the minimal number of workstation and minimal idle time for each workstation. In particular, the proposed approach first introduces the concept of protection time that closely works with the uncertain operation time. Four dominance rules and the mechanism of determining up and low bounds are subsequently put forward, which serve as the basis for the proposed branch and bound algorithm. Experimental results show that the proposed work verified on a benchmark data set is able to solve the uncertainties efficiently.

Keywords: assembly lines, SALBP-UOT, uncertain operation time, branch and bound algorithm.

Procedia PDF Downloads 148
6335 Solving Dimensionality Problem and Finding Statistical Constructs on Latent Regression Models: A Novel Methodology with Real Data Application

Authors: Sergio Paez Moncaleano, Alvaro Mauricio Montenegro

Abstract:

This paper presents a novel statistical methodology for measuring and founding constructs in Latent Regression Analysis. This approach uses the qualities of Factor Analysis in binary data with interpretations on Item Response Theory (IRT). In addition, based on the fundamentals of submodel theory and with a convergence of many ideas of IRT, we propose an algorithm not just to solve the dimensionality problem (nowadays an open discussion) but a new research field that promises more fear and realistic qualifications for examiners and a revolution on IRT and educational research. In the end, the methodology is applied to a set of real data set presenting impressive results for the coherence, speed and precision. Acknowledgments: This research was financed by Colciencias through the project: 'Multidimensional Item Response Theory Models for Practical Application in Large Test Designed to Measure Multiple Constructs' and both authors belong to SICS Research Group from Universidad Nacional de Colombia.

Keywords: item response theory, dimensionality, submodel theory, factorial analysis

Procedia PDF Downloads 348
6334 Annoyance Caused by Air Pollution: A Comparative Study of Two Industrialized Regions

Authors: Milena M. Melo, Jane M. Santos, Severine Frere, Valderio A. Reisen, Neyval C. Reis Jr., Mariade Fátima S. Leite

Abstract:

Although there had been a many studies that shows the impact of air pollution on physical health, comparatively less was known of human behavioral responses and annoyance impacts. Annoyance caused by air pollution is a public health problem because it can be an ambient stressor causing stress and disease and can affect quality of life. The objective of this work is to evaluate the annoyance caused by air pollution in two different industrialized urban areas, Dunkirk (France) and Vitoria (Brazil). The populations of these cities often report feeling annoyed by dust. Surveys were conducted, and the collected data were analyzed using statistical analyses. The results show that sociodemographic variables, importance of air quality, perceived industrial risk, perceived air pollution and occurrence of health problems play important roles in the perceived annoyance. These results show the existence of a common problem in geographically distant areas and allow stakeholders to develop prevention strategies.

Keywords: air pollution, annoyance, industrial risks, public health, perception of pollution, settled dust

Procedia PDF Downloads 668
6333 Automated Method Time Measurement System for Redesigning Dynamic Facility Layout

Authors: Salam Alzubaidi, G. Fantoni, F. Failli, M. Frosolini

Abstract:

The dynamic facility layout problem is a really critical issue in the competitive industrial market; thus, solving this problem requires robust design and effective simulation systems. The sustainable simulation requires inputting reliable and accurate data into the system. So this paper describes an automated system integrated into the real environment to measure the duration of the material handling operations, collect the data in real-time, and determine the variances between the actual and estimated time schedule of the operations in order to update the simulation software and redesign the facility layout periodically. The automated method- time measurement system collects the real data through using Radio Frequency-Identification (RFID) and Internet of Things (IoT) technologies. Hence, attaching RFID- antenna reader and RFID tags enables the system to identify the location of the objects and gathering the time data. The real duration gathered will be manipulated by calculating the moving average duration of the material handling operations, choosing the shortest material handling path, and then updating the simulation software to redesign the facility layout accommodating with the shortest/real operation schedule. The periodic simulation in real-time is more sustainable and reliable than the simulation system relying on an analysis of historical data. The case study of this methodology is in cooperation with a workshop team for producing mechanical parts. Although there are some technical limitations, this methodology is promising, and it can be significantly useful in the redesigning of the manufacturing layout.

Keywords: dynamic facility layout problem, internet of things, method time measurement, radio frequency identification, simulation

Procedia PDF Downloads 109
6332 Virtual Academy Next: Addressing Transition Challenges Through a Gamified Virtual Transition Program for Students with Disabilities

Authors: Jennifer Gallup, Joel Bocanegra, Greg Callan, Abigail Vaughn

Abstract:

Students with disabilities (SWD) engaged in a distance summer program delivered over multiple virtual mediums that used gaming principles to teach and practice self-regulated learning (SRL) through the process of exploring possible jobs. Gaming quests were developed to explore jobs and teach transition skills. Students completed specially designed quests that taught and reinforced SRL and problem-solving through individual, group, and teacher-led experiences. SRL skills learned were reinforced through guided job explorations over the context of MinecraftEDU, zoom with experts in the career, collaborations with a team over Marco Polo, and Zoom. The quests were developed and laid out on an accessible web page, with active learning opportunities and feedback conducted within multiple virtual mediums including MinecraftEDU. Gaming mediums actively engage players in role-playing, problem-solving, critical thinking, and collaboration. Gaming has been used as a medium for education since the inception of formal education. Games, and specifically board games, are pre-historic, meaning we had board games before we had written language. Today, games are widely used in education, often as a reinforcer for behavior or for rewards for work completion. Games are not often used as a direct method of instruction and assessment; however, the inclusion of games as an assessment tool and as a form of instruction increases student engagement and participation. Games naturally include collaboration, problem-solving, and communication. Therefore, our summer program was developed using gaming principles and MinecraftEDU. This manuscript describes a virtual learning summer program called Virtual Academy New and Exciting Transitions (VAN) that was redesigned from a face-to-face setting to a completely online setting with a focus on SWD aged 14-21. The focus of VAN was to address transition planning needs such as problem-solving skills, self-regulation, interviewing, job exploration, and communication for transition-aged youth diagnosed with various disabilities (e.g., learning disabilities, attention-deficit hyperactivity disorder, intellectual disability, down syndrome, autism spectrum disorder).

Keywords: autism, disabilities, transition, summer program, gaming, simulations

Procedia PDF Downloads 57
6331 A Study on Computational Fluid Dynamics (CFD)-Based Design Optimization Techniques Using Multi-Objective Evolutionary Algorithms (MOEA)

Authors: Ahmed E. Hodaib, Mohamed A. Hashem

Abstract:

In engineering applications, a design has to be as fully perfect as possible in some defined case. The designer has to overcome many challenges in order to reach the optimal solution to a specific problem. This process is called optimization. Generally, there is always a function called “objective function” that is required to be maximized or minimized by choosing input parameters called “degrees of freedom” within an allowed domain called “search space” and computing the values of the objective function for these input values. It becomes more complex when we have more than one objective for our design. As an example for Multi-Objective Optimization Problem (MOP): A structural design that aims to minimize weight and maximize strength. In such case, the Pareto Optimal Frontier (POF) is used, which is a curve plotting two objective functions for the best cases. At this point, a designer should make a decision to choose the point on the curve. Engineers use algorithms or iterative methods for optimization. In this paper, we will discuss the Evolutionary Algorithms (EA) which are widely used with Multi-objective Optimization Problems due to their robustness, simplicity, suitability to be coupled and to be parallelized. Evolutionary algorithms are developed to guarantee the convergence to an optimal solution. An EA uses mechanisms inspired by Darwinian evolution principles. Technically, they belong to the family of trial and error problem solvers and can be considered global optimization methods with a stochastic optimization character. The optimization is initialized by picking random solutions from the search space and then the solution progresses towards the optimal point by using operators such as Selection, Combination, Cross-over and/or Mutation. These operators are applied to the old solutions “parents” so that new sets of design variables called “children” appear. The process is repeated until the optimal solution to the problem is reached. Reliable and robust computational fluid dynamics solvers are nowadays commonly utilized in the design and analyses of various engineering systems, such as aircraft, turbo-machinery, and auto-motives. Coupling of Computational Fluid Dynamics “CFD” and Multi-Objective Evolutionary Algorithms “MOEA” has become substantial in aerospace engineering applications, such as in aerodynamic shape optimization and advanced turbo-machinery design.

Keywords: mathematical optimization, multi-objective evolutionary algorithms "MOEA", computational fluid dynamics "CFD", aerodynamic shape optimization

Procedia PDF Downloads 240
6330 Compassion Fade: Effects of Mass Perception and Intertemporal Choice on Non-Volunteering Behavior

Authors: Mariel L. Alonzo, Patricia Mae T. Chi, Juliana Patrice P. Mayormita, Sanjana A. Sorio

Abstract:

Compassion fade proposes an inverse relationship between the magnitude of stimuli to elicited compassion. This phenomenon is viewed within a framework that integrates a 3-Act Compassion structure with Latané and Darley’s Unresponsive Bystander Model and Prospect Theory of Decision-making under risk. Students (N=211) from Ateneo de Davao were sampled to examine the effects of mass perception (increasing number of needy persons) and intertemporal choice (soon versus later) on volunteering behavior. Collegiate classes in their natural setting were randomly assigned to five different treatment groups and were presented with audiovisual presentations featuring an increasing number of needy persons. The students were deceived to believe that two hypothetical feeding programs for Marawi refugees, taking place in 1 month and 6 months, were in need of volunteers for its preparatory phase. Results show a statistically significant (p=0.000; p=0.013) non-linear trend consistently for both feeding programs. There was a decrease in volunteered time means as identifiable victims increased from 0-47 and an increase as it progressed towards 267 non-identifiable victims. Highest interest was expressed for the 0 needy people shown and least for 47. The 0 hours volunteered was consistently the mode and median in all treatments. There was no statistically significant temporal discounting effect.

Keywords: compassion, group perception, identifiable victim, intertemporal choice, prosocial behavior, unresponsive bystander

Procedia PDF Downloads 192
6329 A Systematic Review Examining the Experimental methodology behind in vivo testing of hiatus hernia and Diaphragmatic Hernia Mesh

Authors: Whitehead-Clarke T., Beynon V., Banks J., Karanjia R., Mudera V., Windsor A., Kureshi A.

Abstract:

Introduction: Mesh implants are regularly used to help repair both hiatus hernias (HH) and diaphragmatic hernias (DH). In vivo studies are used to test not only mesh safety but increasingly comparative efficacy. Our work examines the field of in vivo mesh testing for HH and DH models to establish current practices and standards. Method: This systematic review was registered with PROSPERO. Medline and Embase databases were searched for relevant in vivo studies. 44 articles were identified and underwent abstract review, where 22 were excluded. 4 further studies were excluded after full text review – leaving 18 to undergo data extraction. Results: Of 18 studies identified, 9 used an in vivo HH model and 9 a DH model. 5 studies undertook mechanical testing on tissue samples – all uniaxial in nature. Testing strip widths ranged from 1-20mm (median 3mm). Testing speeds varied from 1.5-60mm/minute. Upon histology, the most commonly assessed structural and cellular factors were neovascularization and macrophages, respectively (n=9 each). Structural analysis was mostly qualitative, where cellular analysis was equally likely to be quantitative. 11 studies assessed adhesion formation, of which 8 used one of four scoring systems. 8 studies measured mesh shrinkage. Discussion: In vivo studies assessing mesh for HH and DH repair are uncommon. Within this relatively young field, we encourage surgical and materials testing institutions to discuss its standardisation.

Keywords: hiatus, diaphragmatic, hernia, mesh, materials testing, in vivo

Procedia PDF Downloads 198
6328 Doing Cause-and-Effect Analysis Using an Innovative Chat-Based Focus Group Method

Authors: Timothy Whitehill

Abstract:

This paper presents an innovative chat-based focus group method for collecting qualitative data to construct a cause-and-effect analysis in business research. This method was developed in response to the research and data collection challenges faced by the Covid-19 outbreak in the United Kingdom during 2020-21. This paper discusses the methodological approaches and builds a contemporary argument for its effectiveness in exploring cause-and-effect relationships in the context of focus group research, systems thinking and problem structuring methods. The pilot for this method was conducted between October 2020 and March 2021 and collected more than 7,000 words of chat-based data which was used to construct a consensus drawn cause-and-effect analysis. This method was developed in support of an ongoing Doctorate in Business Administration (DBA) thesis, which is using Design Science Research methodology to operationalize organisational resilience in UK construction sector firms.

Keywords: cause-and-effect analysis, focus group research, problem structuring methods, qualitative research, systems thinking

Procedia PDF Downloads 204
6327 Finding the Longest Common Subsequence in Normal DNA and Disease Affected Human DNA Using Self Organizing Map

Authors: G. Tamilpavai, C. Vishnuppriya

Abstract:

Bioinformatics is an active research area which combines biological matter as well as computer science research. The longest common subsequence (LCSS) is one of the major challenges in various bioinformatics applications. The computation of the LCSS plays a vital role in biomedicine and also it is an essential task in DNA sequence analysis in genetics. It includes wide range of disease diagnosing steps. The objective of this proposed system is to find the longest common subsequence which presents in a normal and various disease affected human DNA sequence using Self Organizing Map (SOM) and LCSS. The human DNA sequence is collected from National Center for Biotechnology Information (NCBI) database. Initially, the human DNA sequence is separated as k-mer using k-mer separation rule. Mean and median values are calculated from each separated k-mer. These calculated values are fed as input to the Self Organizing Map for the purpose of clustering. Then obtained clusters are given to the Longest Common Sub Sequence (LCSS) algorithm for finding common subsequence which presents in every clusters. It returns nx(n-1)/2 subsequence for each cluster where n is number of k-mer in a specific cluster. Experimental outcomes of this proposed system produce the possible number of longest common subsequence of normal and disease affected DNA data. Thus the proposed system will be a good initiative aid for finding disease causing sequence. Finally, performance analysis is carried out for different DNA sequences. The obtained values show that the retrieval of LCSS is done in a shorter time than the existing system.

Keywords: clustering, k-mers, longest common subsequence, SOM

Procedia PDF Downloads 243
6326 Adapting Tools for Text Monitoring and for Scenario Analysis Related to the Field of Social Disasters

Authors: Svetlana Cojocaru, Mircea Petic, Inga Titchiev

Abstract:

Humanity faces more and more often with different social disasters, which in turn can generate new accidents and catastrophes. To mitigate their consequences, it is important to obtain early possible signals about the events which are or can occur and to prepare the corresponding scenarios that could be applied. Our research is focused on solving two problems in this domain: identifying signals related that an accident occurred or may occur and mitigation of some consequences of disasters. To solve the first problem, methods of selecting and processing texts from global network Internet are developed. Information in Romanian is of special interest for us. In order to obtain the mentioned tools, we should follow several steps, divided into preparatory stage and processing stage. Throughout the first stage, we manually collected over 724 news articles and classified them into 10 categories of social disasters. It constitutes more than 150 thousand words. Using this information, a controlled vocabulary of more than 300 keywords was elaborated, that will help in the process of classification and identification of the texts related to the field of social disasters. To solve the second problem, the formalism of Petri net has been used. We deal with the problem of inhabitants’ evacuation in useful time. The analysis methods such as reachability or coverability tree and invariants technique to determine dynamic properties of the modeled systems will be used. To perform a case study of properties of extended evacuation system by adding time, the analysis modules of PIPE such as Generalized Stochastic Petri Nets (GSPN) Analysis, Simulation, State Space Analysis, and Invariant Analysis have been used. These modules helped us to obtain the average number of persons situated in the rooms and the other quantitative properties and characteristics related to its dynamics.

Keywords: lexicon of disasters, modelling, Petri nets, text annotation, social disasters

Procedia PDF Downloads 187
6325 Influential Factors Affecting the Creativity Scientific Problem Finding Ability of Social Science Ph.D. Students

Authors: Yuanyuan Song

Abstract:

For Ph.D. students, the skill of formulating incisive inquiries holds immense importance, as adept questioning can significantly unravel research complexities. Social Science Ph.D. students should possess specific abilities to formulate creative research questions, and identifying the most influential factors is essential. To respond to these questions, in this study, we engaged with Ph.D. candidates with social sciences backgrounds through interviews and questionnaires. Our objective was to identify the predominant determinants influencing their capacity to formulate inventive research queries, ultimately aiming to enhance the academic journey of social science doctoral candidates. Insights gleaned from semi-structured interviews and questionnaires with 15 doctoral scholars from different universities around the world highlighted that mentorship and scholarly exchanges, prior knowledge, positive mindset, and personal interests played pivotal roles in catalyzing these students' contemplation of research inquiries.

Keywords: Ph.D. education, higher education, creativity cultivation, creativity scientific problem finding ability

Procedia PDF Downloads 41
6324 A Machine Learning Approach for Detecting and Locating Hardware Trojans

Authors: Kaiwen Zheng, Wanting Zhou, Nan Tang, Lei Li, Yuanhang He

Abstract:

The integrated circuit industry has become a cornerstone of the information society, finding widespread application in areas such as industry, communication, medicine, and aerospace. However, with the increasing complexity of integrated circuits, Hardware Trojans (HTs) implanted by attackers have become a significant threat to their security. In this paper, we proposed a hardware trojan detection method for large-scale circuits. As HTs introduce physical characteristic changes such as structure, area, and power consumption as additional redundant circuits, we proposed a machine-learning-based hardware trojan detection method based on the physical characteristics of gate-level netlists. This method transforms the hardware trojan detection problem into a machine-learning binary classification problem based on physical characteristics, greatly improving detection speed. To address the problem of imbalanced data, where the number of pure circuit samples is far less than that of HTs circuit samples, we used the SMOTETomek algorithm to expand the dataset and further improve the performance of the classifier. We used three machine learning algorithms, K-Nearest Neighbors, Random Forest, and Support Vector Machine, to train and validate benchmark circuits on Trust-Hub, and all achieved good results. In our case studies based on AES encryption circuits provided by trust-hub, the test results showed the effectiveness of the proposed method. To further validate the method’s effectiveness for detecting variant HTs, we designed variant HTs using open-source HTs. The proposed method can guarantee robust detection accuracy in the millisecond level detection time for IC, and FPGA design flows and has good detection performance for library variant HTs.

Keywords: hardware trojans, physical properties, machine learning, hardware security

Procedia PDF Downloads 124
6323 Investigation into Relationship between Spaced Repetitions and Problems Solving Efficiency

Authors: Sidharth Talan, Rajlakshmi G. Majumdar

Abstract:

Problem-solving skill is one the few skills which is constantly endeavored to improve upon by the professionals and academicians around the world in order to sustain themselves in the ever-growing competitive environment. The given paper focuses on evaluating a hypothesized relationship between the problems solving efficiency of an individual with spaced repetitions, conducted with a time interval of one day over a period of two weeks. The paper has utilized uni-variate regression analysis technique to assess the best fit curve that can explain the significant relationship between the given two variables. The paper has incorporated Anagrams solving as the appropriate testing process for the analysis. Since Anagrams solving involves rearranging a jumbled word to form a correct word, it projects to be an efficient process to observe the attention span, visual- motor coordination and the verbal ability of an individual. Based on the analysis for a sample population of 30, it was observed that problem-solving efficiency of an individual, measured in terms of the score in each test was found to be significantly correlated with time period measured in days.

Keywords: Anagrams, histogram plot, moving average curve, spacing effect

Procedia PDF Downloads 143
6322 Temporal and Spatio-Temporal Stability Analyses in Mixed Convection of a Viscoelastic Fluid in a Porous Medium

Authors: P. Naderi, M. N. Ouarzazi, S. C. Hirata, H. Ben Hamed, H. Beji

Abstract:

The stability of mixed convection in a Newtonian fluid medium heated from below and cooled from above, also known as the Poiseuille-Rayleigh-Bénard problem, has been extensively investigated in the past decades. To our knowledge, mixed convection in porous media has received much less attention in the published literature. The present paper extends the mixed convection problem in porous media for the case of a viscoelastic fluid flow owing to its numerous environmental and industrial applications such as the extrusion of polymer fluids, solidification of liquid crystals, suspension solutions and petroleum activities. Without a superimposed through-flow, the natural convection problem of a viscoelastic fluid in a saturated porous medium has already been treated. The effects of the viscoelastic properties of the fluid on the linear and nonlinear dynamics of the thermoconvective instabilities have also been treated in this work. Consequently, the elasticity of the fluid can lead either to a Hopf bifurcation, giving rise to oscillatory structures in the strongly elastic regime, or to a stationary bifurcation in the weakly elastic regime. The objective of this work is to examine the influence of the main horizontal flow on the linear and characteristics of these two types of instabilities. Under the Boussinesq approximation and Darcy's law extended to a viscoelastic fluid, a temporal stability approach shows that the conditions for the appearance of longitudinal rolls are identical to those found in the absence of through-flow. For the general three-dimensional (3D) perturbations, a Squire transformation allows the deduction of the complex frequencies associated with the 3D problem using those obtained by solving the two-dimensional one. The numerical resolution of the eigenvalue problem concludes that the through-flow has a destabilizing effect and selects a convective configuration organized in purely transversal rolls which oscillate in time and propagate in the direction of the main flow. In addition, by using the mathematical formalism of absolute and convective instabilities, we study the nature of unstable three-dimensional disturbances. It is shown that for a non-vanishing through-flow, general three-dimensional instabilities are convectively unstable which means that in the absence of a continuous noise source these instabilities are drifted outside the porous medium, and no long-term pattern is observed. In contrast, purely transversal rolls may exhibit a transition to absolute instability regime and therefore affect the porous medium everywhere including in the absence of a noise source. The absolute instability threshold, the frequency and the wave number associated with purely transversal rolls are determined as a function of the Péclet number and the viscoelastic parameters. Results are discussed and compared to those obtained from laboratory experiments in the case of Newtonian fluids.

Keywords: instability, mixed convection, porous media, and viscoelastic fluid

Procedia PDF Downloads 325
6321 Hybrid Knowledge and Data-Driven Neural Networks for Diffuse Optical Tomography Reconstruction in Medical Imaging

Authors: Paola Causin, Andrea Aspri, Alessandro Benfenati

Abstract:

Diffuse Optical Tomography (DOT) is an emergent medical imaging technique which employs NIR light to estimate the spatial distribution of optical coefficients in biological tissues for diagnostic purposes, in a noninvasive and non-ionizing manner. DOT reconstruction is a severely ill-conditioned problem due to prevalent scattering of light in the tissue. In this contribution, we present our research in adopting hybrid knowledgedriven/data-driven approaches which exploit the existence of well assessed physical models and build upon them neural networks integrating the availability of data. Namely, since in this context regularization procedures are mandatory to obtain a reasonable reconstruction [1], we explore the use of neural networks as tools to include prior information on the solution. 2. Materials and Methods The idea underlying our approach is to leverage neural networks to solve PDE-constrained inverse problems of the form 𝒒 ∗ = 𝒂𝒓𝒈 𝒎𝒊𝒏𝒒 𝐃(𝒚, 𝒚̃), (1) where D is a loss function which typically contains a discrepancy measure (or data fidelity) term plus other possible ad-hoc designed terms enforcing specific constraints. In the context of inverse problems like (1), one seeks the optimal set of physical parameters q, given the set of observations y. Moreover, 𝑦̃ is the computable approximation of y, which may be as well obtained from a neural network but also in a classic way via the resolution of a PDE with given input coefficients (forward problem, Fig.1 box ). Due to the severe ill conditioning of the reconstruction problem, we adopt a two-fold approach: i) we restrict the solutions (optical coefficients) to lie in a lower-dimensional subspace generated by auto-decoder type networks. This procedure forms priors of the solution (Fig.1 box ); ii) we use regularization procedures of type 𝒒̂ ∗ = 𝒂𝒓𝒈𝒎𝒊𝒏𝒒 𝐃(𝒚, 𝒚̃)+ 𝑹(𝒒), where 𝑹(𝒒) is a regularization functional depending on regularization parameters which can be fixed a-priori or learned via a neural network in a data-driven modality. To further improve the generalizability of the proposed framework, we also infuse physics knowledge via soft penalty constraints (Fig.1 box ) in the overall optimization procedure (Fig.1 box ). 3. Discussion and Conclusion DOT reconstruction is severely hindered by ill-conditioning. The combined use of data-driven and knowledgedriven elements is beneficial and allows to obtain improved results, especially with a restricted dataset and in presence of variable sources of noise.

Keywords: inverse problem in tomography, deep learning, diffuse optical tomography, regularization

Procedia PDF Downloads 58