Search results for: combinatorial optimization problems
7315 Study on Ecological Water Demand Evaluation of Typical Mountainous Rivers in Zhejiang Province: Taking Kaihua River as an Example
Authors: Kaiping Xu, Aiju You, Lei Hua
Abstract:
In view of the ecological environmental problems and protection needs of mountainous rivers in Zhejiang province, a suitable ecological water demand evaluation system was established based on investigation and monitoring. Taking the Kaihua river as an example, the research on ecological water demand and the current situation evaluation were carried out. The main types of ecological water demand in Majin River are basic ecological flow and lake wetland outside the river, and instream flow and water demands for water quality in Zhongcun river. In the wet season, each ecological water demand is 18.05m3/s and 2.56m3 / s, and in the dry season is 3.00m3/s and 0.61m3/s. Three indexes of flow, duration and occurrence time are used to evaluate the ecological water demand. The degree of ecological water demand in the past three years is low level of satisfaction. Meanwhile, the existing problems are analyzed, and put forward reasonable and operable safeguards and suggestions.Keywords: Zhejiang province, mountainous river, ecological water demand, Kaihua river, evaluation
Procedia PDF Downloads 2467314 Robotic Arm-Automated Spray Painting with One-Shot Object Detection and Region-Based Path Optimization
Authors: Iqraq Kamal, Akmal Razif, Sivadas Chandra Sekaran, Ahmad Syazwan Hisaburi
Abstract:
Painting plays a crucial role in the aerospace manufacturing industry, serving both protective and cosmetic purposes for components. However, the traditional manual painting method is time-consuming and labor-intensive, posing challenges for the sector in achieving higher efficiency. Additionally, the current automated robot path planning has been a bottleneck for spray painting processes, as typical manual teaching methods are time-consuming, error-prone, and skill-dependent. Therefore, it is essential to develop automated tool path planning methods to replace manual ones, reducing costs and improving product quality. Focusing on flat panel painting in aerospace manufacturing, this study aims to address issues related to unreliable part identification techniques caused by the high-mixture, low-volume nature of the industry. The proposed solution involves using a spray gun and a UR10 robotic arm with a vision system that utilizes one-shot object detection (OS2D) to identify parts accurately. Additionally, the research optimizes path planning by concentrating on the region of interest—specifically, the identified part, rather than uniformly covering the entire painting tray.Keywords: aerospace manufacturing, one-shot object detection, automated spray painting, vision-based path optimization, deep learning, automation, robotic arm
Procedia PDF Downloads 857313 Stability Optimization of NABH₄ via PH and H₂O:NABH₄ Ratios for Large Scale Hydrogen Production
Authors: Parth Mehta, Vedasri Bai Khavala, Prabhu Rajagopal, Tiju Thomas
Abstract:
There is an increasing need for alternative clean fuels, and hydrogen (H₂) has long been considered a promising solution with a high calorific value (142MJ/kg). However, the storage of H₂ and expensive processes for its generation have hindered its usage. Sodium borohydride (NaBH₄) can potentially be used as an economically viable means of H₂ storage. Thus far, there have been attempts to optimize the life of NaBH₄ (half-life) in aqueous media by stabilizing it with sodium hydroxide (NaOH) for various pH values. Other reports have shown that H₂ yield and reaction kinetics remained constant for all ratios of H₂O to NaBH₄ > 30:1, without any acidic catalysts. Here we highlight the importance of pH and H₂O: NaBH₄ ratio (80:1, 40:1, 20:1 and 10:1 by weight), for NaBH₄ stabilization (half-life reaction time at room temperature) and corrosion minimization of H₂ reactor components. It is interesting to observe that at any particular pH>10 (e.g., pH = 10, 11 and 12), the H₂O: NaBH₄ ratio does not have the expected linear dependence with stability. On the contrary, high stability was observed at the ratio of 10:1 H₂O: NaBH₄ across all pH>10. When the H₂O: NaBH₄ ratio is increased from 10:1 to 20:1 and beyond (till 80:1), constant stability (% degradation) is observed with respect to time. For practical usage (consumption within 6 hours of making NaBH₄ solution), 15% degradation at pH 11 and NaBH₄: H₂O ratio of 10:1 is recommended. Increasing this ratio demands higher NaOH concentration at the same pH, thus requiring a higher concentration or volume of acid (e.g., HCl) for H₂ generation. The reactions are done with tap water to render the results useful from an industrial standpoint. The observed stability regimes are rationalized based on complexes associated with NaBH₄ when solvated in water, which depend sensitively on both pH and NaBH₄: H₂O ratio.Keywords: hydrogen, sodium borohydride, stability optimization, H₂O:NaBH₄ ratio
Procedia PDF Downloads 1277312 Chaotic Sequence Noise Reduction and Chaotic Recognition Rate Improvement Based on Improved Local Geometric Projection
Authors: Rubin Dan, Xingcai Wang, Ziyang Chen
Abstract:
A chaotic time series noise reduction method based on the fusion of the local projection method, wavelet transform, and particle swarm algorithm (referred to as the LW-PSO method) is proposed to address the problem of false recognition due to noise in the recognition process of chaotic time series containing noise. The method first uses phase space reconstruction to recover the original dynamical system characteristics and removes the noise subspace by selecting the neighborhood radius; then it uses wavelet transform to remove D1-D3 high-frequency components to maximize the retention of signal information while least-squares optimization is performed by the particle swarm algorithm. The Lorenz system containing 30% Gaussian white noise is simulated and verified, and the phase space, SNR value, RMSE value, and K value of the 0-1 test method before and after noise reduction of the Schreiber method, local projection method, wavelet transform method, and LW-PSO method are compared and analyzed, which proves that the LW-PSO method has a better noise reduction effect compared with the other three common methods. The method is also applied to the classical system to evaluate the noise reduction effect of the four methods and the original system identification effect, which further verifies the superiority of the LW-PSO method. Finally, it is applied to the Chengdu rainfall chaotic sequence for research, and the results prove that the LW-PSO method can effectively reduce the noise and improve the chaos recognition rate.Keywords: Schreiber noise reduction, wavelet transform, particle swarm optimization, 0-1 test method, chaotic sequence denoising
Procedia PDF Downloads 2047311 The Management of Urban Facilities in the City of Chlef
Authors: Belakhdar Salah Brahim
Abstract:
The Urban management is a major element of social control of public space and thus the functioning of society. As such, it is a key element of a social conception of sustainable development. Also, it is a cross-cutting sector that relies on land management, infrastructure management, habitat management, management of social services, the management of economic development, etc. This study aims to study how urban management focusing on the study of problems related to urban waste management in developing countries. It appears from the study that the city management is to improve infrastructure and urban services in order to increase the city's development and improve living conditions in cities. It covers various aspects including management of urban space, economic management, administrative management, asset management or infrastructure and finally waste management. Environmental management is important because it solves the pollution problems of life and preserve resources for future generations. Changing perceptions of waste has led to the definition of new policies for integrated waste management requirements appropriate to the urban site.Keywords: urbanization, urban management, environmental management, waste management
Procedia PDF Downloads 4377310 Maximizing the Role of Companion Teachers for the Achievement of Professional Competencies and Pedagogics Workshop Activities of Teacher Professional Participants in the Faculty of Teaching and Education of Mulawarman University
Authors: Makrina Tindangen
Abstract:
The problems faced by participants of teacher profession program in Faculty of teaching and education Mulawarman University is professional and pedagogic competence. Professional competence related to the mastery of teaching materials, while pedagogic competence related with the ability to plan and to implement learning. Based on the problems, the purpose of the research is to maximize the role of companion teacher for the achievement of professional and pedagogic competencies in the workshop of the participants of teacher professional education in the Faculty of Teaching and Education of Mulawarman University. Qualitative research method with interview guidance and document to get in-depth data on how to maximize the role of companion teachers in the achievement of professional and pedagogic competencies in the workshop participants of professional education participants. Location of this research is on the Faculty of Teaching and Education of Mulawarman University, Samarinda City, East Kalimantan Province. Research respondents were 12 teachers of workshop facilitator. Descriptive data analysis is through interpretation of interview data. The conclusion of the research result, how to maximize the role of assistant teachers in workshop activities for the professional competence and pedagogic competence of professional teacher training program participants, through facilitation activities conducted by teachers of companion related to real problems faced by students in school, so that the workshop participants have professional competence and pedagogic as an initial competence before carrying out practical activities of field experience in school.Keywords: companion teacher, professional and pedagogical competence, activities, workshop participants
Procedia PDF Downloads 1927309 Development of Electronic Governance as an Element of Reforming State Governance According to the Adjarian Example
Authors: Irakli Manvelidze, Genadi Iashvili, Giga Phartenadze, Giorgi Katamadze
Abstract:
Establishment of electronic governance in the region is facing serious problems. Organizational, technical, social and methodological problems have been identified after the research. These problems currently create serious barriers and prevent the development of effective e-governance. Lack of human resources, difference in program targets of the centre and the region, lack of citizens’ awareness about the project of electronic governance are other issues that should be mentioned. In spite of positive changes the overall situation concerning development of modern information-communication technologies in Adjara is not satisfactory. The information systems in the region can be described as transforming in a democratic way which needs serious reforms. Current situation shows that unsystematic, uncoordinated actions were made which overall represents more chaotic rather than coordinated systematic process. Therefore, a strategic document ‘Adjarian Electronic Government’ should be created which will ensure systematic development of electronic governance in the region. The implementation of the strategy of ‘Adjarian Electronic Government’ should be based on not only conceptual and instrumental but also legal basics. A legal normative basis should be created which will include formation of electronic government’s instrumental basis as well as creation of united regional system of electronic document management. Meanwhile types of documents which would be used in inter institutional relations should be defined under a legal norm. Creation of regional united system of e-filing will regulate regional public institutions, relations between local self-government and public organizations as well as it will ensure coordinated work of all regional public institutions.Keywords: e-government, information society, public administration, reforming state governance, public institutions
Procedia PDF Downloads 2877308 A Robust Optimization of Chassis Durability/Comfort Compromise Using Chebyshev Polynomial Chaos Expansion Method
Authors: Hanwei Gao, Louis Jezequel, Eric Cabrol, Bernard Vitry
Abstract:
The chassis system is composed of complex elements that take up all the loads from the tire-ground contact area and thus it plays an important role in numerous specifications such as durability, comfort, crash, etc. During the development of new vehicle projects in Renault, durability validation is always the main focus while deployment of comfort comes later in the project. Therefore, sometimes design choices have to be reconsidered because of the natural incompatibility between these two specifications. Besides, robustness is also an important point of concern as it is related to manufacturing costs as well as the performance after the ageing of components like shock absorbers. In this paper an approach is proposed aiming to realize a multi-objective optimization between chassis endurance and comfort while taking the random factors into consideration. The adaptive-sparse polynomial chaos expansion method (PCE) with Chebyshev polynomial series has been applied to predict responses’ uncertainty intervals of a system according to its uncertain-but-bounded parameters. The approach can be divided into three steps. First an initial design of experiments is realized to build the response surfaces which represent statistically a black-box system. Secondly within several iterations an optimum set is proposed and validated which will form a Pareto front. At the same time the robustness of each response, served as additional objectives, is calculated from the pre-defined parameter intervals and the response surfaces obtained in the first step. Finally an inverse strategy is carried out to determine the parameters’ tolerance combination with a maximally acceptable degradation of the responses in terms of manufacturing costs. A quarter car model has been tested as an example by applying the road excitations from the actual road measurements for both endurance and comfort calculations. One indicator based on the Basquin’s law is defined to compare the global chassis durability of different parameter settings. Another indicator related to comfort is obtained from the vertical acceleration of the sprung mass. An optimum set with best robustness has been finally obtained and the reference tests prove a good robustness prediction of Chebyshev PCE method. This example demonstrates the effectiveness and reliability of the approach, in particular its ability to save computational costs for a complex system.Keywords: chassis durability, Chebyshev polynomials, multi-objective optimization, polynomial chaos expansion, ride comfort, robust design
Procedia PDF Downloads 1577307 Optimizing Data Transfer and Processing in Multi-Cloud Environments for Big Data Workloads
Authors: Gaurav Kumar Sinha
Abstract:
In an era defined by the proliferation of data and the utilization of cloud computing environments, the efficient transfer and processing of big data workloads across multi-cloud platforms have emerged as critical challenges. This research paper embarks on a comprehensive exploration of the complexities associated with managing and optimizing big data in a multi-cloud ecosystem.The foundation of this study is rooted in the recognition that modern enterprises increasingly rely on multiple cloud providers to meet diverse business needs, enhance redundancy, and reduce vendor lock-in. As a consequence, managing data across these heterogeneous cloud environments has become intricate, necessitating innovative approaches to ensure data integrity, security, and performance.The primary objective of this research is to investigate strategies and techniques for enhancing the efficiency of data transfer and processing in multi-cloud scenarios. It recognizes that big data workloads are characterized by their sheer volume, variety, velocity, and complexity, making traditional data management solutions insufficient for harnessing the full potential of multi-cloud architectures.The study commences by elucidating the challenges posed by multi-cloud environments in the context of big data. These challenges encompass data fragmentation, latency, security concerns, and cost optimization. To address these challenges, the research explores a range of methodologies and solutions. One of the key areas of focus is data transfer optimization. The paper delves into techniques for minimizing data movement latency, optimizing bandwidth utilization, and ensuring secure data transmission between different cloud providers. It evaluates the applicability of dedicated data transfer protocols, intelligent data routing algorithms, and edge computing approaches in reducing transfer times.Furthermore, the study examines strategies for efficient data processing across multi-cloud environments. It acknowledges that big data processing requires distributed and parallel computing capabilities that span across cloud boundaries. The research investigates containerization and orchestration technologies, serverless computing models, and interoperability standards that facilitate seamless data processing workflows.Security and data governance are paramount concerns in multi-cloud environments. The paper explores methods for ensuring data security, access control, and compliance with regulatory frameworks. It considers encryption techniques, identity and access management, and auditing mechanisms as essential components of a robust multi-cloud data security strategy.The research also evaluates cost optimization strategies, recognizing that the dynamic nature of multi-cloud pricing models can impact the overall cost of data transfer and processing. It examines approaches for workload placement, resource allocation, and predictive cost modeling to minimize operational expenses while maximizing performance.Moreover, this study provides insights into real-world case studies and best practices adopted by organizations that have successfully navigated the challenges of multi-cloud big data management. It presents a comparative analysis of various multi-cloud management platforms and tools available in the market.Keywords: multi-cloud environments, big data workloads, data transfer optimization, data processing strategies
Procedia PDF Downloads 727306 Applications of AI, Machine Learning, and Deep Learning in Cyber Security
Authors: Hailyie Tekleselase
Abstract:
Deep learning is increasingly used as a building block of security systems. However, neural networks are hard to interpret and typically solid to the practitioner. This paper presents a detail survey of computing methods in cyber security, and analyzes the prospects of enhancing the cyber security capabilities by suggests that of accelerating the intelligence of the security systems. There are many AI-based applications used in industrial scenarios such as Internet of Things (IoT), smart grids, and edge computing. Machine learning technologies require a training process which introduces the protection problems in the training data and algorithms. We present machine learning techniques currently applied to the detection of intrusion, malware, and spam. Our conclusions are based on an extensive review of the literature as well as on experiments performed on real enterprise systems and network traffic. We conclude that problems can be solved successfully only when methods of artificial intelligence are being used besides human experts or operators.Keywords: artificial intelligence, machine learning, deep learning, cyber security, big data
Procedia PDF Downloads 1317305 Flow Field Optimization for Proton Exchange Membrane Fuel Cells
Authors: Xiao-Dong Wang, Wei-Mon Yan
Abstract:
The flow field design in the bipolar plates affects the performance of the proton exchange membrane (PEM) fuel cell. This work adopted a combined optimization procedure, including a simplified conjugate-gradient method and a completely three-dimensional, two-phase, non-isothermal fuel cell model, to look for optimal flow field design for a single serpentine fuel cell of size 9×9 mm with five channels. For the direct solution, the two-fluid method was adopted to incorporate the heat effects using energy equations for entire cells. The model assumes that the system is steady; the inlet reactants are ideal gases; the flow is laminar; and the porous layers such as the diffusion layer, catalyst layer and PEM are isotropic. The model includes continuity, momentum and species equations for gaseous species, liquid water transport equations in the channels, gas diffusion layers, and catalyst layers, water transport equation in the membrane, electron and proton transport equations. The Bulter-Volumer equation was used to describe electrochemical reactions in the catalyst layers. The cell output power density Pcell is maximized subjected to an optimal set of channel heights, H1-H5, and channel widths, W2-W5. The basic case with all channel heights and widths set at 1 mm yields a Pcell=7260 Wm-2. The optimal design displays a tapered characteristic for channels 1, 3 and 4, and a diverging characteristic in height for channels 2 and 5, producing a Pcell=8894 Wm-2, about 22.5% increment. The reduced channel heights of channels 2-4 significantly increase the sub-rib convection and widths for effectively removing liquid water and oxygen transport in gas diffusion layer. The final diverging channel minimizes the leakage of fuel to outlet via sub-rib convection from channel 4 to channel 5. Near-optimal design without huge loss in cell performance but is easily manufactured is tested. The use of a straight, final channel of 0.1 mm height has led to 7.37% power loss, while the design with all channel widths to be 1 mm with optimal channel heights obtained above yields only 1.68% loss of current density. The presence of a final, diverging channel has greater impact on cell performance than the fine adjustment of channel width at the simulation conditions set herein studied.Keywords: optimization, flow field design, simplified conjugate-gradient method, serpentine flow field, sub-rib convection
Procedia PDF Downloads 3027304 A Mixed-Integer Nonlinear Program to Optimally Pace and Fuel Ultramarathons
Authors: Kristopher A. Pruitt, Justin M. Hill
Abstract:
The purpose of this research is to determine the pacing and nutrition strategies which minimize completion time and carbohydrate intake for athletes competing in ultramarathon races. The model formulation consists of a two-phase optimization. The first-phase mixed-integer nonlinear program (MINLP) determines the minimum completion time subject to the altitude, terrain, and distance of the race, as well as the mass and cardiovascular fitness of the athlete. The second-phase MINLP determines the minimum total carbohydrate intake required for the athlete to achieve the completion time prescribed by the first phase, subject to the flow of carbohydrates through the stomach, liver, and muscles. Consequently, the second phase model provides the optimal pacing and nutrition strategies for a particular athlete for each kilometer of a particular race. Validation of the model results over a wide range of athlete parameters against completion times for real competitive events suggests strong agreement. Additionally, the kilometer-by-kilometer pacing and nutrition strategies, the model prescribes for a particular athlete suggest unconventional approaches could result in lower completion times. Thus, the MINLP provides prescriptive guidance that athletes can leverage when developing pacing and nutrition strategies prior to competing in ultramarathon races. Given the highly-variable topographical characteristics common to many ultramarathon courses and the potential inexperience of many athletes with such courses, the model provides valuable insight to competitors who might otherwise fail to complete the event due to exhaustion or carbohydrate depletion.Keywords: nutrition, optimization, pacing, ultramarathons
Procedia PDF Downloads 1917303 Facilitating Waste Management to Achieve Sustainable Residential Built Environments
Authors: Ingy Ibrahim El-Darwish, Neveen Youssef Azmy
Abstract:
The endowment of a healthy environment can be implemented by endorsing sustainable fundamentals. Design of sustainable buildings through recycling of waste, can reduce health problems, provide good environments and contribute to the aesthetically pleasing entourage. Such environments can help in providing energy-saving alternatives to consolidate the principles of sustainability. The poor community awareness and the absence of laws and legislation in Egypt for waste management specifically in residential areas have led to an inability to provide an integrated system for waste management in urban and rural areas. Many problems and environmental challenges face the Egyptian urban environments. From these problems, is the lack of a cohesive vision for waste collection and recycling for energy-saving. The second problem is the lack public awareness of the short term and long term vision of waste management. Bad practices have adversely affected the efficiency of environmental management systems due to lack of urban legislations that codify collection and recycling of residential communities in Egyptian urban environments. Hence, this research tries to address residents on waste management matters to facilitate legislative process on waste collection and classification within residential units and outside them in a preparation phase for recycling in the Egyptian urban environments. In order to achieve this goal, one of the Egyptian communities has been addressed, analyzed and studied. Waste collection, classification, separation and access to recycling places in the urban city are proposed in preparation for a legislation ruling and regulating the process. Hence, sustainable principles are to be achieved.Keywords: recycling, residential buildings, sustainability, waste
Procedia PDF Downloads 3337302 High Aspect Ratio Micropillar Array Based Microfluidic Viscometer
Authors: Ahmet Erten, Adil Mustafa, Ayşenur Eser, Özlem Yalçın
Abstract:
We present a new viscometer based on a microfluidic chip with elastic high aspect ratio micropillar arrays. The displacement of pillar tips in flow direction can be used to analyze viscosity of liquid. In our work, Computational Fluid Dynamics (CFD) is used to analyze pillar displacement of various micropillar array configurations in flow direction at different viscosities. Following CFD optimization, micro-CNC based rapid prototyping is used to fabricate molds for microfluidic chips. Microfluidic chips are fabricated out of polydimethylsiloxane (PDMS) using soft lithography methods with molds machined out of aluminum. Tip displacements of micropillar array (300 µm in diameter and 1400 µm in height) in flow direction are recorded using a microscope mounted camera, and the displacements are analyzed using image processing with an algorithm written in MATLAB. Experiments are performed with water-glycerol solutions mixed at 4 different ratios to attain 1 cP, 5 cP, 10 cP and 15 cP viscosities at room temperature. The prepared solutions are injected into the microfluidic chips using a syringe pump at flow rates from 10-100 mL / hr and the displacement versus flow rate is plotted for different viscosities. A displacement of around 1.5 µm was observed for 15 cP solution at 60 mL / hr while only a 1 µm displacement was observed for 10 cP solution. The presented viscometer design optimization is still in progress for better sensitivity and accuracy. Our microfluidic viscometer platform has potential for tailor made microfluidic chips to enable real time observation and control of viscosity changes in biological or chemical reactions.Keywords: Computational Fluid Dynamics (CFD), high aspect ratio, micropillar array, viscometer
Procedia PDF Downloads 2497301 Coupling of Microfluidic Droplet Systems with ESI-MS Detection for Reaction Optimization
Authors: Julia R. Beulig, Stefan Ohla, Detlev Belder
Abstract:
In contrast to off-line analytical methods, lab-on-a-chip technology delivers direct information about the observed reaction. Therefore, microfluidic devices make an important scientific contribution, e.g. in the field of synthetic chemistry. Herein, the rapid generation of analytical data can be applied for the optimization of chemical reactions. These microfluidic devices enable a fast change of reaction conditions as well as a resource saving method of operation. In the presented work, we focus on the investigation of multiphase regimes, more specifically on a biphasic microfluidic droplet systems. Here, every single droplet is a reaction container with customized conditions. The biggest challenge is the rapid qualitative and quantitative readout of information as most detection techniques for droplet systems are non-specific, time-consuming or too slow. An exception is the electrospray mass spectrometry (ESI-MS). The combination of a reaction screening platform with a rapid and specific detection method is an important step in droplet-based microfluidics. In this work, we present a novel approach for synthesis optimization on the nanoliter scale with direct ESI-MS detection. The development of a droplet-based microfluidic device, which enables the modification of different parameters while simultaneously monitoring the effect on the reaction within a single run, is shown. By common soft- and photolithographic techniques a polydimethylsiloxane (PDMS) microfluidic chip with different functionalities is developed. As an interface for the MS detection, we use a steel capillary for ESI and improve the spray stability with a Teflon siphon tubing, which is inserted underneath the steel capillary. By optimizing the flow rates, it is possible to screen parameters of various reactions, this is exemplarity shown by a Domino Knoevenagel Hetero-Diels-Alder reaction. Different starting materials, catalyst concentrations and solvent compositions are investigated. Due to the high repetition rate of the droplet production, each set of reaction condition is examined hundreds of times. As a result, of the investigation, we receive possible reagents, the ideal water-methanol ratio of the solvent and the most effective catalyst concentration. The developed system can help to determine important information about the optimal parameters of a reaction within a short time. With this novel tool, we make an important step on the field of combining droplet-based microfluidics with organic reaction screening.Keywords: droplet, mass spectrometry, microfluidics, organic reaction, screening
Procedia PDF Downloads 3047300 Solving Ill-Posed Initial Value Problems for Switched Differential Equations
Authors: Eugene Stepanov, Arcady Ponosov
Abstract:
To model gene regulatory networks one uses ordinary differential equations with switching nonlinearities, where the initial value problem is known to be well-posed if the trajectories cross the discontinuities transversally. Otherwise, the initial value problem is usually ill-posed, which lead to theoretical and numerical complications. In the presentation, it is proposed to apply the theory of hybrid dynamical systems, rather than switched ones, to regularize the problem. 'Hybridization' of the switched system means that one attaches a dynamic discrete component ('automaton'), which follows the trajectories of the original system and governs its dynamics at the points of ill-posedness of the initial value problem making it well-posed. The construction of the automaton is based on the classification of the attractors of the specially designed adjoint dynamical system. Several examples are provided in the presentation, which support the suggested analysis. The method can also be of interest in other applied fields, where differential equations contain switchings, e.g. in neural field models.Keywords: hybrid dynamical systems, ill-posed problems, singular perturbation analysis, switching nonlinearities
Procedia PDF Downloads 1907299 A Computational Framework for Decoding Hierarchical Interlocking Structures with SL Blocks
Authors: Yuxi Liu, Boris Belousov, Mehrzad Esmaeili Charkhab, Oliver Tessmann
Abstract:
This paper presents a computational solution for designing reconfigurable interlocking structures that are fully assembled with SL Blocks. Formed by S-shaped and L-shaped tetracubes, SL Block is a specific type of interlocking puzzle. Analogous to molecular self-assembly, the aggregation of SL blocks will build a reversible hierarchical and discrete system where a single module can be numerously replicated to compose semi-interlocking components that further align, wrap, and braid around each other to form complex high-order aggregations. These aggregations can be disassembled and reassembled, responding dynamically to design inputs and changes with a unique capacity for reconfiguration. To use these aggregations as architectural structures, we developed computational tools that automate the configuration of SL blocks based on architectural design objectives. There are three critical phases in our work. First, we revisit the hierarchy of the SL block system and devise a top-down-type design strategy. From this, we propose two key questions: 1) How to translate 3D polyominoes into SL block assembly? 2) How to decompose the desired voxelized shapes into a set of 3D polyominoes with interlocking joints? These two questions can be considered the Hamiltonian path problem and the 3D polyomino tiling problem. Then, we derive our solution to each of them based on two methods. The first method is to construct the optimal closed path from an undirected graph built from the voxelized shape and translate the node sequence of the resulting path into the assembly sequence of SL blocks. The second approach describes interlocking relationships of 3D polyominoes as a joint connection graph. Lastly, we formulate the desired shapes and leverage our methods to achieve their reconfiguration within different levels. We show that our computational strategy will facilitate the efficient design of hierarchical interlocking structures with a self-replicating geometric module.Keywords: computational design, SL-blocks, 3D polyomino puzzle, combinatorial problem
Procedia PDF Downloads 1367298 Use of Galileo Advanced Features in Maritime Domain
Authors: Olivier Chaigneau, Damianos Oikonomidis, Marie-Cecile Delmas
Abstract:
GAMBAS (Galileo Advanced features for the Maritime domain: Breakthrough Applications for Safety and security) is a project funded by the European Space Program Agency (EUSPA) aiming at identifying the search-and-rescue and ship security alert system needs for maritime users (including operators and fishing stakeholders) and developing operational concepts to answer these needs. The general objective of the GAMBAS project is to support the deployment of Galileo exclusive features in the maritime domain in order to improve safety and security at sea, detection of illegal activities and associated surveillance means, resilience to natural and human-induced emergency situations, and develop, integrate, demonstrate, standardize and disseminate these new associated capabilities. The project aims to demonstrate: improvement of the SAR (Search And Rescue) and SSAS (Ship Security Alert System) detection and response to maritime distress through the integration of new features into the beacon for SSAS in terms of cost optimization, user-friendly aspects, integration of Galileo and OS NMA (Open Service Navigation Message Authentication) reception for improved authenticated localization performance and reliability, and at sea triggering capabilities, optimization of the responsiveness of RCCs (Rescue Co-ordination Centre) towards the distress situations affecting vessels, the adaptation of the MCCs (Mission Control Center) and MEOLUT (Medium Earth Orbit Local User Terminal) to the data distribution of SSAS alerts.Keywords: Galileo new advanced features, maritime, safety, security
Procedia PDF Downloads 967297 An Integrated Approach for Optimal Selection of Machining Parameters in Laser Micro-Machining Process
Authors: A. Gopala Krishna, M. Lakshmi Chaitanya, V. Kalyana Manohar
Abstract:
In the existent analysis, laser micro machining (LMM) of Silicon carbide (SiCp) reinforced Aluminum 7075 Metal Matrix Composite (Al7075/SiCp MMC) was studied. While machining, Because of the intense heat generated, A layer gets formed on the work piece surface which is called recast layer and this layer is detrimental to the surface quality of the component. The recast layer needs to be as small as possible for precise applications. Therefore, The height of recast layer and the depth of groove which are conflicting in nature were considered as the significant manufacturing criteria, Which determines the pursuit of a machining process obtained in LMM of Al7075/10%SiCp composite. The present work formulates the depth of groove and height of recast layer in relation to the machining parameters using the Response Surface Methodology (RSM) and correspondingly, The formulated mathematical models were put to use for optimization. Since the effect of machining parameters on the depth of groove and height of recast layer was contradictory, The problem was explicated as a multi objective optimization problem. Moreover, An evolutionary Non-dominated sorting genetic algorithm (NSGA-II) was employed to optimize the model established by RSM. Subsequently this algorithm was also adapted to achieve the Pareto optimal set of solutions that provide a detailed illustration for making the optimal solutions. Eventually experiments were conducted to affirm the results obtained from RSM and NSGA-II.Keywords: Laser Micro Machining (LMM), depth of groove, Height of recast layer, Response Surface Methodology (RSM), non-dominated sorting genetic algorithm
Procedia PDF Downloads 3497296 Problems Arising in Visual Perception
Authors: K. A. Tharanga, K. H. H. Damayanthi
Abstract:
Perception is an epistemological concept discussed in Philosophy. Perception, in other word, vision, is one of the ways that human beings get empirical knowledge after five senses. However, we face innumerable problems when achieving knowledge from perception, and therefore the knowledge gained through perception is uncertain. what we see in the external world is not real. These are the major issues that we face when receiving knowledge through perception. Sometimes there is no physical existence of what we really see. In such cases, the perception is relative. The following frames will be taken into consideration when perception is analyzed illusions and delusions, the figure of a physical object, appearance and the reality of a physical object, time factor, and colour of a physical object.seeing and knowing become vary according to the above conceptual frames. We cannot come to a proper conclusion of what we see in the empirical world. Because the things that we see are not really there. Hence the scientific knowledge which is gained from observation is doubtful. All the factors discussed in science remain in the physical world. There is a leap from ones existence to the existence of a world outside his/her mind. Indeed, one can suppose that what he/she takes to be real is just anmassive deception. However, depending on the above facts, if someone begins to doubt about the whole world, it is unavoidable to become his/her view a scepticism or nihilism. This is a certain reality.Keywords: empirical, perception, sceptisism, nihilism
Procedia PDF Downloads 987295 Formulation of Famotidine Solid Lipid Nanoparticles (SLN): Preparation, Evaluation and Release Study
Authors: Rachmat Mauludin, Nurmazidah
Abstract:
Background and purpose: Famotidine is an H2 receptor blocker. Absorption orally is rapid enough, but famotidine can be degraded by stomach acid causing dose reduction until 35.8% after 50 minutes. This drug also undergoes first-pass metabolism which reduced its bio availability only until 40-50%. To overcome these problems, Solid Lipid Nano particles (SLNs) as alternative delivery systems can be formulated. SLNs is a lipid-based drug delivery technology with 50-1000 nm particle size, where the drug incorporated into the bio compatible lipids and the lipid particles are stabilized using appropriate stabilizers. When the particle size is 200 nm or below, lipid containing famotidine can be absorbed through the lymphatic vessels to the subclavian vein, so first-pass metabolism can be avoided. Method: Famotidine SLNs with various compositions of stabilizer was prepared using a high-speed homogenization and sonication method. Then, the particle size distribution, zeta potential, entrapment efficiency, particle morphology and in vitro release profiles were evaluated. Optimization of sonication time also carried out. Result: Particle size of SLN by Particle Size Analyzer was in range 114.6 up to 455.267 nm. Ultrasonicated SLNs within 5 minutes generated smaller particle size than SLNs which was ultrasonicated for 10 and 15 minutes. Entrapment efficiency of SLNs were 74.17 up to 79.45%. Particle morphology of the SLNs was spherical and distributed individually. Release study of Famotidine revealed that in acid medium, 28.89 up to 80.55% of famotidine could be released after 2 hours. Nevertheless in basic medium, famotidine was released 40.5 up to 86.88% in the same period. Conclusion: The best formula was SLNs which stabilized by 4% Poloxamer 188 and 1 % Span 20, that had particle size 114.6 nm in diameter, 77.14% famotidine entrapped, and the particle morphology was spherical and distributed individually. SLNs with the best drug release profile was SLNs which stabilized by 4% Eudragit L 100-55 and 1% Tween 80 which had released 36.34 % in pH 1.2 solution, and 74.13% in pH 7.4 solution after 2 hours. The optimum sonication time was 5 minutes.Keywords: famotodine, SLN, high speed homogenization, particle size, release study
Procedia PDF Downloads 8657294 Quality of Service Based Routing Algorithm for Real Time Applications in MANETs Using Ant Colony and Fuzzy Logic
Authors: Farahnaz Karami
Abstract:
Routing is an important, challenging task in mobile ad hoc networks due to node mobility, lack of central control, unstable links, and limited resources. An ant colony has been found to be an attractive technique for routing in Mobile Ad Hoc Networks (MANETs). However, existing swarm intelligence based routing protocols find an optimal path by considering only one or two route selection metrics without considering correlations among such parameters making them unsuitable lonely for routing real time applications. Fuzzy logic combines multiple route selection parameters containing uncertain information or imprecise data in nature, but does not have multipath routing property naturally in order to provide load balancing. The objective of this paper is to design a routing algorithm using fuzzy logic and ant colony that can solve some of routing problems in mobile ad hoc networks, such as nodes energy consumption optimization to increase network lifetime, link failures rate reduction to increase packet delivery reliability and providing load balancing to optimize available bandwidth. In proposed algorithm, the path information will be given to fuzzy inference system by ants. Based on the available path information and considering the parameters required for quality of service (QoS), the fuzzy cost of each path is calculated and the optimal paths will be selected. NS2.35 simulation tools are used for simulation and the results are compared and evaluated with the newest QoS based algorithms in MANETs according to packet delivery ratio, end-to-end delay and routing overhead ratio criterions. The simulation results show significant improvement in the performance of these networks in terms of decreasing end-to-end delay, and routing overhead ratio, and also increasing packet delivery ratio.Keywords: mobile ad hoc networks, routing, quality of service, ant colony, fuzzy logic
Procedia PDF Downloads 697293 Mechanism of in Vitro Inhibition of Alpha-Amylase, Alpha-Glucosidase by Ethanolic Extracts of Polyalthia Longifolia, Its in Vitro Cytotoxicity on L6, Vero Cell-Lines and Influence of Glucose Uptake by Rat Hemi-Diaphragm
Authors: P. Gayathri, G. P. Jeyanthi
Abstract:
The bark of Polyalthia longifolia is used in ayurvedic system of medicine for the manangement of various ailments including diabetes mellitus. The bark of P. longifolia extracts was extracted using various polar and non-polar solvents and tested for inhibition of alpha-amylase and alpha-glucosidase among which the ethanolic extracts were found to be more potent. The ethanolic extracts of the bark were tested for the in vitro inhibition of alpha-amylase using starch as substrate and alpha-glucosidase using p-nitro phenyl alpha-D-gluco pyranoside as substrate to establish its in vitro antidiabetic effect. The mechanism of inhibition was determined by Dixon plot and Cornish-Bowden plot. The cytotoxic effect of the extract was tested on L6 and Vero cell-lines. The extract was partially purified by TLC. The individual effect of the ethanolic extract, TLC fractions and its combinatorial effect with insulin and glibenclamide on glucose uptake by rat hemi-diaphragm were studied.Results revealed that the ethanolic extracts of Polyalthia longifolia bark exhibited competitive inhibition of alpha-amylase and alpha-glucosidase. The extracts were also found not to be cytotoxic at the highest dose of 1 mg/mL. Glucose uptake study revealed that the extract alone and when combined with insulin, decreased the glucose uptake when compared to insulin control, however the purified TLC fractions exhibited significantly higher (p<0.05) glucose uptake by the rat hemi-diaphragm when compared to insulin. The study shows various possible mechanism of in vitro antidiabetic effect of the P. longifolia bark.Keywords: alpha-amylase, alpha-glucosidase, dixon, cornish-bowden, L6 , Vero cell-lines, glucose uptake, polyalthia longifolia bark, ethanolic extract, TLC fractions
Procedia PDF Downloads 4767292 Finding a Set of Long Common Substrings with Repeats from m Input Strings
Authors: Tiantian Li, Lusheng Wang, Zhaohui Zhan, Daming Zhu
Abstract:
In this paper, we propose two string problems, and study algorithms and complexity of various versions for those problems. Let S = {s₁, s₂, . . . , sₘ} be a set of m strings. A common substring of S is a substring appearing in every string in S. Given a set of m strings S = {s₁, s₂, . . . , sₘ} and a positive integer k, we want to find a set C of k common substrings of S such that the k common substrings in C appear in the same order and have no overlap among the m input strings in S, and the total length of the k common substring in C is maximized. This problem is referred to as the longest total length of k common substrings from m input strings (LCSS(k, m) for short). The other problem we study here is called the longest total length of a set of common substrings with length more than l from m input string (LSCSS(l, m) for short). Given a set of m strings S = {s₁, s₂, . . . , sₘ} and a positive integer l, for LSCSS(l, m), we want to find a set of common substrings of S, each is of length more than l, such that the total length of all the common substrings is maximized. We show that both problems are NP-hard when k and m are variables. We propose dynamic programming algorithms with time complexity O(k n₁n₂) and O(n₁n₂) to solve LCSS(k, 2) and LSCSS(l, 2), respectively, where n1 and n₂ are the lengths of the two input strings. We then design an algorithm for LSCSS(l, m) when every length > l common substring appears once in each of the m − 1 input strings. The running time is O(n₁²m), where n1 is the length of the input string with no restriction on length > l common substrings. Finally, we propose a fixed parameter algorithm for LSCSS(l, m), where each length > l common substring appears m − 1 + c times among the m − 1 input strings (other than s1). In other words, each length > l common substring may repeatedly appear at most c times among the m − 1 input strings {s₂, s₃, . . . , sₘ}. The running time of the proposed algorithm is O((n12ᶜ)²m), where n₁ is the input string with no restriction on repeats. The LSCSS(l, m) is proposed to handle whole chromosome sequence alignment for different strains of the same species, where more than 98% of letters in core regions are identical.Keywords: dynamic programming, algorithm, common substrings, string
Procedia PDF Downloads 277291 Remote Sensing Application on Snow Products and Analyzing Disaster-Forming Environments Xinjiang, China
Authors: Gulijianati Abake, Ryutaro Tateishi
Abstract:
Snow is one kind of special underlying surface, has high reflectivity, low thermal conductivity, and snow broth hydrological effect. Every year, frequent snow disaster in Xinjiang causing considerable economic loss and serious damage to towns and farms, such as livestock casualties, traffic jams and other disaster, therefore monitoring SWE (snow volume) in Xinjiang has a great significance. The problems of how this disaster distributes and what disaster-forming environments are important to its occurrence are the most pressing problems in disaster risk assessment and salvage material arrangement. The present study aims 1) to monitor accurate SWE using MODIS, AMSRE, and CMC data, 2) to establish the regularity of snow disaster outbreaks and the important disaster-forming environmental factors. And a spatial autocorrelation analysis method and a canonical correlation analysis method are used to answer these two questions separately, 3) to prepare the way to salvage material arrangements for snow disasters.Keywords: snow water equivalent (snow volume), AMSR-E, CMC snow depth, snow disaster
Procedia PDF Downloads 3737290 A Pilot Epidemiological Survey of Parasitic Problems of Goats in and Around Derawar Fort Area, Cholistan, Pakistan
Authors: Muhammad Tahir Riaz, Khalid Mehmood, Ahmad Waseem Akhtar, Tariq Abbas, Sadaqat Ali, Muhammad Altaf
Abstract:
Livestock sector contributes around 55.9 and 11.8% to agriculture and GDP respectively, according to economic survey of Pakistan 2013-2014. The goats population has been estimated about 66.6 million (M). Parasitic infestation is a major health problem in goats causing loss in body weight, poor body condition, low birth weights, and difficulty in kidding. Keeping in view the utilization of these animals in the country, a pilot epidemiological survey was conducted to find out the major parasitic problems of goats in and around Derawar fort area, Cholistan. Data regarding 662 fecal samples of goats was collected from 25 tobas of Cholistan during June 2012 to June 2013. All the fecal samples were examined through Direct Smear Method and Salt Flotation Technique for the presence of helminth eggs. External parasites were taken from the various components of the carcass of goat and were conserved in 70% alcohol in hygienic, properly enclosed glass jars that were tagged thoroughly. The collected date was analyzed statistically by Chi-square test to find out the prevalence in goats. Out of 662 goats, 261 (39.42%) were found positive for parasites. 233 (35.20%) goats were found positive for gastrointestinal parasites while 28 (4.23%) were positive for external parasites including ticks 20 (3.02%) and mange 8 (1.21%). The higher prevalence of parasites in the study area may be due to pasture grazing, poor management and lack of extension work. In this regards proper management and control measures should be adopted to minimize the Parasitic Problems.Keywords: Cholistan, goats, parasite, surveillance
Procedia PDF Downloads 5727289 Automation of Finite Element Simulations for the Design Space Exploration and Optimization of Type IV Pressure Vessel
Authors: Weili Jiang, Simon Cadavid Lopera, Klaus Drechsler
Abstract:
Fuel cell vehicle has become the most competitive solution for the transportation sector in the hydrogen economy. Type IV pressure vessel is currently the most popular and widely developed technology for the on-board storage, based on their high reliability and relatively low cost. Due to the stringent requirement on mechanical performance, the pressure vessel is subject to great amount of composite material, a major cost driver for the hydrogen tanks. Evidently, the optimization of composite layup design shows great potential in reducing the overall material usage, yet requires comprehensive understanding on underlying mechanisms as well as the influence of different design parameters on mechanical performance. Given the type of materials and manufacturing processes by which the type IV pressure vessels are manufactured, the design and optimization are a nuanced subject. The manifold of stacking sequence and fiber orientation variation possibilities have an out-standing effect on vessel strength due to the anisotropic property of carbon fiber composites, which make the design space high dimensional. Each variation of design parameters requires computational resources. Using finite element analysis to evaluate different designs is the most common method, however, the model-ing, setup and simulation process can be very time consuming and result in high computational cost. For this reason, it is necessary to build a reliable automation scheme to set up and analyze the di-verse composite layups. In this research, the simulation process of different tank designs regarding various parameters is conducted and automatized in a commercial finite element analysis framework Abaqus. Worth mentioning, the modeling of the composite overwrap is automatically generated using an Abaqus-Python scripting interface. The prediction of the winding angle of each layer and corresponding thickness variation on dome region is the most crucial step of the modeling, which is calculated and implemented using analytical methods. Subsequently, these different composites layups are simulated as axisymmetric models to facilitate the computational complexity and reduce the calculation time. Finally, the results are evaluated and compared regarding the ultimate tank strength. By automatically modeling, evaluating and comparing various composites layups, this system is applicable for the optimization of the tanks structures. As mentioned above, the mechanical property of the pressure vessel is highly dependent on composites layup, which requires big amount of simulations. Consequently, to automatize the simulation process gains a rapid way to compare the various designs and provide an indication of the optimum one. Moreover, this automation process can also be operated for creating a data bank of layups and corresponding mechanical properties with few preliminary configuration steps for the further case analysis. Subsequently, using e.g. machine learning to gather the optimum by the data pool directly without the simulation process.Keywords: type IV pressure vessels, carbon composites, finite element analy-sis, automation of simulation process
Procedia PDF Downloads 1397288 On the Accuracy of Basic Modal Displacement Method Considering Various Earthquakes
Authors: Seyed Sadegh Naseralavi, Sadegh Balaghi, Ehsan Khojastehfar
Abstract:
Time history seismic analysis is supposed to be the most accurate method to predict the seismic demand of structures. On the other hand, the required computational time of this method toward achieving the result is its main deficiency. While being applied in optimization process, in which the structure must be analyzed thousands of time, reducing the required computational time of seismic analysis of structures makes the optimization algorithms more practical. Apparently, the invented approximate methods produce some amount of errors in comparison with exact time history analysis but the recently proposed method namely, Complete Quadratic Combination (CQC) and Sum Root of the Sum of Squares (SRSS) drastically reduces the computational time by combination of peak responses in each mode. In the present research, the Basic Modal Displacement (BMD) method is introduced and applied towards estimation of seismic demand of main structure. Seismic demand of sampled structure is estimated by calculation of modal displacement of basic structure (in which the modal displacement has been calculated). Shear steel sampled structures are selected as case studies. The error applying the introduced method is calculated by comparison of the estimated seismic demands with exact time history dynamic analysis. The efficiency of the proposed method is demonstrated by application of three types of earthquakes (in view of time of peak ground acceleration).Keywords: time history dynamic analysis, basic modal displacement, earthquake-induced demands, shear steel structures
Procedia PDF Downloads 3607287 Translation and Validation of the Thai Version of the Japanese Sleep Questionnaire for Preschoolers
Authors: Natcha Lueangapapong, Chariya Chuthapisith, Lunliya Thampratankul
Abstract:
Background: There is a need to find an appropriate tool to help healthcare providers determine sleep problems in children for early diagnosis and management. The Japanese Sleep Questionnaire for Preschoolers (JSQ-P) is a parent-reported sleep questionnaire that has good psychometric properties and can be used in the context of Asian culture, which is likely suitable for Thai children. Objectives: This study aimed to translate and validate the Japanese Sleep Questionnaire for Preschoolers (JSQ-P) into a Thai version and to evaluate factors associated with sleep disorders in preschoolers. Methods: After approval by the original developer, the cross-cultural adaptation process of JSQ-P was performed, including forward translation, reconciliation, backward translation, and final approval of the Thai version of JSQ-P (TH-JSQ-P) by the original creator. This study was conducted between March 2021 and February 2022. The TH-JSQ-P was completed by 2,613 guardians whose children were aged 2-6 years twice in 10-14 days to assess its reliability and validity. Content validity was measured by an index of item-objective congruence (IOC) and a content validity index (CVI). Face validity, content validity, structural validity, construct validity (discriminant validity), criterion validity and predictive validity were assessed. The sensitivity and specificity of the TH-JSQ-P were also measured by using a total JSQ-P score cutoff point 84, recommended by the original JSQ-P and each subscale score among the clinical samples of obstructive sleep apnea syndrome. Results: Internal consistency reliability, evaluated by Cronbach’s α coefficient, showed acceptable reliability in all subscales of JSQ-P. It also had good test-retest reliability, as the intraclass correlation coefficient (ICC) for all items ranged between 0.42-0.84. The content validity was acceptable. For structural validity, our results indicated that the final factor solution for the Th-JSQ-P was comparable to the original JSQ-P. For construct validity, age group was one of the clinical parameters associated with some sleep problems. In detail, parasomnias, insomnia, daytime excessive sleepiness and sleep habits significantly decreased when the children got older; on the other hand, insufficient sleep was significantly increased with age. For criterion validity, all subscales showed a correlation with the Epworth Sleepiness Scale (r = -0.049-0.349). In predictive validity, the Epworth Sleepiness Scale was significantly a strong factor that correlated to sleep problems in all subscales of JSQ-P except in the subscale of sleep habit. The sensitivity and specificity of the total JSQ-P score were 0.72 and 0.66, respectively. Conclusion: The Thai version of JSQ-P has good internal consistency reliability and test-retest reliability. It passed 6 validity tests, and this can be used to evaluate sleep problems in preschool children in Thailand. Furthermore, it has satisfactory general psychometric properties and good reliability and validity. The data collected in examining the sensitivity of the Thai version revealed that the JSQ-P could detect differences in sleep problems among children with obstructive sleep apnea syndrome. This confirmed that the measure is sensitive and can be used to discriminate sleep problems among different children.Keywords: preschooler, questionnaire, validation, Thai version
Procedia PDF Downloads 1107286 Application of Observational Medical Outcomes Partnership-Common Data Model (OMOP-CDM) Database in Nursing Health Problems with Prostate Cancer-a Pilot Study
Authors: Hung Lin-Zin, Lai Mei-Yen
Abstract:
Prostate cancer is the most commonly diagnosed male cancer in the U.S. The prevalence is around 1 in 8. The etiology of prostate cancer is still unknown, but some predisposing factors, such as age, black race, family history, and obesity, may increase the risk of the disease. In 2020, a total of 7,178 Taiwanese people were nearly diagnosed with prostate cancer, accounting for 5.88% of all cancer cases, and the incidence rate ranked fifth among men. In that year, the total number of deaths from prostate cancer was 1,730, accounting for 3.45% of all cancer deaths, and the death rate ranked 6th among men, accounting for 94.34% of the cases of male reproductive organs. Looking for domestic and foreign literature on the use of OMOP (Observational Medical Outcomes Partnership, hereinafter referred to as OMOP) database analysis, there are currently nearly a hundred literature published related to nursing-related health problems and nursing measures built in the OMOP general data model database of medical institutions are extremely rare. The OMOP common data model construction analysis platform is a system developed by the FDA in 2007, using a common data model (common data model, CDM) to analyze and monitor healthcare data. It is important to build up relevant nursing information from the OMOP- CDM database to assist our daily practice. Therefore, we choose prostate cancer patients who are our popular care objects and use the OMOP- CDM database to explore the common associated health problems. With the assistance of OMOP-CDM database analysis, we can expect early diagnosis and prevention of prostate cancer patients' comorbidities to improve patient care.Keywords: OMOP, nursing diagnosis, health problem, prostate cancer
Procedia PDF Downloads 76