Search results for: elliptic curve digital signature algorithm
4600 Optimization of Solar Rankine Cycle by Exergy Analysis and Genetic Algorithm
Authors: R. Akbari, M. A. Ehyaei, R. Shahi Shavvon
Abstract:
Nowadays, solar energy is used for energy purposes such as the use of thermal energy for domestic, industrial and power applications, as well as the conversion of the sunlight into electricity by photovoltaic cells. In this study, the thermodynamic simulation of the solar Rankin cycle with phase change material (paraffin) was first studied. Then energy and exergy analyses were performed. For optimization, a single and multi-objective genetic optimization algorithm to maximize thermal and exergy efficiency was used. The parameters discussed in this paper included the effects of input pressure on turbines, input mass flow to turbines, the surface of converters and collector angles on thermal and exergy efficiency. In the organic Rankin cycle, where solar energy is used as input energy, the fluid selection is considered as a necessary factor to achieve reliable and efficient operation. Therefore, silicon oil is selected for a high-temperature cycle and water for a low-temperature cycle as an operating fluid. The results showed that increasing the mass flow to turbines 1 and 2 would increase thermal efficiency, while it reduces and increases the exergy efficiency in turbines 1 and 2, respectively. Increasing the inlet pressure to the turbine 1 decreases the thermal and exergy efficiency, and increasing the inlet pressure to the turbine 2 increases the thermal efficiency and exergy efficiency. Also, increasing the angle of the collector increased thermal efficiency and exergy. The thermal efficiency of the system was 22.3% which improves to 33.2 and 27.2% in single-objective and multi-objective optimization, respectively. Also, the exergy efficiency of the system was 1.33% which has been improved to 1.719 and 1.529% in single-objective and multi-objective optimization, respectively. These results showed that the thermal and exergy efficiency in a single-objective optimization is greater than the multi-objective optimization.Keywords: exergy analysis, genetic algorithm, rankine cycle, single and multi-objective function
Procedia PDF Downloads 1474599 Machine Vision System for Measuring the Quality of Bulk Sun-dried Organic Raisins
Authors: Navab Karimi, Tohid Alizadeh
Abstract:
An intelligent vision-based system was designed to measure the quality and purity of raisins. A machine vision setup was utilized to capture the images of bulk raisins in ranges of 5-50% mixed pure-impure berries. The textural features of bulk raisins were extracted using Grey-level Histograms, Co-occurrence Matrix, and Local Binary Pattern (a total of 108 features). Genetic Algorithm and neural network regression were used for selecting and ranking the best features (21 features). As a result, the GLCM features set was found to have the highest accuracy (92.4%) among the other sets. Followingly, multiple feature combinations of the previous stage were fed into the second regression (linear regression) to increase accuracy, wherein a combination of 16 features was found to be the optimum. Finally, a Support Vector Machine (SVM) classifier was used to differentiate the mixtures, producing the best efficiency and accuracy of 96.2% and 97.35%, respectively.Keywords: sun-dried organic raisin, genetic algorithm, feature extraction, ann regression, linear regression, support vector machine, south azerbaijan.
Procedia PDF Downloads 734598 Modeling the Effects of Leachate-Impacted Groundwater on the Water Quality of a Large Tidal River
Authors: Emery Coppola Jr., Marwan Sadat, Il Kim, Diane Trube, Richard Kurisko
Abstract:
Contamination sites like landfills often pose significant risks to receptors like surface water bodies. Surface water bodies are often a source of recreation, including fishing and swimming, which not only enhances their value but also serves as a direct exposure pathway to humans, increasing their need for protection from water quality degradation. In this paper, a case study presents the potential effects of leachate-impacted groundwater from a large closed sanitary landfill on the surface water quality of the nearby Raritan River, situated in New Jersey. The study, performed over a two year period, included in-depth field evaluation of both the groundwater and surface water systems, and was supplemented by computer modeling. The analysis required delineation of a representative average daily groundwater discharge from the Landfill shoreline into the large, highly tidal Raritan River, with a corresponding estimate of daily mass loading of potential contaminants of concern. The average daily groundwater discharge into the river was estimated from a high-resolution water level study and a 24-hour constant-rate aquifer pumping test. The significant tidal effects induced on groundwater levels during the aquifer pumping test were filtered out using an advanced algorithm, from which aquifer parameter values were estimated using conventional curve match techniques. The estimated hydraulic conductivity values obtained from individual observation wells closely agree with tidally-derived values for the same wells. Numerous models were developed and used to simulate groundwater contaminant transport and surface water quality impacts. MODFLOW with MT3DMS was used to simulate the transport of potential contaminants of concern from the down-gradient edge of the Landfill to the Raritan River shoreline. A surface water dispersion model based upon a bathymetric and flow study of the river was used to simulate the contaminant concentrations over space within the river. The modeling results helped demonstrate that because of natural attenuation, the Landfill does not have a measurable impact on the river, which was confirmed by an extensive surface water quality study.Keywords: groundwater flow and contaminant transport modeling, groundwater/surface water interaction, landfill leachate, surface water quality modeling
Procedia PDF Downloads 2624597 Artificial Bee Colony Based Modified Energy Efficient Predictive Routing in MANET
Authors: Akhil Dubey, Rajnesh Singh
Abstract:
In modern days there occur many rapid modifications in field of ad hoc network. These modifications create many revolutionary changes in the routing. Predictive energy efficient routing is inspired on the bee’s behavior of swarm intelligence. Predictive routing improves the efficiency of routing in the energetic point of view. The main aim of this routing is the minimum energy consumption during communication and maximized intermediate node’s remaining battery power. This routing is based on food searching behavior of bees. There are two types of bees for the exploration phase the scout bees and for the evolution phase forager bees use by this routing. This routing algorithm computes the energy consumption, fitness ratio and goodness of the path. In this paper we review the literature related with predictive routing, presenting modified routing and simulation result of this algorithm comparison with artificial bee colony based routing schemes in MANET and see the results of path fitness and probability of fitness.Keywords: mobile ad hoc network, artificial bee colony, PEEBR, modified predictive routing
Procedia PDF Downloads 4164596 Multidirectional Product Support System for Decision Making in Textile Industry Using Collaborative Filtering Methods
Authors: A. Senthil Kumar, V. Murali Bhaskaran
Abstract:
In the information technology ground, people are using various tools and software for their official use and personal reasons. Nowadays, people are worrying to choose data accessing and extraction tools at the time of buying and selling their products. In addition, worry about various quality factors such as price, durability, color, size, and availability of the product. The main purpose of the research study is to find solutions to these unsolved existing problems. The proposed algorithm is a Multidirectional Rank Prediction (MDRP) decision making algorithm in order to take an effective strategic decision at all the levels of data extraction, uses a real time textile dataset and analyzes the results. Finally, the results are obtained and compared with the existing measurement methods such as PCC, SLCF, and VSS. The result accuracy is higher than the existing rank prediction methods.Keywords: Knowledge Discovery in Database (KDD), Multidirectional Rank Prediction (MDRP), Pearson’s Correlation Coefficient (PCC), VSS (Vector Space Similarity)
Procedia PDF Downloads 2874595 Surgical Hip Dislocation of Femoroacetabular Impingement: Survivorship and Functional Outcomes at 10 Years
Authors: L. Hoade, O. O. Onafowokan, K. Anderson, G. E. Bartlett, E. D. Fern, M. R. Norton, R. G. Middleton
Abstract:
Aims: Femoroacetabular impingement (FAI) was first recognised as a potential driver for hip pain at the turn of the last millennium. While there is an increasing trend towards surgical management of FAI by arthroscopic means, open surgical hip dislocation and debridement (SHD) remains the Gold Standard of care in terms of reported outcome measures. (1) Long-term functional and survivorship outcomes of SHD as a treatment for FAI are yet to be sufficiently reported in the literature. This study sets out to help address this imbalance. Methods: We undertook a retrospective review of our institutional database for all patients who underwent SHD for FAI between January 2003 and December 2008. A total of 223 patients (241 hips) were identified and underwent a ten year review with a standardised radiograph and patient-reported outcome measures questionnaire. The primary outcome measure of interest was survivorship, defined as progression to total hip arthroplasty (THA). Negative predictive factors were analysed. Secondary outcome measures of interest were survivorship to further (non-arthroplasty) surgery, functional outcomes as reflected by patient reported outcome measure scores (PROMS) scores, and whether a learning curve could be identified. Results: The final cohort consisted of 131 females and 110 males, with a mean age of 34 years. There was an overall native hip joint survival rate of 85.4% at ten years. Those who underwent a THA were significantly older at initial surgery, had radiographic evidence of preoperative osteoarthritis and pre- and post-operative acetabular undercoverage. In those whom had not progressed to THA, the average Non-arthritic Hip Score and Oxford Hip Score at ten year follow-up were 72.3% and 36/48, respectively, and 84% still deemed their surgery worthwhile. A learning curve was found to exist that was predicated on case selection rather than surgical technique. Conclusion: This is only the second study to evaluate the long-term outcomes (beyond ten years) of SHD for FAI and the first outside the originating centre. Our results suggest that, with correct patient selection, this remains an operation with worthwhile outcomes at ten years. How the results of open surgery compared to those of arthroscopy remains to be answered. While these results precede the advent of collison software modelling tools, this data helps set a benchmark for future comparison of other techniques effectiveness at the ten year mark.Keywords: femoroacetabular impingement, hip pain, surgical hip dislocation, hip debridement
Procedia PDF Downloads 844594 Bitplanes Gray-Level Image Encryption Approach Using Arnold Transform
Authors: Ali Abdrhman M. Ukasha
Abstract:
Data security needed in data transmission, storage, and communication to ensure the security. The single step parallel contour extraction (SSPCE) method is used to create the edge map as a key image from the different Gray level/Binary image. Performing the X-OR operation between the key image and each bit plane of the original image for image pixel values change purpose. The Arnold transform used to changes the locations of image pixels as image scrambling process. Experiments have demonstrated that proposed algorithm can fully encrypt 2D Gary level image and completely reconstructed without any distortion. Also shown that the analyzed algorithm have extremely large security against some attacks like salt & pepper and JPEG compression. Its proof that the Gray level image can be protected with a higher security level. The presented method has easy hardware implementation and suitable for multimedia protection in real time applications such as wireless networks and mobile phone services.Keywords: SSPCE method, image compression-salt- peppers attacks, bitplanes decomposition, Arnold transform, lossless image encryption
Procedia PDF Downloads 4384593 Advancing Phenological Understanding of Plants/Trees Through Phenocam Digital Time-lapse Images
Authors: Siddhartha Khare, Suyash Khare
Abstract:
Phenology, a crucial discipline in ecology, offers insights into the seasonal dynamics of organisms within natural ecosystems and the underlying environmental triggers. Leveraging the potent capabilities of digital repeat photography, PhenoCams capture invaluable data on the phenology of crops, plants, and trees. These cameras yield digital imagery in Red Green Blue (RGB) color channels, and some advanced systems even incorporate Near Infrared (NIR) bands. This study presents compelling case studies employing PhenoCam technology to unravel the phenology of black spruce trees. Through the analysis of RGB color channels, a range of essential color metrics including red chromatic coordinate (RCC), green chromatic coordinate (GCC), blue chromatic coordinate (BCC), vegetation contrast index (VCI), and excess green index (ExGI) are derived. These metrics illuminate variations in canopy color across seasons, shedding light on bud and leaf development. This, in turn, facilitates a deeper understanding of phenological events and aids in delineating the growth periods of trees and plants. The initial phase of this study addresses critical questions surrounding the fidelity of continuous canopy greenness records in representing bud developmental phases. Additionally, it discerns which color-based index most accurately tracks the seasonal variations in tree phenology within evergreen forest ecosystems. The subsequent section of this study delves into the transition dates of black spruce (Picea mariana (Mill.) B.S.P.) phenology. This is achieved through a fortnightly comparative analysis of the MODIS normalized difference vegetation index (NDVI) and the enhanced vegetation index (EVI). By employing PhenoCam technology and leveraging advanced color metrics, this study significantly advances our comprehension of black spruce tree phenology, offering valuable insights for ecological research and management.Keywords: phenology, remote sensing, phenocam, color metrics, NDVI, GCC
Procedia PDF Downloads 604592 Application of Causal Inference and Discovery in Curriculum Evaluation and Continuous Improvement
Authors: Lunliang Zhong, Bin Duan
Abstract:
The undergraduate graduation project is a vital part of the higher education curriculum, crucial for engineering accreditation. Current evaluations often summarize data without identifying underlying issues. This study applies the Peter-Clark algorithm to analyze causal relationships within the graduation project data of an Electronics and Information Engineering program, creating a causal model. Structural equation modeling confirmed the model's validity. The analysis reveals key teaching stages affecting project success, uncovering problems in the process. Introducing causal discovery and inference into project evaluation helps identify issues and propose targeted improvement measures. The effectiveness of these measures is validated by comparing the learning outcomes of two student cohorts, stratified by confounding factors, leading to improved teaching quality.Keywords: causal discovery, causal inference, continuous improvement, Peter-Clark algorithm, structural equation modeling
Procedia PDF Downloads 184591 Margin-Based Feed-Forward Neural Network Classifiers
Authors: Xiaohan Bookman, Xiaoyan Zhu
Abstract:
Margin-Based Principle has been proposed for a long time, it has been proved that this principle could reduce the structural risk and improve the performance in both theoretical and practical aspects. Meanwhile, feed-forward neural network is a traditional classifier, which is very hot at present with a deeper architecture. However, the training algorithm of feed-forward neural network is developed and generated from Widrow-Hoff Principle that means to minimize the squared error. In this paper, we propose a new training algorithm for feed-forward neural networks based on Margin-Based Principle, which could effectively promote the accuracy and generalization ability of neural network classifiers with less labeled samples and flexible network. We have conducted experiments on four UCI open data sets and achieved good results as expected. In conclusion, our model could handle more sparse labeled and more high-dimension data set in a high accuracy while modification from old ANN method to our method is easy and almost free of work.Keywords: Max-Margin Principle, Feed-Forward Neural Network, classifier, structural risk
Procedia PDF Downloads 3424590 Sensor and Actuator Fault Detection in Connected Vehicles under a Packet Dropping Network
Authors: Z. Abdollahi Biron, P. Pisu
Abstract:
Connected vehicles are one of the promising technologies for future Intelligent Transportation Systems (ITS). A connected vehicle system is essentially a set of vehicles communicating through a network to exchange their information with each other and the infrastructure. Although this interconnection of the vehicles can be potentially beneficial in creating an efficient, sustainable, and green transportation system, a set of safety and reliability challenges come out with this technology. The first challenge arises from the information loss due to unreliable communication network which affects the control/management system of the individual vehicles and the overall system. Such scenario may lead to degraded or even unsafe operation which could be potentially catastrophic. Secondly, faulty sensors and actuators can affect the individual vehicle’s safe operation and in turn will create a potentially unsafe node in the vehicular network. Further, sending that faulty sensor information to other vehicles and failure in actuators may significantly affect the safe operation of the overall vehicular network. Therefore, it is of utmost importance to take these issues into consideration while designing the control/management algorithms of the individual vehicles as a part of connected vehicle system. In this paper, we consider a connected vehicle system under Co-operative Adaptive Cruise Control (CACC) and propose a fault diagnosis scheme that deals with these aforementioned challenges. Specifically, the conventional CACC algorithm is modified by adding a Kalman filter-based estimation algorithm to suppress the effect of lost information under unreliable network. Further, a sliding mode observer-based algorithm is used to improve the sensor reliability under faults. The effectiveness of the overall diagnostic scheme is verified via simulation studies.Keywords: fault diagnostics, communication network, connected vehicles, packet drop out, platoon
Procedia PDF Downloads 2394589 Predicting Daily Patient Hospital Visits Using Machine Learning
Authors: Shreya Goyal
Abstract:
The study aims to build user-friendly software to understand patient arrival patterns and compute the number of potential patients who will visit a particular health facility for a given period by using a machine learning algorithm. The underlying machine learning algorithm used in this study is the Support Vector Machine (SVM). Accurate prediction of patient arrival allows hospitals to operate more effectively, providing timely and efficient care while optimizing resources and improving patient experience. It allows for better allocation of staff, equipment, and other resources. If there's a projected surge in patients, additional staff or resources can be allocated to handle the influx, preventing bottlenecks or delays in care. Understanding patient arrival patterns can also help streamline processes to minimize waiting times for patients and ensure timely access to care for patients in need. Another big advantage of using this software is adhering to strict data protection regulations such as the Health Insurance Portability and Accountability Act (HIPAA) in the United States as the hospital will not have to share the data with any third party or upload it to the cloud because the software can read data locally from the machine. The data needs to be arranged in. a particular format and the software will be able to read the data and provide meaningful output. Using software that operates locally can facilitate compliance with these regulations by minimizing data exposure. Keeping patient data within the hospital's local systems reduces the risk of unauthorized access or breaches associated with transmitting data over networks or storing it in external servers. This can help maintain the confidentiality and integrity of sensitive patient information. Historical patient data is used in this study. The input variables used to train the model include patient age, time of day, day of the week, seasonal variations, and local events. The algorithm uses a Supervised learning method to optimize the objective function and find the global minima. The algorithm stores the values of the local minima after each iteration and at the end compares all the local minima to find the global minima. The strength of this study is the transfer function used to calculate the number of patients. The model has an output accuracy of >95%. The method proposed in this study could be used for better management planning of personnel and medical resources.Keywords: machine learning, SVM, HIPAA, data
Procedia PDF Downloads 654588 Complex Network Approach to International Trade of Fossil Fuel
Authors: Semanur Soyyigit Kaya, Ercan Eren
Abstract:
Energy has a prominent role for development of nations. Countries which have energy resources also have strategic power in the international trade of energy since it is essential for all stages of production in the economy. Thus, it is important for countries to analyze the weakness and strength of the system. On the other side, it is commonly believed that international trade has complex network properties. Complex network is a tool for the analysis of complex systems with heterogeneous agents and interaction between them. A complex network consists of nodes and the interactions between these nodes. Total properties which emerge as a result of these interactions are distinct from the sum of small parts (more or less) in complex systems. Thus, standard approaches to international trade are superficial to analyze these systems. Network analysis provides a new approach to analyze international trade as a network. In this network countries constitute nodes and trade relations (export or import) constitute edges. It becomes possible to analyze international trade network in terms of high degree indicators which are specific to complex systems such as connectivity, clustering, assortativity/disassortativity, centrality, etc. In this analysis, international trade of crude oil and coal which are types of fossil fuel has been analyzed from 2005 to 2014 via network analysis. First, it has been analyzed in terms of some topological parameters such as density, transitivity, clustering etc. Afterwards, fitness to Pareto distribution has been analyzed. Finally, weighted HITS algorithm has been applied to the data as a centrality measure to determine the real prominence of countries in these trade networks. Weighted HITS algorithm is a strong tool to analyze the network by ranking countries with regards to prominence of their trade partners. We have calculated both an export centrality and an import centrality by applying w-HITS algorithm to data.Keywords: complex network approach, fossil fuel, international trade, network theory
Procedia PDF Downloads 3364587 International Broadcasting of Public Diplomacy in the Era of Social Media in Nigeria
Authors: Henry Okechukwu Onyeiwu
Abstract:
In today’s Nigerian digital age, the landscape of public diplomacy has been significantly altered by the rise of social media platforms like YouTube, Facebook, Twitter, and Instagram. In recent years, social media platforms have emerged as powerful tools for public diplomacy, transforming how countries communicate with both domestic and global audiences. International broadcasting as a tool of public diplomacy has undergone a significant transformation. Traditional methods of state-run media and controlled broadcasting have evolved to incorporate the dynamic, interactive, and decentralized nature of digital platforms. Understanding how Nigerian governments engages in international broadcasting of public diplomacy, the influence of social media on broadcasting public diplomacy, focusing on the advantages and disadvantages of controlling media outlets for diplomatic purposes and also covers the changing nature of global communication in this digital era. As countries navigate the complexities of international relations, the effectiveness of controlled media in shaping public perception and engagement raises significant questions worth exploring. The vast amount of content available can make it challenging to capture and retain audience attention. The ease of spreading false information on social media requires international broadcasters to maintain credibility and counteract misleading narratives. Addressing these challenges requires a comprehensive research that integrates digital communication tools, cultural sensitivity, cybersecurity measures and ongoing evaluation to enhance Nigeria’s international broadcasting of public diplomacy. This study employed a mixed-methods approach, combining qualitative and quantitative research methods. A content analysis of Nigeria’s international broadcasting content was conducted to assess its themes, narratives, and engagement strategies. Additionally, surveys and interviews with communications professionals, diplomats, and social media users were carried out to gather insights on perceptions and effectiveness of public diplomacy initiatives. It has highlighted some of the present trends in technology and the international environmental in which public diplomacy must work, and show how the past can illuminate the road for those navigating this new world. The rise of the social network creates more opportunities than it closes for public diplomacy. This evolution highlights the increasing importance of engagement, mutual understanding, and cooperation in international relations. By Adopting a more inclusive and participatory approach, public diplomacy can more effectively address global challenges and build stronger, more resilient relationships between nations. As Nigeria navigates the complexities of its international relations, this abstract will provide a vital examination of how it can better utilize the dual platforms of international broadcasting and social media in its public diplomacy efforts. The outcome will bear significance not only for Nigeria but also for other nations grappling with similar challenges in the digital age. As social media continues to play a crucial role in public diplomacy, understanding the dynamics of controlled media outlets becomes ever more critical. This abstract shed light on the advantages and disadvantages of such control, ultimately contributing valuable insights to practitioners in the field of diplomacy as they adapt to the rapidly changing communication landscape.Keywords: international broadcasting, public diplomacy, social media, international relation, polities
Procedia PDF Downloads 314586 A Wireless Feedback Control System as a Base of Bio-Inspired Structure System to Mitigate Vibration in Structures
Authors: Gwanghee Heo, Geonhyeok Bang, Chunggil Kim, Chinok Lee
Abstract:
This paper attempts to develop a wireless feedback control system as a primary step eventually toward a bio-inspired structure system where inanimate structure behaves like a life form autonomously. It is a standalone wireless control system which is supposed to measure externally caused structural responses, analyze structural state from acquired data, and take its own action on the basis of the analysis with an embedded logic. For an experimental examination of its effectiveness, we applied it on a model of two-span bridge and performed a wireless control test. Experimental tests have been conducted for comparison on both the wireless and the wired system under the conditions of Un-control, Passive-off, Passive-on, and Lyapunov control algorithm. By proving the congruence of the test result of the wireless feedback control system with the wired control system, its control performance was proven to be effective. Besides, it was found to be economical in energy consumption and also autonomous by means of a command algorithm embedded into it, which proves its basic capacity as a bio-inspired system.Keywords: structural vibration control, wireless system, MR damper, feedback control, embedded system
Procedia PDF Downloads 2114585 Detectability of Malfunction in Turboprop Engine
Authors: Tomas Vampola, Michael Valášek
Abstract:
On the basis of simulation-generated failure states of structural elements of a turboprop engine suitable for the busy-jet class of aircraft, an algorithm for early prediction of damage or reduction in functionality of structural elements of the engine is designed and verified with real data obtained at dynamometric testing facilities of aircraft engines. Based on an expanding database of experimentally determined data from temperature and pressure sensors during the operation of turboprop engines, this strategy is constantly modified with the aim of using the minimum number of sensors to detect an inadmissible or deteriorated operating mode of specific structural elements of an aircraft engine. The assembled algorithm for the early prediction of reduced functionality of the aircraft engine significantly contributes to the safety of air traffic and to a large extent, contributes to the economy of operation with positive effects on the reduction of the energy demand of operation and the elimination of adverse effects on the environment.Keywords: detectability of malfunction, dynamometric testing, prediction of damage, turboprop engine
Procedia PDF Downloads 944584 Experimental and Theoretical Approach, Hirshfeld Surface, Reduced Density Gradient, Molecular Docking of a Thiourea Derivative
Authors: Noureddine Benharkat, Abdelkader Chouaih, Nourdine Boukabcha
Abstract:
A thiourea derivative compound was synthesized and subjected to structural analysis using single-crystal X-ray diffraction (XRD). The crystallographic data unveiled its crystallization in the P21/c space group within the monoclinic system. Examination of the dihedral angles indicated a notable non-planar structure. To support and interpret these resulats, density functional theory (DFT) calculations were conducted utilizing the B3LYP functional along with a 6–311 G (d, p) basis set. Additionally, to assess the contribution of intermolecular interactions, Hirshfeld surface analysis and 2D fingerprint plots were employed. Various types of interactions, whether weak intramolecular or intermolecular, within a molecule can significantly impact its stability. The distinctive signature of non-covalent interactions can be detected solely through electron density analysis. The NCI-RDG analysis was employed to investigate both repulsive and attractive van der Waals interactions while also calculating the energies associated with intermolecular interactions and their characteristics. Additionally, a molecular docking study was studied to explain the structure-activity relationship, revealing that the title compound exhibited an affinity energy of -6.8 kcal/mol when docked with B-DNA (1BNA).Keywords: computational chemistry, density functional theory, crystallography, molecular docking, molecular structure, powder x-ray diffraction, single crystal x-ray diffraction
Procedia PDF Downloads 604583 A Location-Allocation-Routing Model for a Home Health Care Supply Chain Problem
Authors: Amir Mohammad Fathollahi Fard, Mostafa Hajiaghaei-Keshteli, Mohammad Mahdi Paydar
Abstract:
With increasing life expectancy in developed countries, the role of home care services is highlighted by both academia and industrial contributors in Home Health Care Supply Chain (HHCSC) companies. The main decisions in such supply chain systems are the location of pharmacies, the allocation of patients to these pharmacies and also the routing and scheduling decisions of nurses to visit their patients. In this study, for the first time, an integrated model is proposed to consist of all preliminary and necessary decisions in these companies, namely, location-allocation-routing model. This model is a type of NP-hard one. Therefore, an Imperialist Competitive Algorithm (ICA) is utilized to solve the model, especially in large sizes. Results confirm the efficiency of the developed model for HHCSC companies as well as the performance of employed ICA.Keywords: home health care supply chain, location-allocation-routing problem, imperialist competitive algorithm, optimization
Procedia PDF Downloads 3974582 Linear Semi Active Controller of Magneto-Rheological Damper for Seismic Vibration Attenuation
Authors: Zizouni Khaled, Fali Leyla, Sadek Younes, Bousserhane Ismail Khalil
Abstract:
In structural vibration caused principally by an earthquake excitation, the most vibration’s attenuation system used recently is the semi active control with a Magneto Rheological Damper device. This control was a subject of many researches and works in the last years. The big challenges of searchers in this case is to propose an adequate controller with a robust algorithm of current or tension adjustment. In this present paper, a linear controller is proposed to control the MR damper using to reduce a vibrations of three story structure exposed to El Centro’s 1940 and Boumerdès 2003 earthquakes. In this example, the MR damper is installed in the first floor of the structure. The numerical simulations results of the proposed linear control with a feedback law based on clipped optimal algorithm showed the feasibility of the semi active control to protecting civil structures. The comparison of the controlled structure and uncontrolled structures responses illustrate clearly the performance and the effectiveness of the simple proposed approach.Keywords: MR damper, seismic vibration, semi-active control
Procedia PDF Downloads 2854581 A Subband BSS Structure with Reduced Complexity and Fast Convergence
Authors: Salah Al-Din I. Badran, Samad Ahmadi, Ismail Shahin
Abstract:
A blind source separation method is proposed; in this method, we use a non-uniform filter bank and a novel normalisation. This method provides a reduced computational complexity and increased convergence speed comparing to the full-band algorithm. Recently, adaptive sub-band scheme has been recommended to solve two problems: reduction of computational complexity and increase the convergence speed of the adaptive algorithm for correlated input signals. In this work, the reduction in computational complexity is achieved with the use of adaptive filters of orders less than the full-band adaptive filters, which operate at a sampling rate lower than the sampling rate of the input signal. The decomposed signals by analysis bank filter are less correlated in each subband than the input signal at full bandwidth, and can promote better rates of convergence.Keywords: blind source separation, computational complexity, subband, convergence speed, mixture
Procedia PDF Downloads 5804580 Bitplanes Image Encryption/Decryption Using Edge Map (SSPCE Method) and Arnold Transform
Authors: Ali A. Ukasha
Abstract:
Data security needed in data transmission, storage, and communication to ensure the security. The single step parallel contour extraction (SSPCE) method is used to create the edge map as a key image from the different Gray level/Binary image. Performing the X-OR operation between the key image and each bit plane of the original image for image pixel values change purpose. The Arnold transform used to changes the locations of image pixels as image scrambling process. Experiments have demonstrated that proposed algorithm can fully encrypt 2D Gary level image and completely reconstructed without any distortion. Also shown that the analyzed algorithm have extremely large security against some attacks like salt & pepper and JPEG compression. Its proof that the Gray level image can be protected with a higher security level. The presented method has easy hardware implementation and suitable for multimedia protection in real time applications such as wireless networks and mobile phone services.Keywords: SSPCE method, image compression, salt and peppers attacks, bitplanes decomposition, Arnold transform, lossless image encryption
Procedia PDF Downloads 4994579 Investigation of Soil Slopes Stability
Authors: Nima Farshidfar, Navid Daryasafar
Abstract:
In this paper, the seismic stability of reinforced soil slopes is studied using pseudo-dynamic analysis. Equilibrium equations that are applicable to the every kind of failure surface are written using Horizontal Slices Method. In written equations, the balance of the vertical and horizontal forces and moment equilibrium is fully satisfied. Failure surface is assumed to be log-spiral, and non-linear equilibrium equations obtained for the system are solved using Newton-Raphson Method. Earthquake effects are applied as horizontal and vertical pseudo-static coefficients to the problem. To solve this problem, a code was developed in MATLAB, and the critical failure surface is calculated using genetic algorithm. At the end, comparing the results obtained in this paper, effects of various parameters and the effect of using pseudo - dynamic analysis in seismic forces modeling is presented.Keywords: soil slopes, pseudo-dynamic, genetic algorithm, optimization, limit equilibrium method, log-spiral failure surface
Procedia PDF Downloads 3394578 Multilabel Classification with Neural Network Ensemble Method
Authors: Sezin Ekşioğlu
Abstract:
Multilabel classification has a huge importance for several applications, it is also a challenging research topic. It is a kind of supervised learning that contains binary targets. The distance between multilabel and binary classification is having more than one class in multilabel classification problems. Features can belong to one class or many classes. There exists a wide range of applications for multi label prediction such as image labeling, text categorization, gene functionality. Even though features are classified in many classes, they may not always be properly classified. There are many ensemble methods for the classification. However, most of the researchers have been concerned about better multilabel methods. Especially little ones focus on both efficiency of classifiers and pairwise relationships at the same time in order to implement better multilabel classification. In this paper, we worked on modified ensemble methods by getting benefit from k-Nearest Neighbors and neural network structure to address issues within a beneficial way and to get better impacts from the multilabel classification. Publicly available datasets (yeast, emotion, scene and birds) are performed to demonstrate the developed algorithm efficiency and the technique is measured by accuracy, F1 score and hamming loss metrics. Our algorithm boosts benchmarks for each datasets with different metrics.Keywords: multilabel, classification, neural network, KNN
Procedia PDF Downloads 1554577 Chinese “Wolf Warrior” Diplomacy And Foreign Public Opinion
Authors: Chaohong Pan
Abstract:
Through public diplomacy on social media, governments have attempted to influence foreign public opinion. What is the impact of digital public diplomacy? Public diplomacy research often relies on content analysis to study the strategies employed by communicators but has rarely examined its actual impact on the audience. In addition, we do not know if giving a communicator an explicit label, as Twitter does with “government account”, would change the effects of the messages. Can the government label reduce the percussiveness of public diplomacy messages by sending a warning signal? Using a 2 × 2 survey experiment, the present paper contributes to the study of public diplomacy by randomly exposing American participants to four types of tweets from Chinese diplomats. The stimulus materials vary in terms of the tweets’ content (“positive-china” vs. “negative-US) and Twitter government labels (with vs. without the labels). I found that positive tweets about China have a significant positive effect on Americans’ attitudes toward China, whereas negative tweets about the US have little effect on their opinions. Furthermore, positive-China tweets are effective only on China-related issues, which indicates that Chinese diplomats’ tweets have limited effects on shaping a foreign audience’s attitudes toward their own country. Lastly, I find that labels largely have no impact on a diplomatic tweet’s effect. These results contribute to our understanding of the effects of public diplomacy in the digital age.Keywords: public diplomacy, china, foreign public opinion, twitter
Procedia PDF Downloads 1924576 Using Artificial Vision Techniques for Dust Detection on Photovoltaic Panels
Authors: Gustavo Funes, Eduardo Peters, Jose Delpiano
Abstract:
It is widely known that photovoltaic technology has been massively distributed over the last decade despite its low-efficiency ratio. Dust deposition reduces this efficiency even more, lowering the energy production and module lifespan. In this work, we developed an artificial vision algorithm based on CIELAB color space to identify dust over panels in an autonomous way. We performed several experiments photographing three different types of panels, 30W, 340W and 410W. Those panels were soiled artificially with uniform and non-uniform distributed dust. The algorithm proposed uses statistical tools to provide a simulation with a 100% soiled panel and then performs a comparison to get the percentage of dirt in the experimental data set. The simulation uses a seed that is obtained by taking a dust sample from the maximum amount of dust from the dataset. The final result is the dirt percentage and the possible distribution of dust over the panel. Dust deposition is a key factor for plant owners to determine cleaning cycles or identify nonuniform depositions that could lead to module failure and hot spots.Keywords: dust detection, photovoltaic, artificial vision, soiling
Procedia PDF Downloads 504575 Method for Requirements Analysis and Decision Making for Restructuring Projects in Factories
Authors: Rene Hellmuth
Abstract:
The requirements for the factory planning and the building concerned have changed in the last years. Factory planning has the task of designing products, plants, processes, organization, areas, and the building of a factory. Regular restructuring gains more importance in order to maintain the competitiveness of a factory. Restrictions regarding new areas, shorter life cycles of product and production technology as well as a VUCA (volatility, uncertainty, complexity and ambiguity) world cause more frequently occurring rebuilding measures within a factory. Restructuring of factories is the most common planning case today. Restructuring is more common than new construction, revitalization and dismantling of factories. The increasing importance of restructuring processes shows that the ability to change was and is a promising concept for the reaction of companies to permanently changing conditions. The factory building is the basis for most changes within a factory. If an adaptation of a construction project (factory) is necessary, the inventory documents must be checked and often time-consuming planning of the adaptation must take place to define the relevant components to be adapted, in order to be able to finally evaluate them. The different requirements of the planning participants from the disciplines of factory planning (production planner, logistics planner, automation planner) and industrial construction planning (architect, civil engineer) come together during reconstruction and must be structured. This raises the research question: Which requirements do the disciplines involved in the reconstruction planning place on a digital factory model? A subordinate research question is: How can model-based decision support be provided for a more efficient design of the conversion within a factory? Because of the high adaptation rate of factories and its building described above, a methodology for rescheduling factories based on the requirements engineering method from software development is conceived and designed for practical application in factory restructuring projects. The explorative research procedure according to Kubicek is applied. Explorative research is suitable if the practical usability of the research results has priority. Furthermore, it will be shown how to best use a digital factory model in practice. The focus will be on mobile applications to meet the needs of factory planners on site. An augmented reality (AR) application will be designed and created to provide decision support for planning variants. The aim is to contribute to a shortening of the planning process and model-based decision support for more efficient change management. This requires the application of a methodology that reduces the deficits of the existing approaches. The time and cost expenditure are represented in the AR tablet solution based on a building information model (BIM). Overall, the requirements of those involved in the planning process for a digital factory model in the case of restructuring within a factory are thus first determined in a structured manner. The results are then applied and transferred to a construction site solution based on augmented reality.Keywords: augmented reality, digital factory model, factory planning, restructuring
Procedia PDF Downloads 1344574 Flood-prone Urban Area Mapping Using Machine Learning, a Case Sudy of M'sila City (Algeria)
Authors: Medjadj Tarek, Ghribi Hayet
Abstract:
This study aims to develop a flood sensitivity assessment tool using machine learning (ML) techniques and geographic information system (GIS). The importance of this study is integrating the geographic information systems (GIS) and machine learning (ML) techniques for mapping flood risks, which help decision-makers to identify the most vulnerable areas and take the necessary precautions to face this type of natural disaster. To reach this goal, we will study the case of the city of M'sila, which is among the areas most vulnerable to floods. This study drew a map of flood-prone areas based on the methodology where we have made a comparison between 3 machine learning algorithms: the xGboost model, the Random Forest algorithm and the K Nearest Neighbour algorithm. Each of them gave an accuracy respectively of 97.92 - 95 - 93.75. In the process of mapping flood-prone areas, the first model was relied upon, which gave the greatest accuracy (xGboost).Keywords: Geographic information systems (GIS), machine learning (ML), emergency mapping, flood disaster management
Procedia PDF Downloads 954573 Development of an Automatic Computational Machine Learning Pipeline to Process Confocal Fluorescence Images for Virtual Cell Generation
Authors: Miguel Contreras, David Long, Will Bachman
Abstract:
Background: Microscopy plays a central role in cell and developmental biology. In particular, fluorescence microscopy can be used to visualize specific cellular components and subsequently quantify their morphology through development of virtual-cell models for study of effects of mechanical forces on cells. However, there are challenges with these imaging experiments, which can make it difficult to quantify cell morphology: inconsistent results, time-consuming and potentially costly protocols, and limitation on number of labels due to spectral overlap. To address these challenges, the objective of this project is to develop an automatic computational machine learning pipeline to predict cellular components morphology for virtual-cell generation based on fluorescence cell membrane confocal z-stacks. Methods: Registered confocal z-stacks of nuclei and cell membrane of endothelial cells, consisting of 20 images each, were obtained from fluorescence confocal microscopy and normalized through software pipeline for each image to have a mean pixel intensity value of 0.5. An open source machine learning algorithm, originally developed to predict fluorescence labels on unlabeled transmitted light microscopy cell images, was trained using this set of normalized z-stacks on a single CPU machine. Through transfer learning, the algorithm used knowledge acquired from its previous training sessions to learn the new task. Once trained, the algorithm was used to predict morphology of nuclei using normalized cell membrane fluorescence images as input. Predictions were compared to the ground truth fluorescence nuclei images. Results: After one week of training, using one cell membrane z-stack (20 images) and corresponding nuclei label, results showed qualitatively good predictions on training set. The algorithm was able to accurately predict nuclei locations as well as shape when fed only fluorescence membrane images. Similar training sessions with improved membrane image quality, including clear lining and shape of the membrane, clearly showing the boundaries of each cell, proportionally improved nuclei predictions, reducing errors relative to ground truth. Discussion: These results show the potential of pre-trained machine learning algorithms to predict cell morphology using relatively small amounts of data and training time, eliminating the need of using multiple labels in immunofluorescence experiments. With further training, the algorithm is expected to predict different labels (e.g., focal-adhesion sites, cytoskeleton), which can be added to the automatic machine learning pipeline for direct input into Principal Component Analysis (PCA) for generation of virtual-cell mechanical models.Keywords: cell morphology prediction, computational machine learning, fluorescence microscopy, virtual-cell models
Procedia PDF Downloads 2054572 Supervised-Component-Based Generalised Linear Regression with Multiple Explanatory Blocks: THEME-SCGLR
Authors: Bry X., Trottier C., Mortier F., Cornu G., Verron T.
Abstract:
We address component-based regularization of a Multivariate Generalized Linear Model (MGLM). A set of random responses Y is assumed to depend, through a GLM, on a set X of explanatory variables, as well as on a set T of additional covariates. X is partitioned into R conceptually homogeneous blocks X1, ... , XR , viewed as explanatory themes. Variables in each Xr are assumed many and redundant. Thus, Generalised Linear Regression (GLR) demands regularization with respect to each Xr. By contrast, variables in T are assumed selected so as to demand no regularization. Regularization is performed searching each Xr for an appropriate number of orthogonal components that both contribute to model Y and capture relevant structural information in Xr. We propose a very general criterion to measure structural relevance (SR) of a component in a block, and show how to take SR into account within a Fisher-scoring-type algorithm in order to estimate the model. We show how to deal with mixed-type explanatory variables. The method, named THEME-SCGLR, is tested on simulated data.Keywords: Component-Model, Fisher Scoring Algorithm, GLM, PLS Regression, SCGLR, SEER, THEME
Procedia PDF Downloads 3964571 Determination of Myocardial Function Using Heart Accumulated Radiopharmaceuticals
Authors: C. C .D. Kulathilake, M. Jayatilake, T. Takahashi
Abstract:
The myocardium is composed of specialized muscle which relies mainly on fatty acid and sugar metabolism and it is widely contribute to the heart functioning. The changes of the cardiac energy-producing system during heart failure have been proved using autoradiography techniques. This study focused on evaluating sugar and fatty acid metabolism in myocardium as cardiac energy getting system using heart-accumulated radiopharmaceuticals. Two sets of autoradiographs of heart cross sections of Lewis male rats were analyzed and the time- accumulation curve obtained with use of the MATLAB image processing software to evaluate fatty acid and sugar metabolic functions.Keywords: autoradiographs, fatty acid, radiopharmaceuticals, sugar
Procedia PDF Downloads 451