Search results for: multi-objective fractional programming
423 AM/E/c Queuing Hub Maximal Covering Location Model with Fuzzy Parameter
Authors: M. H. Fazel Zarandi, N. Moshahedi
Abstract:
The hub location problem appears in a variety of applications such as medical centers, firefighting facilities, cargo delivery systems and telecommunication network design. The location of service centers has a strong influence on the congestion at each of them, and, consequently, on the quality of service. This paper presents a fuzzy maximal hub covering location problem (FMCHLP) in which travel costs between any pair of nodes is considered as a fuzzy variable. In order to consider the quality of service, we model each hub as a queue. Arrival rate follows Poisson distribution and service rate follows Erlang distribution. In this paper, at first, a nonlinear mathematical programming model is presented. Then, we convert it to the linear one. We solved the linear model using GAMS software up to 25 nodes and for large sizes due to the complexity of hub covering location problems, and simulated annealing algorithm is developed to solve and test the model. Also, we used possibilistic c-means clustering method in order to find an initial solution.Keywords: fuzzy modeling, location, possibilistic clustering, queuing
Procedia PDF Downloads 393422 Mondoc: Informal Lightweight Ontology for Faceted Semantic Classification of Hypernymy
Authors: M. Regina Carreira-Lopez
Abstract:
Lightweight ontologies seek to concrete union relationships between a parent node, and a secondary node, also called "child node". This logic relation (L) can be formally defined as a triple ontological relation (LO) equivalent to LO in ⟨LN, LE, LC⟩, and where LN represents a finite set of nodes (N); LE is a set of entities (E), each of which represents a relationship between nodes to form a rooted tree of ⟨LN, LE⟩; and LC is a finite set of concepts (C), encoded in a formal language (FL). Mondoc enables more refined searches on semantic and classified facets for retrieving specialized knowledge about Atlantic migrations, from the Declaration of Independence of the United States of America (1776) and to the end of the Spanish Civil War (1939). The model looks forward to increasing documentary relevance by applying an inverse frequency of co-ocurrent hypernymy phenomena for a concrete dataset of textual corpora, with RMySQL package. Mondoc profiles archival utilities implementing SQL programming code, and allows data export to XML schemas, for achieving semantic and faceted analysis of speech by analyzing keywords in context (KWIC). The methodology applies random and unrestricted sampling techniques with RMySQL to verify the resonance phenomena of inverse documentary relevance between the number of co-occurrences of the same term (t) in more than two documents of a set of texts (D). Secondly, the research also evidences co-associations between (t) and their corresponding synonyms and antonyms (synsets) are also inverse. The results from grouping facets or polysemic words with synsets in more than two textual corpora within their syntagmatic context (nouns, verbs, adjectives, etc.) state how to proceed with semantic indexing of hypernymy phenomena for subject-heading lists and for authority lists for documentary and archival purposes. Mondoc contributes to the development of web directories and seems to achieve a proper and more selective search of e-documents (classification ontology). It can also foster on-line catalogs production for semantic authorities, or concepts, through XML schemas, because its applications could be used for implementing data models, by a prior adaptation of the based-ontology to structured meta-languages, such as OWL, RDF (descriptive ontology). Mondoc serves to the classification of concepts and applies a semantic indexing approach of facets. It enables information retrieval, as well as quantitative and qualitative data interpretation. The model reproduces a triple tuple ⟨LN, LE, LT, LCF L, BKF⟩ where LN is a set of entities that connect with other nodes to concrete a rooted tree in ⟨LN, LE⟩. LT specifies a set of terms, and LCF acts as a finite set of concepts, encoded in a formal language, L. Mondoc only resolves partial problems of linguistic ambiguity (in case of synonymy and antonymy), but neither the pragmatic dimension of natural language nor the cognitive perspective is addressed. To achieve this goal, forthcoming programming developments should target at oriented meta-languages with structured documents in XML.Keywords: hypernymy, information retrieval, lightweight ontology, resonance
Procedia PDF Downloads 125421 Dynamic Thermal Modelling of a PEMFC-Type Fuel Cell
Authors: Marco Avila Lopez, Hasnae Ait-Douchi, Silvia De Los Santos, Badr Eddine Lebrouhi, Pamela Ramírez Vidal
Abstract:
In the context of the energy transition, fuel cell technology has emerged as a solution for harnessing hydrogen energy and mitigating greenhouse gas emissions. An in-depth study was conducted on a PEMFC-type fuel cell, with an initiation of an analysis of its operational principles and constituent components. Subsequently, the modelling of the fuel cell was undertaken using the Python programming language, encompassing both steady-state and transient regimes. In the case of the steady-state regime, the physical and electrochemical phenomena occurring within the fuel cell were modelled, with the assumption of uniform temperature throughout all cell compartments. Parametric identification was carried out, resulting in a remarkable mean error of only 1.62% when the model results were compared to experimental data documented in the literature. The dynamic model that was developed enabled the scrutiny of the fuel cell's response in terms of temperature and voltage under varying current conditions.Keywords: fuel cell, modelling, dynamic, thermal model, PEMFC
Procedia PDF Downloads 81420 Triassic Magmatism in Southern Beishan Orogen, Northwest China: Zircon U–Pb Geochronology, Petrogenesis and Tectonic Implications
Authors: Zengda Li
Abstract:
The tectonic evolution of the Beishan orogen, which forms part of the Central Asian Orogenic Belt, remains debated. This study reports the identification of three Triassic granitic plutons representing two distinct stages of magmatism in southern Beishan orogen. Zircon U–Pb dating constrains the early stage as 238–237 Ma and the late stage as 229–227 Ma. The granitoids belong to high-K calc-alkaline and shoshonitic series and exhibit alkalic-calcic and calc-alkalic features, and are weakly peraluminous rocks. Most of these granitoids are highly fractionated I-type and A-type granites. They have relatively high Isr values (0.7049–0.7086) and weak negative εNd(t) values of −1.5 to −2.1, with young Nd model ages of 1.04–0.91 Ga, indicating a crustal contribution. They also show markedly positive zircon εHf(t) values (+3.4 to +11.8) and two-stage Hf model ages of 1.06–0.69 Ga, indicating a mixture of mantle and crustal components. The lithospheric mantle beneath this region incorporating older subducted materials was metasomatized by fluids or melts. Partial melting of the metasomatized lithospheric mantle resulted in underplated magmas, which provided the heat and material input to generate the granitoids. The Middle Triassic granitic plutons show moderate negative Eu anomalies, enrichment of LILEs and depletion in Nb, Ta, and Ti suggesting partial melting of crustal components in response to the underplated mantle-derived magmas, probably linked to lithospheric delamination and asthenospheric upwelling. The Late Triassic granitic plutons show characteristics of post-orogenic granite with strong negative anomalies of Eu, Ba, Nb, Sr, P, and Ti, indicating fractional crystallization and crustal contamination during the emplacement process.Keywords: Triassic, magmatism, geochronology, petrogenesis, Beishan orogen
Procedia PDF Downloads 154419 Proposal of a Model Supporting Decision-Making Based on Multi-Objective Optimization Analysis on Information Security Risk Treatment
Authors: Ritsuko Kawasaki (Aiba), Takeshi Hiromatsu
Abstract:
Management is required to understand all information security risks within an organization, and to make decisions on which information security risks should be treated in what level by allocating how much amount of cost. However, such decision-making is not usually easy, because various measures for risk treatment must be selected with the suitable application levels. In addition, some measures may have objectives conflicting with each other. It also makes the selection difficult. Moreover, risks generally have trends and it also should be considered in risk treatment. Therefore, this paper provides the extension of the model proposed in the previous study. The original model supports the selection of measures by applying a combination of weighted average method and goal programming method for multi-objective analysis to find an optimal solution. The extended model includes the notion of weights to the risks, and the larger weight means the priority of the risk.Keywords: information security risk treatment, selection of risk measures, risk acceptance, multi-objective optimization
Procedia PDF Downloads 461418 Low-Cost IoT System for Monitoring Ground Propagation Waves due to Construction and Traffic Activities to Nearby Construction
Authors: Lan Nguyen, Kien Le Tan, Bao Nguyen Pham Gia
Abstract:
Due to the high cost, specialized dynamic measurement devices for industrial lands are difficult for many colleges to equip for hands-on teaching. This study connects a dynamic measurement sensor and receiver utilizing an inexpensive Raspberry Pi 4 board, some 24-bit ADC circuits, a geophone vibration sensor, and embedded Python open-source programming. Gather and analyze signals for dynamic measuring, ground vibration monitoring, and structure vibration monitoring. The system may wirelessly communicate data to the computer and is set up as a communication node network, enabling real-time monitoring of background vibrations at various locations. The device can be utilized for a variety of dynamic measurement and monitoring tasks, including monitoring earthquake vibrations, ground vibrations from construction operations, traffic, and vibrations of building structures.Keywords: sensors, FFT, signal processing, real-time data monitoring, ground propagation wave, python, raspberry Pi 4
Procedia PDF Downloads 103417 Development of a Serial Signal Monitoring Program for Educational Purposes
Authors: Jungho Moon, Lae-Jeong Park
Abstract:
This paper introduces a signal monitoring program developed with a view to helping electrical engineering students get familiar with sensors with digital output. Because the output of digital sensors cannot be simply monitored by a measuring instrument such as an oscilloscope, students tend to have a hard time dealing with digital sensors. The monitoring program runs on a PC and communicates with an MCU that reads the output of digital sensors via an asynchronous communication interface. Receiving the sensor data from the MCU, the monitoring program shows time and/or frequency domain plots of the data in real time. In addition, the monitoring program provides a serial terminal that enables the user to exchange text information with the MCU while the received data is plotted. The user can easily observe the output of digital sensors and configure the digital sensors in real time, which helps students who do not have enough experiences with digital sensors. Though the monitoring program was programmed in the Matlab programming language, it runs without the Matlab since it was compiled as a standalone executable.Keywords: digital sensor, MATLAB, MCU, signal monitoring program
Procedia PDF Downloads 496416 Cooperative Jamming for Implantable Medical Device Security
Authors: Kim Lytle, Tim Talty, Alan Michaels, Jeff Reed
Abstract:
Implantable medical devices (IMDs) are medically necessary devices embedded in the human body that monitor chronic disorders or automatically deliver therapies. Most IMDs have wireless capabilities that allow them to share data with an offboard programming device to help medical providers monitor the patient’s health while giving the patient more insight into their condition. However, serious security concerns have arisen as researchers demonstrated these devices could be hacked to obtain sensitive information or harm the patient. Cooperative jamming can be used to prevent privileged information leaks by maintaining an adequate signal-to-noise ratio at the intended receiver while minimizing signal power elsewhere. This paper uses ray tracing to demonstrate how a low number of friendly nodes abiding by Bluetooth Low Energy (BLE) transmission regulations can enhance IMD communication security in an office environment, which in turn may inform how companies and individuals can protect their proprietary and personal information.Keywords: implantable biomedical devices, communication system security, array signal processing, ray tracing
Procedia PDF Downloads 114415 A Scalable Media Job Framework for an Open Source Search Engine
Authors: Pooja Mishra, Chris Pollett
Abstract:
This paper explores efficient ways to implement various media-updating features like news aggregation, video conversion, and bulk email handling. All of these jobs share the property that they are periodic in nature, and they all benefit from being handled in a distributed fashion. The data for these jobs also often comes from a social or collaborative source. We isolate the class of periodic, one round map reduce jobs as a useful setting to describe and handle media updating tasks. As such tasks are simpler than general map reduce jobs, programming them in a general map reduce platform could easily become tedious. This paper presents a MediaUpdater module of the Yioop Open Source Search Engine Web Portal designed to handle such jobs via an extension of a PHP class. We describe how to implement various media-updating tasks in our system as well as experiments carried out using these implementations on an Amazon Web Services cluster.Keywords: distributed jobs framework, news aggregation, video conversion, email
Procedia PDF Downloads 298414 A Mathematical Model for a Two-Stage Assembly Flow-Shop Scheduling Problem with Batch Delivery System
Authors: Saeedeh Ahmadi Basir, Mohammad Mahdavi Mazdeh, Mohammad Namakshenas
Abstract:
Manufacturers often dispatch jobs in batches to reduce delivery costs. However, sending several jobs in batches can have a negative effect on other scheduling-related objective functions such as minimizing the number of tardy jobs which is often used to rate managers’ performance in many manufacturing environments. This paper aims to minimize the number of weighted tardy jobs and the sum of delivery costs of a two-stage assembly flow-shop problem in a batch delivery system. We present a mixed-integer linear programming (MILP) model to solve the problem. As this is an MILP model, the commercial solver (the CPLEX solver) is not guaranteed to find the optimal solution for large-size problems at a reasonable amount of time. We present several numerical examples to confirm the accuracy of the model.Keywords: scheduling, two-stage assembly flow-shop, tardy jobs, batched delivery system
Procedia PDF Downloads 460413 A Method for Reduction of Association Rules in Data Mining
Authors: Diego De Castro Rodrigues, Marcelo Lisboa Rocha, Daniela M. De Q. Trevisan, Marcos Dias Da Conceicao, Gabriel Rosa, Rommel M. Barbosa
Abstract:
The use of association rules algorithms within data mining is recognized as being of great value in the knowledge discovery in databases. Very often, the number of rules generated is high, sometimes even in databases with small volume, so the success in the analysis of results can be hampered by this quantity. The purpose of this research is to present a method for reducing the quantity of rules generated with association algorithms. Therefore, a computational algorithm was developed with the use of a Weka Application Programming Interface, which allows the execution of the method on different types of databases. After the development, tests were carried out on three types of databases: synthetic, model, and real. Efficient results were obtained in reducing the number of rules, where the worst case presented a gain of more than 50%, considering the concepts of support, confidence, and lift as measures. This study concluded that the proposed model is feasible and quite interesting, contributing to the analysis of the results of association rules generated from the use of algorithms.Keywords: data mining, association rules, rules reduction, artificial intelligence
Procedia PDF Downloads 160412 Transformer Design Optimization Using Artificial Intelligence Techniques
Authors: Zakir Husain
Abstract:
Main objective of a power transformer design optimization problem requires minimizing the total overall cost and/or mass of the winding and core material by satisfying all possible constraints obligatory by the standards and transformer user requirement. The constraints include appropriate limits on winding fill factor, temperature rise, efficiency, no-load current and voltage regulation. The design optimizations tasks are a constrained minimum cost and/or mass solution by optimally setting the parameters, geometry and require magnetic properties of the transformer. In this paper, present the above design problems have been formulated by using genetic algorithm (GA) and simulated annealing (SA) on the MATLAB platform. The importance of the presented approach is stems for two main features. First, proposed technique provides reliable and efficient solution for the problem of design optimization with several variables. Second, it guaranteed to obtained solution is global optimum. This paper includes a demonstration of the application of the genetic programming GP technique to transformer design.Keywords: optimization, power transformer, genetic algorithm (GA), simulated annealing technique (SA)
Procedia PDF Downloads 583411 Modeling and Simulation Frameworks for Cloud Computing Environment: A Critical Evaluation
Authors: Abul Bashar
Abstract:
The recent surge in the adoption of cloud computing systems by various organizations has brought forth the challenge of evaluating their performance. One of the major issues faced by the cloud service providers and customers is to assess the ability of cloud computing systems to provide the desired services in accordance to the QoS and SLA constraints. To this end, an opportunity exists to develop means to ensure that the desired performance levels of such systems are met under simulated environments. This will eventually minimize the service disruptions and performance degradation issues during the commissioning and operational phase of cloud computing infrastructure. However, it is observed that several simulators and modelers are available for simulating the cloud computing systems. Therefore, this paper presents a critical evaluation of the state-of-the-art modeling and simulation frameworks applicable to cloud computing systems. It compares the prominent simulation frameworks in terms of the API features, programming flexibility, operating system requirements, supported services, licensing needs and popularity. Subsequently, it provides recommendations regarding the choice of the most appropriate framework for researchers, administrators and managers of cloud computing systems.Keywords: cloud computing, modeling framework, performance evaluation, simulation tools
Procedia PDF Downloads 502410 Generative AI: A Comparison of Conditional Tabular Generative Adversarial Networks and Conditional Tabular Generative Adversarial Networks with Gaussian Copula in Generating Synthetic Data with Synthetic Data Vault
Authors: Lakshmi Prayaga, Chandra Prayaga. Aaron Wade, Gopi Shankar Mallu, Harsha Satya Pola
Abstract:
Synthetic data generated by Generative Adversarial Networks and Autoencoders is becoming more common to combat the problem of insufficient data for research purposes. However, generating synthetic data is a tedious task requiring extensive mathematical and programming background. Open-source platforms such as the Synthetic Data Vault (SDV) and Mostly AI have offered a platform that is user-friendly and accessible to non-technical professionals to generate synthetic data to augment existing data for further analysis. The SDV also provides for additions to the generic GAN, such as the Gaussian copula. We present the results from two synthetic data sets (CTGAN data and CTGAN with Gaussian Copula) generated by the SDV and report the findings. The results indicate that the ROC and AUC curves for the data generated by adding the layer of Gaussian copula are much higher than the data generated by the CTGAN.Keywords: synthetic data generation, generative adversarial networks, conditional tabular GAN, Gaussian copula
Procedia PDF Downloads 82409 Blending Effects on Crude Oil Stability: An Experimental Study
Authors: Muheddin Hamza, Entisar Etter
Abstract:
This study is a part of investigating the possibility of blending two crude oils obtained from Libyan oil fields, namely crude oil (A) and crude oil (B) with different ratios, prior to blending the crude oils have to be compatible in order to avoid phase out and precipitation of asphaltene from the bulk of crude. The physical properties of both crudes such as density, viscosity, pour point and sulphur content were measured according to (ASTM) method. To examine the stability of both crudes and their blends, the oil compatibility model using microscopic, colloidal instability index (CII) using SARA analysis and asphaltene stabilization test using Turbiscan tests were conducted in the Libyan Petroleum Institute laboratories. Compatibility tests were carried out with both crude oils, the insolubility number (IN), and the solubility blending number (SBN), for both crude oils and their blends were calculated. The criteria for compatibility of any blend is that the volume average solubility blending number (SBN) is greater than the insolubility number (IN) of any component in the blend, the results indicated that both crudes were compatible. To support the results of compatibility tests the SARA analysis was done for the fractional determination of (saturates, aromatics, resins and asphaltenes) content. From this result, the colloidal Instability index (CII) and resin to asphaltenes ratio (R/A) were calculated for crudes and their blends. The results show that crude oil (B) which has higher (R/A) and lower (CII) is more stable than crude oil (A) and as the ratio of crude (B) increases in the blend the (CII) and (R/A) were improved, and the blends becomes more stable. Asphaltene stabilization test was also conducted for the crudes and their blends using Turbiscan MA200 according to the standard test method ASTM D7061-04, the Turbiscan shows that the crude (B) is more stable than crude (A) which shows a fair tendency. The (CII) and (R/A) were compared with the solubility number (SBN) for each crude and the blends along with Turbiscan results. The solubility blending number (SBN) of the crudes and their blends show that the crudes are compatible, also by comparing (R/A) and (SBN) values of the blends, it can be seen that they are complements of each other. All the experimental results show that the blends of both crudes are more stability.Keywords: asphaltene, crude oil, compatibility, oil blends, resin, SARA
Procedia PDF Downloads 511408 Resting-State Functional Connectivity Analysis Using an Independent Component Approach
Authors: Eric Jacob Bacon, Chaoyang Jin, Dianning He, Shuaishuai Hu, Lanbo Wang, Han Li, Shouliang Qi
Abstract:
Objective: Refractory epilepsy is a complicated type of epilepsy that can be difficult to diagnose. Recent technological advancements have made resting-state functional magnetic resonance (rsfMRI) a vital technique for studying brain activity. However, there is still much to learn about rsfMRI. Investigating rsfMRI connectivity may aid in the detection of abnormal activities. In this paper, we propose studying the functional connectivity of rsfMRI candidates to diagnose epilepsy. Methods: 45 rsfMRI candidates, comprising 26 with refractory epilepsy and 19 healthy controls, were enrolled in this study. A data-driven approach known as independent component analysis (ICA) was used to achieve our goal. First, rsfMRI data from both patients and healthy controls were analyzed using group ICA. The components that were obtained were then spatially sorted to find and select meaningful ones. A two-sample t-test was also used to identify abnormal networks in patients and healthy controls. Finally, based on the fractional amplitude of low-frequency fluctuations (fALFF), a chi-square statistic test was used to distinguish the network properties of the patient and healthy control groups. Results: The two-sample t-test analysis yielded abnormal in the default mode network, including the left superior temporal lobe and the left supramarginal. The right precuneus was found to be abnormal in the dorsal attention network. In addition, the frontal cortex showed an abnormal cluster in the medial temporal gyrus. In contrast, the temporal cortex showed an abnormal cluster in the right middle temporal gyrus and the right fronto-operculum gyrus. Finally, the chi-square statistic test was significant, producing a p-value of 0.001 for the analysis. Conclusion: This study offers evidence that investigating rsfMRI connectivity provides an excellent diagnosis option for refractory epilepsy.Keywords: ICA, RSN, refractory epilepsy, rsfMRI
Procedia PDF Downloads 76407 A Parallel Implementation of Artificial Bee Colony Algorithm within CUDA Architecture
Authors: Selcuk Aslan, Dervis Karaboga, Celal Ozturk
Abstract:
Artificial Bee Colony (ABC) algorithm is one of the most successful swarm intelligence based metaheuristics. It has been applied to a number of constrained or unconstrained numerical and combinatorial optimization problems. In this paper, we presented a parallelized version of ABC algorithm by adapting employed and onlooker bee phases to the Compute Unified Device Architecture (CUDA) platform which is a graphical processing unit (GPU) programming environment by NVIDIA. The execution speed and obtained results of the proposed approach and sequential version of ABC algorithm are compared on functions that are typically used as benchmarks for optimization algorithms. Tests on standard benchmark functions with different colony size and number of parameters showed that proposed parallelization approach for ABC algorithm decreases the execution time consumed by the employed and onlooker bee phases in total and achieved similar or better quality of the results compared to the standard sequential implementation of the ABC algorithm.Keywords: Artificial Bee Colony algorithm, GPU computing, swarm intelligence, parallelization
Procedia PDF Downloads 378406 Non-Steroidal Anti-inflammatory Drugs, Plant Extracts, and Characterized Microparticles to Modulate Antimicrobial Resistance of Epidemic Meca Positive S. Aureus of Dairy Origin
Authors: Amjad I. Aqib, Shanza R. Khan, Tanveer Ahmad, Syed A. R. Shah, Muhammad A. Naseer, Muhammad Shoaib, Iqra Sarwar, Muhammad F. A. Kulyar, Zeeshan A. Bhutta, Mumtaz A. Khan, Mahboob Ali, Khadija Yasmeen
Abstract:
The current study focused on resistance modulation of dairy linked epidemic mec A positive S. aureus for resistance modulation by plant extract (Eucalyptus globolus, Calotropis procera), NSAIDs, and star like microparticles. Zinc oxide {ZnO}c and {Zn (OH)₂} microparticles were synthesized by solvothermal method and characterized by calcination, X-ray diffraction (XRD), and scanning electron microscope (SEM). Plant extracts were prepared by the Soxhlet extraction method. The study found 34% of subclinical samples (n=200) positive for S. aureus from dairy milk having significant (p < 0.05) association of assumed risk factors with pathogen. The antimicrobial assay showed 55, 42, 41, and 41% of S. aureus resistant to oxacillin, ciprofloxacin, streptomycin, and enoxacin. Amoxicillin showed the highest percentage of increase in zone of inhibitions (ZOI) at 100mg of Calotropis procera extract (31.29%) followed by 1mg/mL (28.91%) and 10mg/mL (21.68%) of Eucalyptus globolus. Amoxicillin increased ZOI by 42.85, 37.32, 29.05, and 22.78% in combination with 500 ug/ml with each of diclofenac, aspirin, ibuprofen, and meloxicam, respectively. Fractional inhibitory concentration indices (FICIs) showed synergism of amoxicillin with diclofenac and aspirin and indifferent synergy with ibuprofen and meloxicam. The preliminary in vitro finding of combination of microparticles with amoxicillin proved to be synergistic, giving rise to 26.74% and 14.85% increase in ZOI of amoxicillin in combination with zinc oxide and zinc hydroxide, respectively. The modulated antimicrobial resistance incurred by NSAIDs, plant extracts, and microparticles against pathogenic S. aureus invite immediate attention to probe alternative antimicrobial sources.Keywords: antimicrobial resistance, dairy milk, nanoparticles, NSIDs, plant extracts, resistance modulation, S. aureus
Procedia PDF Downloads 212405 Exact Energy Spectrum and Expectation Values of the Inverse Square Root Potential Model
Authors: Benedict Ita, Peter Okoi
Abstract:
In this work, the concept of the extended Nikiforov-Uvarov technique is discussed and employed to obtain the exact bound state energy eigenvalues and the corresponding normalized eigenfunctions of the inverse square root potential. With expressions for the exact energy eigenvalues and corresponding eigenfunctions, the expressions for the expectation values of the inverse separation-squared, kinetic energy, and the momentum-squared of the potential are presented using the Hellmann Feynman theorem. For visualization, algorithms written and implemented in Python language are used to generate tables and plots for l-states of the energy eigenvalues and some expectation values. The results obtained here may find suitable applications in areas like atomic and molecular physics, chemical physics, nuclear physics, and solid-state physics.Keywords: Schrodinger equation, Nikoforov-Uvarov method, inverse square root potential, diatomic molecules, Python programming, Hellmann-Feynman theorem, second order differential equation, matrix algebra
Procedia PDF Downloads 19404 Perspectives and Outcomes of a Long and Shorter Community Mental Health Program
Authors: Danielle Klassen, Reiko Yeap, Margo Schmitt-Boshnick, Scott Oddie
Abstract:
The development of the 7-week Alberta Happiness Basics program was initiated in 2010 in response to the need for community mental health programming. This provincial wide program aims to increase overall happiness and reduce negative thoughts and feelings through a positive psychology intervention. While the 7-week program has proven effective, a shortened 4-week program has additionally been developed to address client needs. In this study, participants were interviewed to determine if the 4- and 7-week programs had similar success of producing lasting behavior change at 3, 6, and 9 months post-program. A health quality of life (HQOL) measure was also used to compare the two programs and examine patient outcomes. Quantitative and qualitative analysis showed significant improvements in HQOL and sustainable behavior change for both programs. Findings indicate that the shorter, patient-centered program was effective in increasing happiness and reducing negative thoughts and feelings.Keywords: primary care, mental health, depression, short duration
Procedia PDF Downloads 270403 A Multicriteria Model for Sustainable Management in Agriculture
Authors: Basil Manos, Thomas Bournaris, Christina Moulogianni
Abstract:
The European agricultural policy supports all member states to apply agricultural development plans for the development of their agricultural sectors. A specific measure of the agricultural development plans refers to young people in order to enter into the agricultural sector. This measure helps the participating young farmers in achieving maximum efficiency, using methods and environmentally friendly practices, by altering their farm plans. This study applies a Multicriteria Mathematical Programming (MCDA) model for the young farmers to find farm plans that achieve the maximum gross margin and the minimum environmental impacts (less use of fertilizers and irrigation water). The analysis was made in the region of Central Macedonia, Greece, among young farmers who have participated in the “Setting up Young Farmers” measure during 2007-2010. The analysis includes the implementation of the MCDA model for the farm plans optimization and the comparison of selected environmental indicators with those of the existent situation.Keywords: multicriteria, optimum farm plans, environmental impacts, sustainable management
Procedia PDF Downloads 340402 Determination of Optimum Water Consumptive Using Deficit Irrigation Model for Barely: A Case Study in Arak, Iran
Authors: Mohsen Najarchi
Abstract:
This research was carried out in five fields (5-15 hectares) in Arak located in center of Iran, to determine optimum level of water consumed for Barely in four stages growth (vegetative, yield formation, flowering, and ripening). Actual evapotranspiration was calculated using measured water requirement in the fields. Five levels of water requirement equal to 50, 60, 70, 80, and 90 percents formed the treatments. To determine the optimum level of water requirement linear programming was used. The study showed 60 percent water requirement (40 percent deficit irrigation) has been the optimum level of irrigation for winter wheat in four stages of growth. Comparison between all of the treatments indicated above with normal condition (100% water requirement) shows increasing in water use efficiency. Although 40% deficit irrigation treatment lead to decrease of 38% in yield, net benefit was increasing in 11.37%. Furthermore, in comparison with normal condition, 70% of water requirement increased water use efficiency as 30%.Keywords: optimum, deficit irrigation, water use efficiency, evapotranspiration
Procedia PDF Downloads 396401 The Necessity to Standardize Procedures of Providing Engineering Geological Data for Designing Road and Railway Tunneling Projects
Authors: Atefeh Saljooghi Khoshkar, Jafar Hassanpour
Abstract:
One of the main problems of the design stage relating to many tunneling projects is the lack of an appropriate standard for the provision of engineering geological data in a predefined format. In particular, this is more reflected in highway and railroad tunnel projects in which there is a number of tunnels and different professional teams involved. In this regard, comprehensive software needs to be designed using the accepted methods in order to help engineering geologists to prepare standard reports, which contain sufficient input data for the design stage. Regarding this necessity, applied software has been designed using macro capabilities and Visual Basic programming language (VBA) through Microsoft Excel. In this software, all of the engineering geological input data, which are required for designing different parts of tunnels, such as discontinuities properties, rock mass strength parameters, rock mass classification systems, boreability classification, the penetration rate, and so forth, can be calculated and reported in a standard format.Keywords: engineering geology, rock mass classification, rock mechanic, tunnel
Procedia PDF Downloads 80400 Inhalable Lipid-Coated-Chitosan Nano-Embedded Microdroplets of an Antifungal Drug for Deep Lung Delivery
Authors: Ranjot Kaur, Om P. Katare, Anupama Sharma, Sarah R. Dennison, Kamalinder K. Singh, Bhupinder Singh
Abstract:
Respiratory microbial infections being among the top leading cause of death worldwide are difficult to treat as the microbes reside deep inside the airways, where only a small fraction of drug can access after traditional oral or parenteral routes. As a result, high doses of drugs are required to maintain drug levels above minimum inhibitory concentrations (MIC) at the infection site, unfortunately leading to severe systemic side-effects. Therefore, delivering antimicrobials directly to the respiratory tract provides an attractive way out in such situations. In this context, current study embarks on the systematic development of lung lia pid-modified chitosan nanoparticles for inhalation of voriconazole. Following the principles of quality by design, the chitosan nanoparticles were prepared by ionic gelation method and further coated with major lung lipid by precipitation method. The factor screening studies were performed by fractional factorial design, followed by optimization of the nanoparticles by Box-Behnken Design. The optimized formulation has a particle size range of 170-180nm, PDI 0.3-0.4, zeta potential 14-17, entrapment efficiency 45-50% and drug loading of 3-5%. The presence of a lipid coating was confirmed by FESEM, FTIR, and X-RD. Furthermore, the nanoparticles were found to be safe upto 40µg/ml on A549 and Calu-3 cell lines. The quantitative and qualitative uptake studies also revealed the uptake of nanoparticles in lung epithelial cells. Moreover, the data from Spraytec and next-generation impactor studies confirmed the deposition of nanoparticles in lower airways. Also, the interaction of nanoparticles with DPPC monolayers signifies its biocompatibility with lungs. Overall, the study describes the methodology and potential of lipid-coated chitosan nanoparticles in futuristic inhalation nanomedicine for the management of pulmonary aspergillosis.Keywords: dipalmitoylphosphatidylcholine, nebulization, DPPC monolayers, quality-by-design
Procedia PDF Downloads 143399 RFID Laptop Monitoring and Management System
Authors: Francis E. Idachaba, Sarah Uyimeh Tommy
Abstract:
This paper describes the design of an RFID laptop monitoring and management system. Laptops embedded with RFID chips are monitored and tracked to provide a monitoring system for the purpose of tracking as well as monitoring movement of the laptops in and out of a building. The proposed system is implemented with both hardware and software components. The hardware architecture consists of RFID passive tag, RFID module (reader), and a server hosting the application and database. The RFID readers are distributed at major exits of a building or premises. The tags are programmed with owner laptop details are concealed in the laptops. The software architecture consists of application software that has the APIs (Applications Programming Interface) necessary to interface the RFID system with the PC, to achieve automated laptop monitoring system. A friendly graphic user interface (GUI) and a database that saves all readings and owners details. The system is capable of reducing laptop theft especially in students’ hostels as laptops can be monitored as they are taken either in or out of the building.Keywords: asset tracking, GUI, laptop monitoring, radio frequency identification, passive tags
Procedia PDF Downloads 390398 Optimal Design and Simulation of a Grid-Connected Photovoltaic (PV) Power System for an Electrical Department in University of Tripoli, Libya
Authors: Mustafa Al-Refai
Abstract:
This paper presents the optimal design and simulation of a grid-connected Photovoltaic (PV) system to supply electric power to meet the energy demand by Electrical Department in University of Tripoli Libya. Solar radiation is the key factor determining electricity produced by photovoltaic (PV) systems. This paper is designed to develop a novel method to calculate the solar photovoltaic generation capacity on the basis of Mean Global Solar Radiation data available for Tripoli Libya and finally develop a system design of possible plant capacity for the available roof area. MatLab/Simulink Programming tools and monthly average solar radiation data are used for this design and simulation. The specifications of equipments are provided based on the availability of the components in the market. Simulation results and analyses are presented to validate the proposed system configuration.Keywords: photovoltaic (PV), grid, Simulink, solar energy, power plant, solar irradiation
Procedia PDF Downloads 301397 Efficient Reconstruction of DNA Distance Matrices Using an Inverse Problem Approach
Authors: Boris Melnikov, Ye Zhang, Dmitrii Chaikovskii
Abstract:
We continue to consider one of the cybernetic methods in computational biology related to the study of DNA chains. Namely, we are considering the problem of reconstructing the not fully filled distance matrix of DNA chains. When applied in a programming context, it is revealed that with a modern computer of average capabilities, creating even a small-sized distance matrix for mitochondrial DNA sequences is quite time-consuming with standard algorithms. As the size of the matrix grows larger, the computational effort required increases significantly, potentially spanning several weeks to months of non-stop computer processing. Hence, calculating the distance matrix on conventional computers is hardly feasible, and supercomputers are usually not available. Therefore, we started publishing our variants of the algorithms for calculating the distance between two DNA chains; then, we published algorithms for restoring partially filled matrices, i.e., the inverse problem of matrix processing. In this paper, we propose an algorithm for restoring the distance matrix for DNA chains, and the primary focus is on enhancing the algorithms that shape the greedy function within the branches and boundaries method framework.Keywords: DNA chains, distance matrix, optimization problem, restoring algorithm, greedy algorithm, heuristics
Procedia PDF Downloads 118396 Design and Development of Data Mining Application for Medical Centers in Remote Areas
Authors: Grace Omowunmi Soyebi
Abstract:
Data Mining is the extraction of information from a large database which helps in predicting a trend or behavior, thereby helping management make knowledge-driven decisions. One principal problem of most hospitals in rural areas is making use of the file management system for keeping records. A lot of time is wasted when a patient visits the hospital, probably in an emergency, and the nurse or attendant has to search through voluminous files before the patient's file can be retrieved; this may cause an unexpected to happen to the patient. This Data Mining application is to be designed using a Structured System Analysis and design method, which will help in a well-articulated analysis of the existing file management system, feasibility study, and proper documentation of the Design and Implementation of a Computerized medical record system. This Computerized system will replace the file management system and help to easily retrieve a patient's record with increased data security, access clinical records for decision-making, and reduce the time range at which a patient gets attended to.Keywords: data mining, medical record system, systems programming, computing
Procedia PDF Downloads 209395 Computational Fluid Dynamicsfd Simulations of Air Pollutant Dispersion: Validation of Fire Dynamic Simulator Against the Cute Experiments of the Cost ES1006 Action
Authors: Virginie Hergault, Siham Chebbah, Bertrand Frere
Abstract:
Following in-house objectives, Central laboratory of Paris police Prefecture conducted a general review on models and Computational Fluid Dynamics (CFD) codes used to simulate pollutant dispersion in the atmosphere. Starting from that review and considering main features of Large Eddy Simulation, Central Laboratory Of Paris Police Prefecture (LCPP) postulates that the Fire Dynamics Simulator (FDS) model, from National Institute of Standards and Technology (NIST), should be well suited for air pollutant dispersion modeling. This paper focuses on the implementation and the evaluation of FDS in the frame of the European COST ES1006 Action. This action aimed at quantifying the performance of modeling approaches. In this paper, the CUTE dataset carried out in the city of Hamburg, and its mock-up has been used. We have performed a comparison of FDS results with wind tunnel measurements from CUTE trials on the one hand, and, on the other, with the models results involved in the COST Action. The most time-consuming part of creating input data for simulations is the transfer of obstacle geometry information to the format required by SDS. Thus, we have developed Python codes to convert automatically building and topographic data to the FDS input file. In order to evaluate the predictions of FDS with observations, statistical performance measures have been used. These metrics include the fractional bias (FB), the normalized mean square error (NMSE) and the fraction of predictions within a factor of two of observations (FAC2). As well as the CFD models tested in the COST Action, FDS results demonstrate a good agreement with measured concentrations. Furthermore, the metrics assessment indicate that FB and NMSE meet the tolerance acceptable.Keywords: numerical simulations, atmospheric dispersion, cost ES1006 action, CFD model, cute experiments, wind tunnel data, numerical results
Procedia PDF Downloads 133394 RS Based SCADA System for Longer Distance Powered Devices
Authors: Harkishen Singh, Gavin Mangeni
Abstract:
This project aims at building an efficient and automatic power monitoring SCADA system, which is capable of monitoring the electrical parameters of high voltage powered devices in real time for example RMS voltage and current, frequency, energy consumed, power factor etc. The system uses RS-485 serial communication interface to transfer data over longer distances. Embedded C programming is the platform used to develop two hardware modules namely: RTU and Master Station modules, which both use the CC2540 BLE 4.0 microcontroller configured in slave / master mode. The Si8900 galvanic ally isolated microchip is used to perform ADC externally. The hardware communicates via UART port and sends data to the user PC using the USB port. Labview software is used to design a user interface to display current state of the power loads being monitored as well as logs data to excel spreadsheet file. An understanding of the Si8900’s auto baud rate process is key to successful implementation of this project.Keywords: SCADA, RS485, CC2540, labview, Si8900
Procedia PDF Downloads 301