Search results for: domain specific modeling language
4009 Black Box Model and Evolutionary Fuzzy Control Methods of Coupled-Tank System
Authors: S. Yaman, S. Rostami
Abstract:
In this study, a black box modeling of the coupled-tank system is obtained by using fuzzy sets. The derived model is tested via adaptive neuro fuzzy inference system (ANFIS). In order to achieve a better control performance, the parameters of three different controller types, classical proportional integral controller (PID), fuzzy PID and function tuner method, are tuned by one of the evolutionary computation method, genetic algorithm. All tuned controllers are applied to the fuzzy model of the coupled-tank experimental setup and analyzed under the different reference input values. According to the results, it is seen that function tuner method demonstrates better robust control performance and guarantees the closed loop stability.
Keywords: Function tuner method, fuzzy modeling, fuzzy PID controller, genetic algorithm.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 16484008 Retrospective Reconstruction of Time Series Data for Integrated Waste Management
Authors: A. Buruzs, M. F. Hatwágner, A. Torma, L. T. Kóczy
Abstract:
The development, operation and maintenance of Integrated Waste Management Systems (IWMS) affects essentially the sustainable concern of every region. The features of such systems have great influence on all of the components of sustainability. In order to reach the optimal way of processes, a comprehensive mapping of the variables affecting the future efficiency of the system is needed such as analysis of the interconnections among the components and modeling of their interactions. The planning of a IWMS is based fundamentally on technical and economical opportunities and the legal framework. Modeling the sustainability and operation effectiveness of a certain IWMS is not in the scope of the present research. The complexity of the systems and the large number of the variables require the utilization of a complex approach to model the outcomes and future risks. This complex method should be able to evaluate the logical framework of the factors composing the system and the interconnections between them. The authors of this paper studied the usability of the Fuzzy Cognitive Map (FCM) approach modeling the future operation of IWMS’s. The approach requires two input data set. One is the connection matrix containing all the factors affecting the system in focus with all the interconnections. The other input data set is the time series, a retrospective reconstruction of the weights and roles of the factors. This paper introduces a novel method to develop time series by content analysis.
Keywords: Content analysis, factors, integrated waste management system, time series.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 20184007 Internet Governance based on Multiple-Stakeholders: Opportunities, Issues and Developments
Authors: Martin Hans Knahl
Abstract:
The Internet is the global data communications infrastructure based on the interconnection of both public and private networks using protocols that implement Internetworking on a global scale. Hence the control of protocol and infrastructure development, resource allocation and network operation are crucial and interlinked aspects. Internet Governance is the hotly debated and contentious subject that refers to the global control and operation of key Internet infrastructure such as domain name servers and resources such as domain names. It is impossible to separate technical and political positions as they are interlinked. Furthermore the existence of a global market, transparency and competition impact upon Internet Governance and related topics such as network neutrality and security. Current trends and developments regarding Internet governance with a focus on the policy-making process, security and control have been observed to evaluate current and future implications on the Internet. The multi stakeholder approach to Internet Governance discussed in this paper presents a number of opportunities, issues and developments that will affect the future direction of the Internet. Internet operation, maintenance and advisory organisations such as the Internet Corporation for Assigned Names and Numbers (ICANN) or the Internet Governance Forum (IGF) are currently in the process of formulating policies for future Internet Governance. Given the controversial nature of the issues at stake and the current lack of agreement it is predicted that institutional as well as market governance will remain present for the network access and content.Keywords: Internet Governance, ICANN, Democracy, Security
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 18694006 Modeling and Simulations of Complex Low- Dimensional systems: Testing the Efficiency of Parallelization
Authors: Ryszard Matysiak, Grzegorz Kamieniarz
Abstract:
The deterministic quantum transfer-matrix (QTM) technique and its mathematical background are presented. This important tool in computational physics can be applied to a class of the real physical low-dimensional magnetic systems described by the Heisenberg hamiltonian which includes the macroscopic molecularbased spin chains, small size magnetic clusters embedded in some supramolecules and other interesting compounds. Using QTM, the spin degrees of freedom are accurately taken into account, yielding the thermodynamical functions at finite temperatures. In order to test the application for the susceptibility calculations to run in the parallel environment, the speed-up and efficiency of parallelization are analyzed on our platform SGI Origin 3800 with p = 128 processor units. Using Message Parallel Interface (MPI) system libraries we find the efficiency of the code of 94% for p = 128 that makes our application highly scalable.Keywords: Deterministic simulations, low-dimensional magnets, modeling of complex systems, parallelization.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 16144005 Determination of Material Properties for Biodegradable Polylactic Acid Plastic Used in 3D Printers
Authors: Juraj Beniak, Ľubomír Šooš, Peter Križan, Miloš Matúš
Abstract:
Within Rapid Prototyping technologies are used many types of materials. Many of them are recyclable but there are still as plastic like, so practically they do not degrade in the landfill. Polylactic acid (PLA) is one of the special plastic materials, which are biodegradable and available for 3D printing within Fused Deposition Modeling (FDM) technology. The question is, if the mechanical properties of produced models are comparable to similar technical plastic materials which are usual for prototype production. Presented paper shows the experiments results for tensile strength measurements for specimens prepared with different 3D printer settings and model orientation. Paper contains also the comparison of tensile strength values with values measured on specimens produced by conventional technologies as injection moulding.Keywords: 3D printing, biodegradable plastic, fused deposition modeling, PLA plastic, rapid prototyping.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 19444004 Data Privacy and Safety with Large Language Models
Authors: Ashly Joseph, Jithu Paulose
Abstract:
Large language models (LLMs) have revolutionized natural language processing capabilities, enabling applications such as chatbots, dialogue agents, image, and video generators. Nevertheless, their trainings on extensive datasets comprising personal information poses notable privacy and safety hazards. This study examines methods for addressing these challenges, specifically focusing on approaches to enhance the security of LLM outputs, safeguard user privacy, and adhere to data protection rules. We explore several methods including post-processing detection algorithms, content filtering, reinforcement learning from human and AI inputs, and the difficulties in maintaining a balance between model safety and performance. The study also emphasizes the dangers of unintentional data leakage, privacy issues related to user prompts, and the possibility of data breaches. We highlight the significance of corporate data governance rules and optimal methods for engaging with chatbots. In addition, we analyze the development of data protection frameworks, evaluate the adherence of LLMs to General Data Protection Regulation (GDPR), and examine privacy legislation in academic and business policies. We demonstrate the difficulties and remedies involved in preserving data privacy and security in the age of sophisticated artificial intelligence by employing case studies and real-life instances. This article seeks to educate stakeholders on practical strategies for improving the security and privacy of LLMs, while also assuring their responsible and ethical implementation.
Keywords: Data privacy, large language models, artificial intelligence, machine learning, cybersecurity, general data protection regulation, data safety.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1124003 Application of Spreadsheet and Queuing Network Model to Capacity Optimization in Product Development
Authors: Muhammad Marsudi, Dzuraidah Abdul Wahab, Che Hassan Che Haron
Abstract:
Modeling of a manufacturing system enables one to identify the effects of key design parameters on the system performance and as a result to make correct decision. This paper proposes a manufacturing system modeling approach using a spreadsheet model based on queuing network theory, in which a static capacity planning model and stochastic queuing model are integrated. The model was used to improve the existing system utilization in relation to product design. The model incorporates few parameters such as utilization, cycle time, throughput, and batch size. The study also showed that the validity of developed model is good enough to apply and the maximum value of relative error is 10%, far below the limit value 32%. Therefore, the model developed in this study is a valuable alternative model in evaluating a manufacturing systemKeywords: Manufacturing system, product design, spreadsheet model, utilization.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 19204002 Wetting Characterization of High Aspect Ratio Nanostructures by Gigahertz Acoustic Reflectometry
Authors: C. Virgilio, J. Carlier, P. Campistron, M. Toubal, P. Garnier, L. Broussous, V. Thomy, B. Nongaillard
Abstract:
Wetting efficiency of microstructures or nanostructures patterned on Si wafers is a real challenge in integrated circuits manufacturing. In fact, bad or non-uniform wetting during wet processes limits chemical reactions and can lead to non-complete etching or cleaning inside the patterns and device defectivity. This issue is more and more important with the transistors size shrinkage and concerns mainly high aspect ratio structures. Deep Trench Isolation (DTI) structures enabling pixels’ isolation in imaging devices are subject to this phenomenon. While low-frequency acoustic reflectometry principle is a well-known method for Non Destructive Test applications, we have recently shown that it is also well suited for nanostructures wetting characterization in a higher frequency range. In this paper, we present a high-frequency acoustic reflectometry characterization of DTI wetting through a confrontation of both experimental and modeling results. The acoustic method proposed is based on the evaluation of the reflection of a longitudinal acoustic wave generated by a 100 µm diameter ZnO piezoelectric transducer sputtered on the silicon wafer backside using MEMS technologies. The transducers have been fabricated to work at 5 GHz corresponding to a wavelength of 1.7 µm in silicon. The DTI studied structures, manufactured on the wafer frontside, are crossing trenches of 200 nm wide and 4 µm deep (aspect ratio of 20) etched into a Si wafer frontside. In that case, the acoustic signal reflection occurs at the bottom and at the top of the DTI enabling its characterization by monitoring the electrical reflection coefficient of the transducer. A Finite Difference Time Domain (FDTD) model has been developed to predict the behavior of the emitted wave. The model shows that the separation of the reflected echoes (top and bottom of the DTI) from different acoustic modes is possible at 5 Ghz. A good correspondence between experimental and theoretical signals is observed. The model enables the identification of the different acoustic modes. The evaluation of DTI wetting is then performed by focusing on the first reflected echo obtained through the reflection at Si bottom interface, where wetting efficiency is crucial. The reflection coefficient is measured with different water / ethanol mixtures (tunable surface tension) deposited on the wafer frontside. Two cases are studied: with and without PFTS hydrophobic treatment. In the untreated surface case, acoustic reflection coefficient values with water show that liquid imbibition is partial. In the treated surface case, the acoustic reflection is total with water (no liquid in DTI). The impalement of the liquid occurs for a specific surface tension but it is still partial for pure ethanol. DTI bottom shape and local pattern collapse of the trenches can explain these incomplete wetting phenomena. This high-frequency acoustic method sensitivity coupled with a FDTD propagative model thus enables the local determination of the wetting state of a liquid on real structures. Partial wetting states for non-hydrophobic surfaces or low surface tension liquids are then detectable with this method.
Keywords: Wetting, acoustic reflectometry, gigahertz, semiconductor.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 13014001 An Empirical Analysis of the Influence of Application Experience on Working Methods of Process Modelers
Authors: A. Nielen, S. Mütze-Niewöhner, C. M. Schlick
Abstract:
In view of growing competition in the service sector, services are as much in need of modeling, analysis and improvement as business or working processes. Graphical process models are important means to capture process-related know-how for an effective management of the service process. In this contribution, a human performance analysis of process model development paying special attention to model development time and the working method was conducted. It was found that modelers with higher application experience need significantly less time for mental activities than modelers with lower application experience, spend more time on labeling graphical elements, and achieved higher process model quality in terms of activity label quality.Keywords: Model quality, predetermined motion time system, process modeling, working method.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 13304000 A Comprehensive model for developing of Steer-By-Wire System
Authors: Reza Kazemi , Iman Mousavinejad
Abstract:
Steer-By-Wire ( SBW ) has several advantages of packaging flexibility , advanced vehicle control system ,and superior performance . SBW has no mechanical linkage between the steering gear and the steering column. It is possible to control the steering wheel and the front-wheel steering independently. SBW system is composed of two motors controlled by ECU. One motor in the steering wheel is to improve the driver's steering feel and the other motor in the steering linkage is to improve the vehicle maneuverability and stability. This paper shows a new approach at modeling of SBW system by Bond Graph theory. The mechanical parts , the steering wheel motor and the front wheel motor will be modeled by this theory. The work in the paper will help to guide further researches on control algorithm of the SBW system .
Keywords: Steer-By-Wire ( SBW ), Bond Graph theory, Electronic-Control-Unit ( ECU ) , Modeling
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 36523999 Urban Growth Analysis Using Multi-Temporal Satellite Images, Non-stationary Decomposition Methods and Stochastic Modeling
Authors: Ali Ben Abbes, ImedRiadh Farah, Vincent Barra
Abstract:
Remotely sensed data are a significant source for monitoring and updating databases for land use/cover. Nowadays, changes detection of urban area has been a subject of intensive researches. Timely and accurate data on spatio-temporal changes of urban areas are therefore required. The data extracted from multi-temporal satellite images are usually non-stationary. In fact, the changes evolve in time and space. This paper is an attempt to propose a methodology for changes detection in urban area by combining a non-stationary decomposition method and stochastic modeling. We consider as input of our methodology a sequence of satellite images I1, I2, … In at different periods (t = 1, 2, ..., n). Firstly, a preprocessing of multi-temporal satellite images is applied. (e.g. radiometric, atmospheric and geometric). The systematic study of global urban expansion in our methodology can be approached in two ways: The first considers the urban area as one same object as opposed to non-urban areas (e.g. vegetation, bare soil and water). The objective is to extract the urban mask. The second one aims to obtain a more knowledge of urban area, distinguishing different types of tissue within the urban area. In order to validate our approach, we used a database of Tres Cantos-Madrid in Spain, which is derived from Landsat for a period (from January 2004 to July 2013) by collecting two frames per year at a spatial resolution of 25 meters. The obtained results show the effectiveness of our method.
Keywords: Multi-temporal satellite image, urban growth, Non-stationarity, stochastic modeling.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 15043998 Talent Management through Integration of Talent Value Chain and Human Capital Analytics Approaches
Authors: Wuttigrai Ngamsirijit
Abstract:
Talent management in today’s modern organizations has become data-driven due to a demand for objective human resource decision making and development of analytics technologies. HR managers have been faced with some obstacles in exploiting data and information to obtain their effective talent management decisions. These include process-based data and records; insufficient human capital-related measures and metrics; lack of capabilities in data modeling in strategic manners; and, time consuming to add up numbers and make decisions. This paper proposes a framework of talent management through integration of talent value chain and human capital analytics approaches. It encompasses key data, measures, and metrics regarding strategic talent management decisions along the organizational and talent value chain. Moreover, specific predictive and prescriptive models incorporating these data and information are recommended to help managers in understanding the state of talent, gaps in managing talent and the organization, and the ways to develop optimized talent strategies.Keywords: Decision making, human capital analytics, talent management, talent value chain.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 9663997 Energy Consumption and GHG Production in Railway and Road Passenger Regional Transport
Authors: Martin Kendra, Tomas Skrucany, Jozef Gnap, Jan Ponicky
Abstract:
Paper deals with the modeling and simulation of energy consumption and GHG production of two different modes of regional passenger transport – road and railway. These two transport modes use the same type of fuel – diesel. Modeling and simulation of the energy consumption in transport is often used due to calculation satisfactory accuracy and cost efficiency. Paper deals with the calculation based on EN standards and information collected from technical information from vehicle producers and characteristics of tracks. Calculation included maximal theoretical capacity of bus and train and real passenger’s measurement from operation. Final energy consumption and GHG production is calculated by using software simulation. In evaluation of the simulation is used system “well to wheel”.
Keywords: Bus, energy consumption, GHG, production, simulation, train.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 15673996 Information Quality Evaluation Framework: Extending ISO 25012 Data Quality Model
Authors: Irfan Rafique, Philip Lew, Maissom Qanber Abbasi, Zhang Li
Abstract:
The world wide web coupled with the ever-increasing sophistication of online technologies and software applications puts greater emphasis on the need of even more sophisticated and consistent quality requirements modeling than traditional software applications. Web sites and Web applications (WebApps) are becoming more information driven and content-oriented raising the concern about their information quality (InQ). The consistent and consolidated modeling of InQ requirements for WebApps at different stages of the life cycle still poses a challenge. This paper proposes an approach to specify InQ requirements for WebApps by reusing and extending the ISO 25012:2008(E) data quality model. We also discuss learnability aspect of information quality for the WebApps. The proposed ISO 25012 based InQ framework is a step towards a standardized approach to evaluate WebApps InQ.Keywords: Data Quality Model, Information learnability, Information Quality, Web applications.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 57893995 Key Competences in Economics and Business Field: The Employers’ Side of the Story
Authors: Bruno Škrinjarić
Abstract:
Rapid technological developments and increase in organizations’ interdependence on international scale are changing the traditional workplace paradigm. A key feature of knowledge based economy is that employers are looking for individuals that possess both specific academic skills and knowledge, and also capability to be proactive and respond to problems creatively and autonomously. The focus of this paper is workers with Economics and Business background and its goals are threefold: (1) to explore wide range of competences and identify which are the most important to employers; (2) to investigate the existence and magnitude of gap between required and possessed level of a certain competency; and (3) to inquire how this gap is connected with performance of a company. A study was conducted on a representative sample of Croatian enterprises during the spring of 2016. Results show that generic, rather than specific, competences are more important to employers and the gap between the relative importance of certain competence and its current representation in existing workforce is greater for generic competences than for specific. Finally, results do not support the hypothesis that this gap is correlated with firms’ performance.
Keywords: Competency gap, competency matching, key competences, firm performance.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 14683994 Mathematical Model for Progressive Phase Distribution of Ku-band Reflectarray Antennas
Authors: M. Y. Ismail, M. Inam, A. F. M. Zain, N. Misran
Abstract:
Progressive phase distribution is an important consideration in reflectarray antenna design which is required to form a planar wave in front of the reflectarray aperture. This paper presents a detailed mathematical model in order to determine the required reflection phase values from individual element of a reflectarray designed in Ku-band frequency range. The proposed technique of obtaining reflection phase can be applied for any geometrical design of elements and is independent of number of array elements. Moreover the model also deals with the solution of reflectarray antenna design with both centre and off-set feed configurations. The theoretical modeling has also been implemented for reflectarrays constructed on 0.508mm thickness of different dielectric substrates. The results show an increase in the slope of the phase curve from 4.61°/mm to 22.35°/mm by varying the material properties.
Keywords: Mathematical modeling, Progressive phase distribution, Reflectarray antenna, Reflection phase.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 20673993 Formulation, Analysis and Validation of Takagi-Sugeno Fuzzy Modeling For Robotic Monipulators
Authors: Rafael Jorge Menezes Santos, Ginalber Luiz de Oliveira Serra, Carlos César Teixeira Ferreira
Abstract:
This paper proposes a methodology for analysis of the dynamic behavior of a robotic manipulator in continuous time. Initially this system (nonlinear system) will be decomposed into linear submodels and analyzed in the context of the Linear and Parameter Varying (LPV) Systems. The obtained linear submodels, which represent the local dynamic behavior of the robotic manipulator in some operating points were grouped in a Takagi-Sugeno fuzzy structure. The obtained fuzzy model was analyzed and validated through analog simulation, as universal approximator of the robotic manipulator.Keywords: modeling of nonlinear dynamic systems, Takagi- Sugeno fuzzy model, Linear and Parameter Varying (LPV) System.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 26093992 Spanning Tree Transformation of Connected Graphs into Single-Row Networks
Authors: S.L. Loh, S. Salleh, N.H. Sarmin
Abstract:
A spanning tree of a connected graph is a tree which consists the set of vertices and some or perhaps all of the edges from the connected graph. In this paper, a model for spanning tree transformation of connected graphs into single-row networks, namely Spanning Tree of Connected Graph Modeling (STCGM) will be introduced. Path-Growing Tree-Forming algorithm applied with Vertex-Prioritized is contained in the model to produce the spanning tree from the connected graph. Paths are produced by Path-Growing and they are combined into a spanning tree by Tree-Forming. The spanning tree that is produced from the connected graph is then transformed into single-row network using Tree Sequence Modeling (TSM). Finally, the single-row routing problem is solved using a method called Enhanced Simulated Annealing for Single-Row Routing (ESSR).Keywords: Graph theory, simulated annealing, single-rowrouting and spanning tree.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 17373991 Improving Quality of Business Networks for Information Systems
Authors: Hazem M. El-Bakry, Ahmed Atwan
Abstract:
Computer networks are essential part in computerbased information systems. The performance of these networks has a great influence on the whole information system. Measuring the usability criteria and customers satisfaction on small computer network is very important. In this article, an effective approach for measuring the usability of business network in an information system is introduced. The usability process for networking provides us with a flexible and a cost-effective way to assess the usability of a network and its products. In addition, the proposed approach can be used to certify network product usability late in the development cycle. Furthermore, it can be used to help in developing usable interfaces very early in the cycle and to give a way to measure, track, and improve usability. Moreover, a new approach for fast information processing over computer networks is presented. The entire data are collected together in a long vector and then tested as a one input pattern. Proposed fast time delay neural networks (FTDNNs) use cross correlation in the frequency domain between the tested data and the input weights of neural networks. It is proved mathematically and practically that the number of computation steps required for the presented time delay neural networks is less than that needed by conventional time delay neural networks (CTDNNs). Simulation results using MATLAB confirm the theoretical computations.Keywords: Usability Criteria, Computer Networks, Fast Information Processing, Cross Correlation, Frequency Domain.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 20353990 Cardiovascular Modeling Software Tools in Medicine
Authors: J. Fernandez, R. Fernandez de Canete, J. Perea-Paizal, J. C. Ramos-Diaz
Abstract:
The high prevalence of cardiovascular diseases has provoked a raising interest in the development of mathematical models in order to evaluate the cardiovascular function both under physiological and pathological conditions. In this paper, a physical model of the cardiovascular system with intrinsic regulation is presented and implemented by using the object-oriented Modelica simulation software tools. For this task, a multi-compartmental system previously validated with physiological data has been built, based on the interconnection of cardiovascular elements such as resistances, capacitances and pumping among others, by following an electrohydraulic analogy. The results obtained under both physiological and pathological scenarios provide an easy interpretative key to analyze the hemodynamic behavior of the patient. The described approach represents a valuable tool in the teaching of physiology for graduate medical and nursing students among others.
Keywords: Cardiovascular system, Modelica simulation software, physical modeling, teaching tool.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 12523989 Gaussian Process Model Identification Using Artificial Bee Colony Algorithm and Its Application to Modeling of Power Systems
Authors: Tomohiro Hachino, Hitoshi Takata, Shigeru Nakayama, Ichiro Iimura, Seiji Fukushima, Yasutaka Igarashi
Abstract:
This paper presents a nonparametric identification of continuous-time nonlinear systems by using a Gaussian process (GP) model. The GP prior model is trained by artificial bee colony algorithm. The nonlinear function of the objective system is estimated as the predictive mean function of the GP, and the confidence measure of the estimated nonlinear function is given by the predictive covariance of the GP. The proposed identification method is applied to modeling of a simplified electric power system. Simulation results are shown to demonstrate the effectiveness of the proposed method.
Keywords: Artificial bee colony algorithm, Gaussian process model, identification, nonlinear system, electric power system.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 15763988 High-Fidelity 1D Dynamic Model of a Hydraulic Servo Valve Using 3D Computational Fluid Dynamics and Electromagnetic Finite Element Analysis
Authors: D. Henninger, A. Zopey, T. Ihde, C. Mehring
Abstract:
The dynamic performance of a 4-way solenoid operated hydraulic spool valve has been analyzed by means of a one-dimensional modeling approach capturing flow, magnetic and fluid forces, valve inertia forces, fluid compressibility, and damping. Increased model accuracy was achieved by analyzing the detailed three-dimensional electromagnetic behavior of the solenoids and flow behavior through the spool valve body for a set of relevant operating conditions, thereby allowing the accurate mapping of flow and magnetic forces on the moving valve body, in lieu of representing the respective forces by lower-order models or by means of simplistic textbook correlations. The resulting high-fidelity one-dimensional model provided the basis for specific and timely design modification eliminating experimentally observed valve oscillations.Keywords: Dynamic performance model, high-fidelity model, 1D-3D decoupled analysis, solenoid-operated hydraulic servo valve, CFD and electromagnetic FEA.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 11533987 Numerical Solution of the Equations of Salt Diffusion into the Potato Tissues
Authors: Behrouz Mosayebi Dehkordi, Frazaneh Hashemi, Ramin Mostafazadeh
Abstract:
Fick's second law equations for unsteady state diffusion of salt into the potato tissues were solved numerically. The set of equations resulted from implicit modeling were solved using Thomas method to find the salt concentration profiles in solid phase. The needed effective diffusivity and equilibrium distribution coefficient were determined experimentally. Cylindrical samples of potato were infused with aqueous NaCl solutions of 1-3% concentrations, and variations in salt concentrations of brine were determined over time. Solute concentrations profiles of samples were determined by measuring salt uptake of potato slices. For the studied conditions, equilibrium distribution coefficients were found to be dependent on salt concentrations, whereas the effective diffusivity was slightly affected by brine concentration.Keywords: Brine, Diffusion, Diffusivity, Modeling, Potato
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 18863986 Total Lipid of Mutant Synechococcus sp. PCC 7002
Authors: Azlin S Azmi, Mus’ab Zainal, Sarina Sulaiman, Azura Amid, Zaki Zainudin
Abstract:
Microalgae lipid is a promising feedstock for biodiesel production. The objective of this work was to study growth factors affecting marine mutant Synechococcus sp. (PCC 7002) for high lipid production. Four growth factors were investigated; nitrogen-phosporus-potassium (NPK) concentration, light intensity, temperature and NaNO3 concentration on mutant strain growth and lipid production were studied. Design Expert v8.0 was used to design the experimental and analyze the data. The experimental design selected was Min-Run Res IV which consists of 12 runs and the response surfaces measured were specific growth rate and lipid concentration. The extraction of lipid was conducted by chloroform/methanol solvents system. Based on the study, mutant Synechococcus sp. PCC 7002 gave the highest specific growth rate of 0.0014 h-1 at 0% NPK, 2500 lux, 40oC and 0% NaNO3. On the other hand, the highest lipid concentration was obtained at 0% NPK, 3500 lux, 30oC and 1% NaNO3.
Keywords: Cyanobacteria, lipid, mutant, marine Synechococcus sp. PCC 7002, specific growth rate.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 26703985 MONARC: A Case Study on Simulation Analysis for LHC Activities
Authors: Ciprian Dobre
Abstract:
The scale, complexity and worldwide geographical spread of the LHC computing and data analysis problems are unprecedented in scientific research. The complexity of processing and accessing this data is increased substantially by the size and global span of the major experiments, combined with the limited wide area network bandwidth available. We present the latest generation of the MONARC (MOdels of Networked Analysis at Regional Centers) simulation framework, as a design and modeling tool for large scale distributed systems applied to HEP experiments. We present simulation experiments designed to evaluate the capabilities of the current real-world distributed infrastructure to support existing physics analysis processes and the means by which the experiments bands together to meet the technical challenges posed by the storage, access and computing requirements of LHC data analysis within the CMS experiment.Keywords: Modeling and simulation, evaluation, large scale distributed systems, LHC experiments, CMS.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 18113984 A Goal-Driven Crime Scripting Framework
Authors: Hashem Dehghanniri
Abstract:
Crime scripting is a simple and effective crime modeling technique that aims to improve understanding of security analysts about security and crime incidents. Low-quality scripts provide a wrong, incomplete, or sophisticated understanding of the crime commission process, which oppose the purpose of their application, e.g., identifying effective and cost-efficient situational crime prevention (SCP) measures. One important and overlooked factor in generating quality scripts is the crime scripting method. This study investigates the problems within the existing crime scripting practices and proposes a crime scripting approach that contributes to generating quality crime scripts. It was validated by experienced crime scripters. This framework helps analysts develop better crime scripts and contributes to their effective application, e.g., SCP measures identification or policy-making.
Keywords: Attack modeling, crime commission process, crime script, situational crime prevention.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 7093983 Numerical Modeling of Determination of in situ Rock Mass Deformation Modulus Using the Plate Load Test
Authors: A. Khodabakhshi, A. Mortazavi
Abstract:
Accurate determination of rock mass deformation modulus, as an important design parameter, is one of the most controversial issues in most engineering projects. A 3D numerical model of standard plate load test (PLT) using the FLAC3D code was carried to investigate the mechanism governing the test process. Five objectives were the focus of this study. The first goal was to employ 3D modeling in the interpretation of PLT conducted at the Bazoft dam site, Iran. The second objective was to investigate the effect of displacements measuring depth from the loading plates on the calculated moduli. The magnitude of rock mass deformation modulus calculated from PLT depends on anchor depth, and in practice, this may be a cause of error in the selection of realistic deformation modulus for the rock mass. The third goal of the study was to investigate the effect of testing plate diameter on the calculated modulus. Moreover, a comparison of the calculated modulus from ISRM formula, numerical modeling and calculated modulus from the actual PLT carried out at right abutment of the Bazoft dam site was another objective of the study. Finally, the effect of plastic strains on the calculated moduli in each of the loading-unloading cycles for three loading plates was investigated. The geometry, material properties, and boundary conditions on the constructed 3D model were selected based on the in-situ conditions of PLT at Bazoft dam site. A good agreement was achieved between numerical model results and the field tests results.
Keywords: Deformation modulus, numerical model, plate loading test, rock mass.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 7723982 Implementing a Database from a Requirement Specification
Abstract:
Creating a database scheme is essentially a manual process. From a requirement specification the information contained within has to be analyzed and reduced into a set of tables, attributes and relationships. This is a time consuming process that has to go through several stages before an acceptable database schema is achieved. The purpose of this paper is to implement a Natural Language Processing (NLP) based tool to produce a relational database from a requirement specification. The Stanford CoreNLP version 3.3.1 and the Java programming were used to implement the proposed model. The outcome of this study indicates that a first draft of a relational database schema can be extracted from a requirement specification by using NLP tools and techniques with minimum user intervention. Therefore this method is a step forward in finding a solution that requires little or no user intervention.
Keywords: Information Extraction, Natural Language Processing, Relation Extraction.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 22283981 Extending the Flipped Classroom Approach: Using Technology in Module Delivery to Students of English Language and Literature at the British University in Egypt
Authors: Azza Taha Zaki
Abstract:
Technology-enhanced teaching has been in the limelight since the 90s when educators started investigating and experimenting with using computers in the classroom as a means of building 21st. century skills and motivating students. The concept of technology-enhanced strategies in education is kaleidoscopic! It has meant different things to different educators. For the purpose of this paper, however, it will be used to refer to the diverse technology-based strategies used to support and enrich the flipped learning process, in the classroom and outside. The paper will investigate how technology is put in the service of teaching and learning to improve the students’ learning experience as manifested in students’ attendance and engagement, achievement rates and finally, students’ projects at the end of the semester. The results will be supported by a student survey about relevant specific aspects of their learning experience in the modules in the study.
Keywords: Attendance, British University, Egypt, flipped, student achievement, student-centred, student engagement, students’ projects.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 6723980 Learning Monte Carlo Data for Circuit Path Length
Authors: Namal A. Senanayake, A. Beg, Withana C. Prasad
Abstract:
This paper analyzes the patterns of the Monte Carlo data for a large number of variables and minterms, in order to characterize the circuit path length behavior. We propose models that are determined by training process of shortest path length derived from a wide range of binary decision diagram (BDD) simulations. The creation of the model was done use of feed forward neural network (NN) modeling methodology. Experimental results for ISCAS benchmark circuits show an RMS error of 0.102 for the shortest path length complexity estimation predicted by the NN model (NNM). Use of such a model can help reduce the time complexity of very large scale integrated (VLSI) circuitries and related computer-aided design (CAD) tools that use BDDs.Keywords: Monte Carlo data, Binary decision diagrams, Neural network modeling, Shortest path length estimation.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1595