Search results for: logic programming
348 Modelling Conceptual Quantities Using Support Vector Machines
Authors: Ka C. Lam, Oluwafunmibi S. Idowu
Abstract:
Uncertainty in cost is a major factor affecting performance of construction projects. To our knowledge, several conceptual cost models have been developed with varying degrees of accuracy. Incorporating conceptual quantities into conceptual cost models could improve the accuracy of early predesign cost estimates. Hence, the development of quantity models for estimating conceptual quantities of framed reinforced concrete structures using supervised machine learning is the aim of the current research. Using measured quantities of structural elements and design variables such as live loads and soil bearing pressures, response and predictor variables were defined and used for constructing conceptual quantities models. Twenty-four models were developed for comparison using a combination of non-parametric support vector regression, linear regression, and bootstrap resampling techniques. R programming language was used for data analysis and model implementation. Gross soil bearing pressure and gross floor loading were discovered to have a major influence on the quantities of concrete and reinforcement used for foundations. Building footprint and gross floor loading had a similar influence on beams and slabs. Future research could explore the modelling of other conceptual quantities for walls, finishes, and services using machine learning techniques. Estimation of conceptual quantities would assist construction planners in early resource planning and enable detailed performance evaluation of early cost predictions.Keywords: bootstrapping, conceptual quantities, modelling, reinforced concrete, support vector regression
Procedia PDF Downloads 205347 Employing a System of Systems Approach in the Maritime RobotX Challenge: Incorporating Information Technology Students in the Development of an Autonomous Catamaran
Authors: Adam Jenkins
Abstract:
The Maritime RobotX Challenge provides a platform for postgraduate students conducting research in autonomous robotic systems to participate in an international competition. Although targeted to postgraduate students, the problem domain lends itself to a wide range of different levels of student expertise. In 2022, undergraduate Information Technology students from the University of South Australia undertook the challenge, utilizing a System of the Systems approach to the project's architecture. Each student group produced an independent solution to an identified task, which was then implemented on a Single Board Computer (SBC). A Central Control System then engaged each solution when appropriate, allowing the encapsulated SBC systems to manage each task as it was encountered. This approach facilitated collaboration among the multiple independent student teams over an 18-month period, and the fundamental system-agnostic architecture allowed for both the variance in student solutions and the limitations caused by the global electronics shortage. By adopting this approach, Information Technology teams were able to work independently yet produce an effective solution, leveraging their expertise to develop and construct an autonomous catamaran capable of meeting the competition's demanding requirements while producing a high level of engagement. The System of Systems approach is recommended to other universities interested in competing at this level and engaging students in a real-world problem.Keywords: case study, robotics, education, programming, system of systems, multi-disciplinary collaboration
Procedia PDF Downloads 76346 Biases in Macroprudential Supervision and Their Legal Implications
Authors: Anat Keller
Abstract:
Given that macro-prudential supervision is a relatively new policy area and its empirical and analytical research are still in their infancy, its theoretical foundations are also lagging behind. This paper contributes to the developing discussion on effective legal and institutional macroprudential supervision frameworks. In the first part of the paper, it is argued that effectiveness as a key benchmark poses some challenges in the context of macroprudential supervision such as the difficulty in proving causality between supervisory actions and the achievement of the supervisor’s mission. The paper suggests that effectiveness in the macroprudential context should, therefore, be assessed at the supervisory decision-making process (to be differentiated from the supervisory outcomes). The second part of the essay examines whether insights from behavioural economics can point to biases in the macroprudential decision-making process. These biases include, inter alia, preference bias, groupthink bias and inaction bias. It is argued that these biases are exacerbated in the multilateral setting of the macroprudential supervision framework in the EU. The paper then examines how legal and institutional frameworks should be designed to acknowledge and perhaps contain these identified biases. The paper suggests that the effectiveness of macroprudential policy will largely depend on the existence of clear and robust transparency and accountability arrangements. Accountability arrangements can be used as a vehicle for identifying and addressing potential biases in the macro-prudential framework, in particular, inaction bias. Inclusiveness of the public in the supervisory process in the form of transparency and awareness of the logic behind policy decisions may assist in minimising their potential unpopularity thus promoting their effectiveness. Furthermore, a governance structure which facilitates coordination of the macroprudential supervisor with other policymakers and incorporates outside perspectives and opinions could ‘break-down’ groupthink bias as well as inaction bias.Keywords: behavioural economics and biases, effectiveness of macroprudential supervision, legal and institutional macroprudential frameworks, macroprudential decision-making process
Procedia PDF Downloads 280345 Cache Analysis and Software Optimizations for Faster on-Chip Network Simulations
Authors: Khyamling Parane, B. M. Prabhu Prasad, Basavaraj Talawar
Abstract:
Fast simulations are critical in reducing time to market in CMPs and SoCs. Several simulators have been used to evaluate the performance and power consumed by Network-on-Chips. Researchers and designers rely upon these simulators for design space exploration of NoC architectures. Our experiments show that simulating large NoC topologies take hours to several days for completion. To speed up the simulations, it is necessary to investigate and optimize the hotspots in simulator source code. Among several simulators available, we choose Booksim2.0, as it is being extensively used in the NoC community. In this paper, we analyze the cache and memory system behaviour of Booksim2.0 to accurately monitor input dependent performance bottlenecks. Our measurements show that cache and memory usage patterns vary widely based on the input parameters given to Booksim2.0. Based on these measurements, the cache configuration having least misses has been identified. To further reduce the cache misses, we use software optimization techniques such as removal of unused functions, loop interchanging and replacing post-increment operator with pre-increment operator for non-primitive data types. The cache misses were reduced by 18.52%, 5.34% and 3.91% by employing above technology respectively. We also employ thread parallelization and vectorization to improve the overall performance of Booksim2.0. The OpenMP programming model and SIMD are used for parallelizing and vectorizing the more time-consuming portions of Booksim2.0. Speedups of 2.93x and 3.97x were observed for the Mesh topology with 30 × 30 network size by employing thread parallelization and vectorization respectively.Keywords: cache behaviour, network-on-chip, performance profiling, vectorization
Procedia PDF Downloads 197344 Design of an Automatic Bovine Feeding Machine
Authors: Huseyin A. Yavasoglu, Yusuf Ziya Tengiz, Ali Göksenli
Abstract:
In this study, an automatic feeding machine for different type and class of bovine animals is designed. Daily nutrition of a bovine consists of grass, corn, straw, silage, oat, wheat and different vitamins and minerals. The amount and mixture amount of each of the nutrition depends on different parameters of the bovine. These parameters are; age, sex, weight and maternity of the bovine, also outside temperature. The problem in a farm is to constitute the correct mixture and amount of nutrition for each animal. Faulty nutrition will cause an insufficient feeding of the animal concluding in an unhealthy bovine. To solve this problem, a new automatic feeding machine is designed. Travelling of the machine is performed by four tires, which is pulled by a tractor. The carrier consists of eight bins, which each of them carries a nutrition type. Capacity of each unit is 250 kg. At the bottom of each chamber is a sensor measuring the weight of the food inside. A funnel is at the bottom of each chamber by which open/close function is controlled by a valve. Each animal will carry a RFID tag including ID on its ear. A receiver on the feeding machine will read this ID and by given previous information by the operator (veterinarian), the system will detect the amount of each nutrition unit which will be given to the selected animal for feeding. In the system, each bin will open its exit gate by the help of the valve under the control of PLC (Programmable Logic Controller). The amount of each nutrition type will be controlled by measuring the open/close time. The exit canals of the bins are collected in a reservoir. To achieve a homogenous nitration, the collected feed will be mixed by a worm gear. Further the mixture will be transported by a help of a funnel to the feeding unit of the animal. The feeding process can be performed in 100 seconds. After feeding of the animal, the tractor pulls the travelling machine to the next animal. By the help of this system animals can be feeded by right amount and mixture of nutritionKeywords: bovine, feeding, nutrition, transportation, automatic
Procedia PDF Downloads 342343 Crashworthiness Optimization of an Automotive Front Bumper in Composite Material
Authors: S. Boria
Abstract:
In the last years, the crashworthiness of an automotive body structure can be improved, since the beginning of the design stage, thanks to the development of specific optimization tools. It is well known how the finite element codes can help the designer to investigate the crashing performance of structures under dynamic impact. Therefore, by coupling nonlinear mathematical programming procedure and statistical techniques with FE simulations, it is possible to optimize the design with reduced number of analytical evaluations. In engineering applications, many optimization methods which are based on statistical techniques and utilize estimated models, called meta-models, are quickly spreading. A meta-model is an approximation of a detailed simulation model based on a dataset of input, identified by the design of experiments (DOE); the number of simulations needed to build it depends on the number of variables. Among the various types of meta-modeling techniques, Kriging method seems to be excellent in accuracy, robustness and efficiency compared to other ones when applied to crashworthiness optimization. Therefore the application of such meta-model was used in this work, in order to improve the structural optimization of a bumper for a racing car in composite material subjected to frontal impact. The specific energy absorption represents the objective function to maximize and the geometrical parameters subjected to some design constraints are the design variables. LS-DYNA codes were interfaced with LS-OPT tool in order to find the optimized solution, through the use of a domain reduction strategy. With the use of the Kriging meta-model the crashworthiness characteristic of the composite bumper was improved.Keywords: composite material, crashworthiness, finite element analysis, optimization
Procedia PDF Downloads 256342 Modal Analysis of Functionally Graded Materials Plates Using Finite Element Method
Authors: S. J. Shahidzadeh Tabatabaei, A. M. Fattahi
Abstract:
Modal analysis of an FGM plate composed of Al2O3 ceramic phase and 304 stainless steel metal phases was performed in this paper by ABAQUS software with the assumption that the behavior of material is elastic and mechanical properties (Young's modulus and density) are variable in the thickness direction of the plate. Therefore, a sub-program was written in FORTRAN programming language and was linked with ABAQUS software. For modal analysis, a finite element analysis was carried out similar to the model of other researchers and the accuracy of results was evaluated after comparing the results. Comparison of natural frequencies and mode shapes reflected the compatibility of results and optimal performance of the program written in FORTRAN as well as high accuracy of finite element model used in this research. After validation of the results, it was evaluated the effect of material (n parameter) on the natural frequency. In this regard, finite element analysis was carried out for different values of n and in simply supported mode. About the effect of n parameter that indicates the effect of material on the natural frequency, it was observed that the natural frequency decreased as n increased; because by increasing n, the share of ceramic phase on FGM plate has decreased and the share of steel phase has increased and this led to reducing stiffness of FGM plate and thereby reduce in the natural frequency. That is because the Young's modulus of Al2O3 ceramic is equal to 380 GPa and Young's modulus of SUS304 steel is 207 GPa.Keywords: FGM plates, modal analysis, natural frequency, finite element method
Procedia PDF Downloads 391341 Old and New Paradigms for Pre-Earthquake Prevention and Post-Earthquake Regeneration of Territories in Crisis in Italy
Authors: Maria Angela Bedini, Fabio Bronzini
Abstract:
Most of the Italian territory is at seismic risk. Many earthquakes have hit Italy, and devastating effects have been generated. The specific objective of the research is to distinguish the negative approaches that have generated unacceptable social situations of marginalization, abandonment, and economic regression, from positive methodological approaches. On the basis of the different situations examined, the study proposes strategies and guidelines to obtain the best possible results, in Italy or abroad, in the event of new earthquakes. At national and international level, many theoretical studies address the aspects of prevention, while the comparisons, carried out in this study, between the techniques and the operative procedures applied and the results obtained are rare. The adopted methodology compares the different pre-earthquake urban-planning approaches, for the emergency (temporary urban planning), and for the post-earthquake (socio-economic-territorial processes) in Italy. Attention is placed on the current consolidated planning and programming acquisitions, pre and post-earthquake. The main results of the study concern the prospects in Italy of protection from seismic risks in the next decades. An integrated settlement system for a new economic and social model, aimed at the rebirth of territories in crisis, is proposed. Finally, the conclusions describe the disciplinary positions, procedures and the fundamental points generally shared by the scientific community for each approach, in order to identify the strategic choices and the disciplinary and management paths that will be followed in the coming decades.Keywords: post-earthquake, seismic emergency, seismic prevention, urban planning interventions in Italy
Procedia PDF Downloads 128340 Conflict, Confusion or Compromise: Violence against Women, A Case Study of Pakistan
Authors: Farhat Jabeen, Syed Asfaq Hussain Bukhari
Abstract:
In the wake of the contemporary period the basic objective of the research paper points out that socio-cultural scenario of Pakistan reveals that gender-based violence is deep rooted in the society irrespective of language and ethnicity. This paper would reconnaissance the possibility reforms in Pakistan for diminishing of violence. Women are not given their due role, rights, and respect. Furthermore, they are treated as chattels. This presentation will cover the socio-customary practices in the context of discrimination, stigmatization, and violence against women. This paper envisages justice in a broader sense of recognition of rights for women, and masculine structure of society, socio-customary practices and discrimination against women are a very serious concern which needs to be understood as a multidimensional problem. The paper will specially focus on understanding the existing obstacles of women in Pakistan in the constitutional scenario. Women stumble across discrimination and human rights manipulations, voluptuous violation and manipulation including domestic viciousness and are disadvantaged by laws, strategies, and programming that do not take their concerns into considerations. This presentation examines the role of honour killings among Pakistani community. This affects their self-assurance and capability to elevation integrity campaign where gender inequalities and discrimination in social, legal domain are to be put right. This paper brings to light the range of practices, laws and legal justice regarding the status of women and also covers attitude towards compensations for murders/killings, domestic violence, rape, adultery, social behavior and recourse to justice.Keywords: discrimination, cultural, women, violence
Procedia PDF Downloads 324339 Method to Find a ε-Optimal Control of Stochastic Differential Equation Driven by a Brownian Motion
Authors: Francys Souza, Alberto Ohashi, Dorival Leao
Abstract:
We present a general solution for finding the ε-optimal controls for non-Markovian stochastic systems as stochastic differential equations driven by Brownian motion, which is a problem recognized as a difficult solution. The contribution appears in the development of mathematical tools to deal with modeling and control of non-Markovian systems, whose applicability in different areas is well known. The methodology used consists to discretize the problem through a random discretization. In this way, we transform an infinite dimensional problem in a finite dimensional, thereafter we use measurable selection arguments, to find a control on an explicit form for the discretized problem. Then, we prove the control found for the discretized problem is a ε-optimal control for the original problem. Our theory provides a concrete description of a rather general class, among the principals, we can highlight financial problems such as portfolio control, hedging, super-hedging, pairs-trading and others. Therefore, our main contribution is the development of a tool to explicitly the ε-optimal control for non-Markovian stochastic systems. The pathwise analysis was made through a random discretization jointly with measurable selection arguments, has provided us with a structure to transform an infinite dimensional problem into a finite dimensional. The theory is applied to stochastic control problems based on path-dependent stochastic differential equations, where both drift and diffusion components are controlled. We are able to explicitly show optimal control with our method.Keywords: dynamic programming equation, optimal control, stochastic control, stochastic differential equation
Procedia PDF Downloads 188338 Examining the Possibility of Establishing Regional Environmental Governance in the Middle East
Authors: Somayeh Bahrami, Seyed Jalal Dehghani Firoozabadi
Abstract:
Environmental governance is an interdisciplinary concept in political ecology and environmental policy focusing on the necessity of embedding the environmental issues in all levels of decision-making and act of states. Similar to sustainable development the concept of environmental governance believes that economic and political life of societies and countries need to be considered as a subset of the environment. This concept has been accepted by North Countries, those that have done the most irreparable environmental damage since the Industrial Revolution. Although North Countries are more responsible for damage to the environment, considering the global fluidity logic of environmental challenges, such an impression doesn’t cause developing countries to disavow responsibility for regional and international cooperation to protect the environment. Establishing an environmental governance at all levels of local, national, regional and global is one of the most significant ways to improve sustainable development. Given to the various political and economic difficulties developing countries including the Middle East face, building environmental governance in these countries is difficult but feasible, as these difficulties have not impeded their mutual partnership for confronting joint environmental issues. However, the environmental issues wouldn’t be solved only by mutual partnership but by establishing environmental governance, establishing regional environmental institutions (an introduction to building Regional Environmental Governance) and delegation of some environmental authorities to the mentioned institutions. The research is aimed at examining necessities, opportunities, and barriers to establishing Regional Environmental Governance in the Middle East. Therefore, this research seeks to answer the question of whether establishing Regional Environmental Governance is possible in the Middle East and if so then why. This study used descriptive-analytical methods and the inferential methodology has been used to reach the goals. Data has been collected by using library and internet sources as well as news sources on the basis of objective-historical data.Keywords: environmental democracy (ED), environmental governance (EG), middle east (ME), regional environmental governance (REG)
Procedia PDF Downloads 456337 Joint Replenishment and Heterogeneous Vehicle Routing Problem with Cyclical Schedule
Authors: Ming-Jong Yao, Chin-Sum Shui, Chih-Han Wang
Abstract:
This paper is developed based on a real-world decision scenario that an industrial gas company that applies the Vendor Managed Inventory model and supplies liquid oxygen with a self-operated heterogeneous vehicle fleet to hospitals in nearby cities. We name it as a Joint Replenishment and Heterogeneous Vehicle Routing Problem with Cyclical Schedule and formulate it as a non-linear mixed-integer linear programming problem which simultaneously determines the length of the planning cycle (PC), the length of the replenishment cycle and the dates of replenishment for each customer and the vehicle routes of each day within PC, such that the average daily operation cost within PC, including inventory holding cost, setup cost, transportation cost, and overtime labor cost, is minimized. A solution method based on genetic algorithm, embedded with an encoding and decoding mechanism and local search operators, is then proposed, and the hash function is adopted to avoid repetitive fitness evaluation for identical solutions. Numerical experiments demonstrate that the proposed solution method can effectively solve the problem under different lengths of PC and number of customers. The method is also shown to be effective in determining whether the company should expand the storage capacity of a customer whose demand increases. Sensitivity analysis of the vehicle fleet composition shows that deploying a mixed fleet can reduce the daily operating cost.Keywords: cyclic inventory routing problem, joint replenishment, heterogeneous vehicle, genetic algorithm
Procedia PDF Downloads 87336 Programmatic Actions of Social Welfare State in Service to Justice: Law, Society and the Third Sector
Authors: Bruno Valverde Chahaira, Matheus Jeronimo Low Lopes, Marta Beatriz Tanaka Ferdinandi
Abstract:
This paper proposes to dissect the meanings and / or directions of the State, in order, to present the State models to elaborate a conceptual framework about its function in the legal scope. To do so, it points out the possible contracts established between the State and the Society, since the general principles immanent in them can guide the models of society in force. From this orientation arise the contracts, whose purpose is by the effect to modify the status (the being and / or the opinion) of each of the subjects in presence - State and Society. In this logic, this paper announces the fiduciary contracts and “veredicção”(portuguese word) contracts, from the perspective of semiotics discourse (or greimasian). Therefore, studies focus on the issue of manifest language in unilateral and bilateral or reciprocal relations between the State and Society. Thus, under the biases of the model of the communicative situation and discourse, the guidelines of these contractual relations will be analyzed in order to see if there is a pragmatic sanction: positive when the contract is signed between the subjects (reward), or negative when the contract between they are broken (punishment). In this way, a third path emerges which, in this specific case, passes through the subject-third sector. In other words, the proposal, which is systemic in nature, is to analyze whether, since the contract of the welfare state is not carried out in the constitutional program on fundamental rights: education, health, housing, an others. Therefore, in the structure of the exchange demanded by the society according to its contractual obligations (others), the third way (Third Sector) advances in the empty space left by the State. In this line, it presents the modalities of action of the third sector in the social scope. Finally, the normative communication organization of these three subjects is sought in the pragmatic model of discourse, namely: State, Society and Third Sector, in an attempt to understand the constant dynamics in the Law and in the language of the relations established between them.Keywords: access to justice, state, social rights, third sector
Procedia PDF Downloads 145335 Multi-Objective Optimization for Aircraft Fleet Management: A Parametric Approach
Authors: Xin-Yu Li, Dung-Ying Lin
Abstract:
Fleet availability is a crucial indicator for an aircraft fleet. However, in practice, fleet planning involves many resource and safety constraints, such as annual and monthly flight training targets and maximum engine usage limits. Due to safety considerations, engines must be removed for mandatory maintenance and replacement of key components. This situation is known as the "threshold." The annual number of thresholds is a key factor in maintaining fleet availability. However, the traditional method heavily relies on experience and manual planning, which may result in ineffective engine usage and affect the flight missions. This study aims to address the challenges of fleet planning and availability maintenance in aircraft fleets with resource and safety constraints. The goal is to effectively optimize engine usage and maintenance tasks. This study has four objectives: minimizing the number of engine thresholds, minimizing the monthly lack of flight hours, minimizing the monthly excess of flight hours, and minimizing engine disassembly frequency. To solve the resulting formulation, this study uses parametric programming techniques and ϵ-constraint method to reformulate multi-objective problems into single-objective problems, efficiently generating Pareto fronts. This method is advantageous when handling multiple conflicting objectives. It allows for an effective trade-off between these competing objectives. Empirical results and managerial insights will be provided.Keywords: aircraft fleet, engine utilization planning, multi-objective optimization, parametric method, Pareto optimality
Procedia PDF Downloads 23334 The Impact of Cognitive Load on Deceit Detection and Memory Recall in Children’s Interviews: A Meta-Analysis
Authors: Sevilay Çankaya
Abstract:
The detection of deception in children’s interviews is essential for statement veracity. The widely used method for deception detection is building cognitive load, which is the logic of the cognitive interview (CI), and its effectiveness for adults is approved. This meta-analysis delves into the effectiveness of inducing cognitive load as a means of enhancing veracity detection during interviews with children. Additionally, the effectiveness of cognitive load on children's total number of events recalled is assessed as a second part of the analysis. The current meta-analysis includes ten effect sizes from search using databases. For the effect size calculation, Hedge’s g was used with a random effect model by using CMA version 2. Heterogeneity analysis was conducted to detect potential moderators. The overall result indicated that cognitive load had no significant effect on veracity outcomes (g =0.052, 95% CI [-.006,1.25]). However, a high level of heterogeneity was found (I² = 92%). Age, participants’ characteristics, interview setting, and characteristics of the interviewer were coded as possible moderators to explain variance. Age was significant moderator (β = .021; p = .03, R2 = 75%) but the analysis did not reveal statistically significant effects for other potential moderators: participants’ characteristics (Q = 0.106, df = 1, p = .744), interview setting (Q = 2.04, df = 1, p = .154), and characteristics of interviewer (Q = 2.96, df = 1, p = .086). For the second outcome, the total number of events recalled, the overall effect was significant (g =4.121, 95% CI [2.256,5.985]). The cognitive load was effective in total recalled events when interviewing with children. All in all, while age plays a crucial role in determining the impact of cognitive load on veracity, the surrounding context, interviewer attributes, and inherent participant traits may not significantly alter the relationship. These findings throw light on the need for more focused, age-specific methods when using cognitive load measures. It may be possible to improve the precision and dependability of deceit detection in children's interviews with the help of more studies in this field.Keywords: deceit detection, cognitive load, memory recall, children interviews, meta-analysis
Procedia PDF Downloads 55333 Stochastic Optimization of a Vendor-Managed Inventory Problem in a Two-Echelon Supply Chain
Authors: Bita Payami-Shabestari, Dariush Eslami
Abstract:
The purpose of this paper is to develop a multi-product economic production quantity model under vendor management inventory policy and restrictions including limited warehouse space, budget, and number of orders, average shortage time and maximum permissible shortage. Since the “costs” cannot be predicted with certainty, it is assumed that data behave under uncertain environment. The problem is first formulated into the framework of a bi-objective of multi-product economic production quantity model. Then, the problem is solved with three multi-objective decision-making (MODM) methods. Then following this, three methods had been compared on information on the optimal value of the two objective functions and the central processing unit (CPU) time with the statistical analysis method and the multi-attribute decision-making (MADM). The results are compared with statistical analysis method and the MADM. The results of the study demonstrate that augmented-constraint in terms of optimal value of the two objective functions and the CPU time perform better than global criteria, and goal programming. Sensitivity analysis is done to illustrate the effect of parameter variations on the optimal solution. The contribution of this research is the use of random costs data in developing a multi-product economic production quantity model under vendor management inventory policy with several constraints.Keywords: economic production quantity, random cost, supply chain management, vendor-managed inventory
Procedia PDF Downloads 129332 Distributed Cost-Based Scheduling in Cloud Computing Environment
Authors: Rupali, Anil Kumar Jaiswal
Abstract:
Cloud computing can be defined as one of the prominent technologies that lets a user change, configure and access the services online. it can be said that this is a prototype of computing that helps in saving cost and time of a user practically the use of cloud computing can be found in various fields like education, health, banking etc. Cloud computing is an internet dependent technology thus it is the major responsibility of Cloud Service Providers(CSPs) to care of data stored by user at data centers. Scheduling in cloud computing environment plays a vital role as to achieve maximum utilization and user satisfaction cloud providers need to schedule resources effectively. Job scheduling for cloud computing is analyzed in the following work. To complete, recreate the task calculation, and conveyed scheduling methods CloudSim3.0.3 is utilized. This research work discusses the job scheduling for circulated processing condition also by exploring on this issue we find it works with minimum time and less cost. In this work two load balancing techniques have been employed: ‘Throttled stack adjustment policy’ and ‘Active VM load balancing policy’ with two brokerage services ‘Advanced Response Time’ and ‘Reconfigure Dynamically’ to evaluate the VM_Cost, DC_Cost, Response Time, and Data Processing Time. The proposed techniques are compared with Round Robin scheduling policy.Keywords: physical machines, virtual machines, support for repetition, self-healing, highly scalable programming model
Procedia PDF Downloads 168331 Development of a Matlab® Program for the Bi-Dimensional Truss Analysis Using the Stiffness Matrix Method
Authors: Angel G. De Leon Hernandez
Abstract:
A structure is defined as a physical system or, in certain cases, an arrangement of connected elements, capable of bearing certain loads. The structures are presented in every part of the daily life, e.g., in the designing of buildings, vehicles and mechanisms. The main goal of a structure designer is to develop a secure, aesthetic and maintainable system, considering the constraint imposed to every case. With the advances in the technology during the last decades, the capabilities of solving engineering problems have increased enormously. Nowadays the computers, play a critical roll in the structural analysis, pitifully, for university students the vast majority of these software are inaccessible due to the high complexity and cost they represent, even when the software manufacturers offer student versions. This is exactly the reason why the idea of developing a more reachable and easy-to-use computing tool. This program is designed as a tool for the university students enrolled in courser related to the structures analysis and designs, as a complementary instrument to achieve a better understanding of this area and to avoid all the tedious calculations. Also, the program can be useful for graduated engineers in the field of structural design and analysis. A graphical user interphase is included in the program to make it even simpler to operate it and understand the information requested and the obtained results. In the present document are included the theoretical basics in which the program is based to solve the structural analysis, the logical path followed in order to develop the program, the theoretical results, a discussion about the results and the validation of those results.Keywords: stiffness matrix method, structural analysis, Matlab® applications, programming
Procedia PDF Downloads 122330 Solving the Economic Load Dispatch Problem Using Differential Evolution
Authors: Alaa Sheta
Abstract:
Economic Load Dispatch (ELD) is one of the vital optimization problems in power system planning. Solving the ELD problems mean finding the best mixture of power unit outputs of all members of the power system network such that the total fuel cost is minimized while sustaining operation requirements limits satisfied across the entire dispatch phases. Many optimization techniques were proposed to solve this problem. A famous one is the Quadratic Programming (QP). QP is a very simple and fast method but it still suffer many problem as gradient methods that might trapped at local minimum solutions and cannot handle complex nonlinear functions. Numbers of metaheuristic algorithms were used to solve this problem such as Genetic Algorithms (GAs) and Particle Swarm Optimization (PSO). In this paper, another meta-heuristic search algorithm named Differential Evolution (DE) is used to solve the ELD problem in power systems planning. The practicality of the proposed DE based algorithm is verified for three and six power generator system test cases. The gained results are compared to existing results based on QP, GAs and PSO. The developed results show that differential evolution is superior in obtaining a combination of power loads that fulfill the problem constraints and minimize the total fuel cost. DE found to be fast in converging to the optimal power generation loads and capable of handling the non-linearity of ELD problem. The proposed DE solution is able to minimize the cost of generated power, minimize the total power loss in the transmission and maximize the reliability of the power provided to the customers.Keywords: economic load dispatch, power systems, optimization, differential evolution
Procedia PDF Downloads 282329 DLtrace: Toward Understanding and Testing Deep Learning Information Flow in Deep Learning-Based Android Apps
Authors: Jie Zhang, Qianyu Guo, Tieyi Zhang, Zhiyong Feng, Xiaohong Li
Abstract:
With the widespread popularity of mobile devices and the development of artificial intelligence (AI), deep learning (DL) has been extensively applied in Android apps. Compared with traditional Android apps (traditional apps), deep learning based Android apps (DL-based apps) need to use more third-party application programming interfaces (APIs) to complete complex DL inference tasks. However, existing methods (e.g., FlowDroid) for detecting sensitive information leakage in Android apps cannot be directly used to detect DL-based apps as they are difficult to detect third-party APIs. To solve this problem, we design DLtrace; a new static information flow analysis tool that can effectively recognize third-party APIs. With our proposed trace and detection algorithms, DLtrace can also efficiently detect privacy leaks caused by sensitive APIs in DL-based apps. Moreover, using DLtrace, we summarize the non-sequential characteristics of DL inference tasks in DL-based apps and the specific functionalities provided by DL models for such apps. We propose two formal definitions to deal with the common polymorphism and anonymous inner-class problems in the Android static analyzer. We conducted an empirical assessment with DLtrace on 208 popular DL-based apps in the wild and found that 26.0% of the apps suffered from sensitive information leakage. Furthermore, DLtrace has a more robust performance than FlowDroid in detecting and identifying third-party APIs. The experimental results demonstrate that DLtrace expands FlowDroid in understanding DL-based apps and detecting security issues therein.Keywords: mobile computing, deep learning apps, sensitive information, static analysis
Procedia PDF Downloads 178328 INCIPIT-CRIS: A Research Information System Combining Linked Data Ontologies and Persistent Identifiers
Authors: David Nogueiras Blanco, Amir Alwash, Arnaud Gaudinat, René Schneider
Abstract:
At a time when the access to and the sharing of information are crucial in the world of research, the use of technologies such as persistent identifiers (PIDs), Current Research Information Systems (CRIS), and ontologies may create platforms for information sharing if they respond to the need of disambiguation of their data by assuring interoperability inside and between other systems. INCIPIT-CRIS is a continuation of the former INCIPIT project, whose goal was to set up an infrastructure for a low-cost attribution of PIDs with high granularity based on Archival Resource Keys (ARKs). INCIPIT-CRIS can be interpreted as a logical consequence and propose a research information management system developed from scratch. The system has been created on and around the Schema.org ontology with a further articulation of the use of ARKs. It is thus built upon the infrastructure previously implemented (i.e., INCIPIT) in order to enhance the persistence of URIs. As a consequence, INCIPIT-CRIS aims to be the hinge between previously separated aspects such as CRIS, ontologies and PIDs in order to produce a powerful system allowing the resolution of disambiguation problems using a combination of an ontology such as Schema.org and unique persistent identifiers such as ARK, allowing the sharing of information through a dedicated platform, but also the interoperability of the system by representing the entirety of the data as RDF triplets. This paper aims to present the implemented solution as well as its simulation in real life. We will describe the underlying ideas and inspirations while going through the logic and the different functionalities implemented and their links with ARKs and Schema.org. Finally, we will discuss the tests performed with our project partner, the Swiss Institute of Bioinformatics (SIB), by the use of large and real-world data sets.Keywords: current research information systems, linked data, ontologies, persistent identifier, schema.org, semantic web
Procedia PDF Downloads 135327 Rising STI Prevalence among MSM Clients in Calabar, Nigeria: A Call to Action
Authors: Ugoh Kelechi Melford, Anene O.
Abstract:
Introduction: Evidence has shown that there are increasing rates of new HIV and other STI infections occurring among Men who have Sex with Men (MSM) in Nigeria, with the prevalence 3 times higher than the general population as reported by the 2011 National Integrated Bio Behavioral Surveillance Survey. The poor state of health care and support services hinders our effort to control the high rates of these new infections among MSM. Methods: The Initiative for Improved Male Health (IMH-Initiative) works to provide a safe space for young MSM living with HIV to access comprehensive palliative care and support, as well as referrals for other services through drama and dance competitions. An STI assessment was conducted in IMH-Initiative’s Community Center in Calabar, for gay men and other MSM. An STI history was conducted for all clients who visited the community clinic specifically for HCT and STI counseling and referrals within a 5 month period, and their data were collated. Results: 61 MSM were diagnosed, and reported the following in the last 6 months. 49 where living with HIV. 46 had previous histories of untreated anal warts. 20 had previous histories of treated Gonorrhea by self-medication and herbs. 21 had untreated boils and rashes around the genitals. 10 clients where living with HIV, and reported untreated penile and rectal gonorrhea. All clients indicated that there were not comfortable discussing STI infections with staff of public hospitals. Conclusion: It is evident that a reasonable number of STI infections among MSM are not completely treated or ignored. This thereby increases the individual’s risk of HIV infection, and cripples HIV prevention programming in Nigeria. HIV programs targeting MSM must incorporate STI syndromic management, so as to increase access to non-stigmatized diagnosis and treatment of STIs. Also, access to STI drugs for clients cannot be overemphasized.Keywords: MSM, IBBSS, STI, IMH
Procedia PDF Downloads 332326 Taleb's Complexity Theory Concept of 'Antifragility' Has a Significant Contribution to Make to Positive Psychology as Applied to Wellbeing
Authors: Claudius Peter Van Wyk
Abstract:
Given the increasingly manifest phenomena, as described in complexity theory, of volatility, uncertainty, complexity and ambiguity (VUCA), Taleb's notion of 'antifragility, has a significant contribution to make to positive psychology applied to wellbeing. Antifragility is argued to be fundamentally different from the concepts of resiliency; as the ability to recover from failure, and robustness; as the ability to resist failure. Rather it describes the capacity to reorganise in the face of stress in such a way as to cope more effectively with systemic challenges. The concept, which has been applied in disciplines ranging from physics, molecular biology, planning, engineering, and computer science, can now be considered for its application in individual human and social wellbeing. There are strong correlations to Antonovsky's model of 'salutogenesis' in which an attitude and competencies are developed of transforming burdening factors into greater resourcefulness. We demonstrate, from the perspective of neuroscience, how technology measuring nervous system coherence can be coupled to acquired psychodynamic approaches to not only identify contextual stressors, utilise biofeedback instruments for facilitating greater coherence, but apply these insights to specific life stressors that compromise well-being. Employing an on-going case study with BMW South Africa, the neurological mapping is demonstrated together with 'reframing' and emotional anchoring techniques from neurolinguistic programming. The argument is contextualised in the discipline of psychoneuroimmunology which describes the stress pathways from the CNS and endocrine systems and their impact on immune function and the capacity to restore homeostasis.Keywords: antifragility, complexity, neuroscience, psychoneuroimmunology, salutogenesis, volatility
Procedia PDF Downloads 376325 Design and Implementation of Low-code Model-building Methods
Authors: Zhilin Wang, Zhihao Zheng, Linxin Liu
Abstract:
This study proposes a low-code model-building approach that aims to simplify the development and deployment of artificial intelligence (AI) models. With an intuitive way to drag and drop and connect components, users can easily build complex models and integrate multiple algorithms for training. After the training is completed, the system automatically generates a callable model service API. This method not only lowers the technical threshold of AI development and improves development efficiency but also enhances the flexibility of algorithm integration and simplifies the deployment process of models. The core strength of this method lies in its ease of use and efficiency. Users do not need to have a deep programming background and can complete the design and implementation of complex models with a simple drag-and-drop operation. This feature greatly expands the scope of AI technology, allowing more non-technical people to participate in the development of AI models. At the same time, the method performs well in algorithm integration, supporting many different types of algorithms to work together, which further improves the performance and applicability of the model. In the experimental part, we performed several performance tests on the method. The results show that compared with traditional model construction methods, this method can make more efficient use, save computing resources, and greatly shorten the model training time. In addition, the system-generated model service interface has been optimized for high availability and scalability, which can adapt to the needs of different application scenarios.Keywords: low-code, model building, artificial intelligence, algorithm integration, model deployment
Procedia PDF Downloads 29324 Optimization of Multi Commodities Consumer Supply Chain: Part 1-Modelling
Authors: Zeinab Haji Abolhasani, Romeo Marian, Lee Luong
Abstract:
This paper and its companions (Part II, Part III) will concentrate on optimizing a class of supply chain problems known as Multi- Commodities Consumer Supply Chain (MCCSC) problem. MCCSC problem belongs to production-distribution (P-D) planning category. It aims to determine facilities location, consumers’ allocation, and facilities configuration to minimize total cost (CT) of the entire network. These facilities can be manufacturer units (MUs), distribution centres (DCs), and retailers/end-users (REs) but not limited to them. To address this problem, three major tasks should be undertaken. At the first place, a mixed integer non-linear programming (MINP) mathematical model is developed. Then, system’s behaviors under different conditions will be observed using a simulation modeling tool. Finally, the most optimum solution (minimum CT) of the system will be obtained using a multi-objective optimization technique. Due to the large size of the problem, and the uncertainties in finding the most optimum solution, integration of modeling and simulation methodologies is proposed followed by developing new approach known as GASG. It is a genetic algorithm on the basis of granular simulation which is the subject of the methodology of this research. In part II, MCCSC is simulated using discrete-event simulation (DES) device within an integrated environment of SimEvents and Simulink of MATLAB® software package followed by a comprehensive case study to examine the given strategy. Also, the effect of genetic operators on the obtained optimal/near optimal solution by the simulation model will be discussed in part III.Keywords: supply chain, genetic algorithm, optimization, simulation, discrete event system
Procedia PDF Downloads 316323 Study of Mobile Game Addiction Using Electroencephalography Data Analysis
Authors: Arsalan Ansari, Muhammad Dawood Idrees, Maria Hafeez
Abstract:
Use of mobile phones has been increasing considerably over the past decade. Currently, it is one of the main sources of communication and information. Initially, mobile phones were limited to calls and messages, but with the advent of new technology smart phones were being used for many other purposes including video games. Despite of positive outcomes, addiction to video games on mobile phone has become a leading cause of psychological and physiological problems among many people. Several researchers examined the different aspects of behavior addiction with the use of different scales. Objective of this study is to examine any distinction between mobile game addicted and non-addicted players with the use of electroencephalography (EEG), based upon psycho-physiological indicators. The mobile players were asked to play a mobile game and EEG signals were recorded by BIOPAC equipment with AcqKnowledge as data acquisition software. Electrodes were places, following the 10-20 system. EEG was recorded at sampling rate of 200 samples/sec (12,000samples/min). EEG recordings were obtained from the frontal (Fp1, Fp2), parietal (P3, P4), and occipital (O1, O2) lobes of the brain. The frontal lobe is associated with behavioral control, personality, and emotions. The parietal lobe is involved in perception, understanding logic, and arithmetic. The occipital lobe plays a role in visual tasks. For this study, a 60 second time window was chosen for analysis. Preliminary analysis of the signals was carried out with Acqknowledge software of BIOPAC Systems. From the survey based on CGS manual study 2010, it was concluded that five participants out of fifteen were in addictive category. This was used as prior information to group the addicted and non-addicted by physiological analysis. Statistical analysis showed that by applying clustering analysis technique authors were able to categorize the addicted and non-addicted players specifically on theta frequency range of occipital area.Keywords: mobile game, addiction, psycho-physiology, EEG analysis
Procedia PDF Downloads 164322 Optimization Technique for the Contractor’s Portfolio in the Bidding Process
Authors: Taha Anjamrooz, Sareh Rajabi, Salwa Bheiry
Abstract:
Selection between the available projects in bidding processes for the contractor is one of the essential areas to concentrate on. It is important for the contractor to choose the right projects within its portfolio during the tendering stage based on certain criteria. It should align the bidding process with its origination strategies and goals as a screening process to have the right portfolio pool to start with. Secondly, it should set the proper framework and use a suitable technique in order to optimize its selection process for concertation purpose and higher efforts during the tender stage with goals of success and winning. In this research paper, a two steps framework proposed to increase the efficiency of the contractor’s bidding process and the winning chance of getting the new projects awarded. In this framework, initially, all the projects pass through the first stage screening process, in which the portfolio basket will be evaluated and adjusted in accordance with the organization strategies to the reduced version of the portfolio pool, which is in line with organization activities. In the second stage, the contractor uses linear programming to optimize the portfolio pool based on available resources such as manpower, light equipment, heavy equipment, financial capability, return on investment, and success rate of winning the bid. Therefore, this optimization model will assist the contractor in utilizing its internal resource to its maximum and increase its winning chance for the new project considering past experience with clients, built-relation between two parties, and complexity in the exertion of the projects. The objective of this research will be to increase the contractor's winning chance in the bidding process based on the success rate and expected return on investment.Keywords: bidding process, internal resources, optimization, contracting portfolio management
Procedia PDF Downloads 142321 Maximizing Profit Using Optimal Control by Exploiting the Flexibility in Thermal Power Plants
Authors: Daud Mustafa Minhas, Raja Rehan Khalid, Georg Frey
Abstract:
The next generation power systems are equipped with abundantly available free renewable energy resources (RES). During their low-cost operations, the price of electricity significantly reduces to a lower value, and sometimes it becomes negative. Therefore, it is recommended not to operate the traditional power plants (e.g. coal power plants) and to reduce the losses. In fact, it is not a cost-effective solution, because these power plants exhibit some shutdown and startup costs. Moreover, they require certain time for shutdown and also need enough pause before starting up again, increasing inefficiency in the whole power network. Hence, there is always a trade-off between avoiding negative electricity prices, and the startup costs of power plants. To exploit this trade-off and to increase the profit of a power plant, two main contributions are made: 1) introducing retrofit technology for state of art coal power plant; 2) proposing optimal control strategy for a power plant by exploiting different flexibility features. These flexibility features include: improving ramp rate of power plant, reducing startup time and lowering minimum load. While, the control strategy is solved as mixed integer linear programming (MILP), ensuring optimal solution for the profit maximization problem. Extensive comparisons are made considering pre and post-retrofit coal power plant having the same efficiencies under different electricity price scenarios. It concludes that if the power plant must remain in the market (providing services), more flexibility reflects direct economic advantage to the plant operator.Keywords: discrete optimization, power plant flexibility, profit maximization, unit commitment model
Procedia PDF Downloads 143320 Web-Based Learning in Nursing: The Sample of Delivery Lesson Program
Authors: Merve Kadioğlu, Nevin H. Şahin
Abstract:
Purpose: This research is organized to determine the influence of the web-based learning program. The program has been developed to gain information about normal delivery skill that is one of the topics of nursing students who take the woman health and illness. Material and Methods: The methodology of this study was applied as pre-test post-test single-group quasi-experimental. The pilot study consisted of 28 nursing student study groups who agreed to participate in the study. The findings were gathered via web-based technologies: student information form, information evaluation tests, Web Based Training Material Evaluation Scale and web-based learning environment feedback form. In the analysis of the data, the percentage, frequency and Wilcoxon Signed Ranks Test were used. The Web Based Instruction Program was developed in the light of full learning model, Mayer's research-based multimedia development principles and Gagne's Instructional Activities Model. Findings: The average scores of it was determined in accordance with the web-based educational material evaluation scale: ‘Instructional Suitability’ 4.45, ‘Suitability to Educational Program’ 4.48, ‘Visual Adequacy’ 4.53, ‘Programming Eligibility / Technical Adequacy’ 4.00. Also, the participants mentioned that the program is successful and useful. A significant difference was found between the pre-test and post-test results of the seven modules (p < 0.05). Results: According to pilot study data, the program was rated ‘very good’ by the study group. It was also found to be effective in increasing knowledge about normal labor.Keywords: normal delivery, web-based learning, nursing students, e-learning
Procedia PDF Downloads 178319 Efficiency of Robust Heuristic Gradient Based Enumerative and Tunneling Algorithms for Constrained Integer Programming Problems
Authors: Vijaya K. Srivastava, Davide Spinello
Abstract:
This paper presents performance of two robust gradient-based heuristic optimization procedures based on 3n enumeration and tunneling approach to seek global optimum of constrained integer problems. Both these procedures consist of two distinct phases for locating the global optimum of integer problems with a linear or non-linear objective function subject to linear or non-linear constraints. In both procedures, in the first phase, a local minimum of the function is found using the gradient approach coupled with hemstitching moves when a constraint is violated in order to return the search to the feasible region. In the second phase, in one optimization procedure, the second sub-procedure examines 3n integer combinations on the boundary and within hypercube volume encompassing the result neighboring the result from the first phase and in the second optimization procedure a tunneling function is constructed at the local minimum of the first phase so as to find another point on the other side of the barrier where the function value is approximately the same. In the next cycle, the search for the global optimum commences in both optimization procedures again using this new-found point as the starting vector. The search continues and repeated for various step sizes along the function gradient as well as that along the vector normal to the violated constraints until no improvement in optimum value is found. The results from both these proposed optimization methods are presented and compared with one provided by popular MS Excel solver that is provided within MS Office suite and other published results.Keywords: constrained integer problems, enumerative search algorithm, Heuristic algorithm, Tunneling algorithm
Procedia PDF Downloads 325