Search results for: highly scalable programming model
20632 Schedule a New Production Plan by Heuristic Methods
Authors: Hanife Merve Öztürk, Sıdıka Dalgan
Abstract:
In this project, a capacity analysis study is done at TAT A. Ş. Maret Plant. Production capacity of products which generate 80% of sales amount are determined. Obtained data entered the LEKIN Scheduling Program and we get production schedules by using heuristic methods. Besides heuristic methods, as mathematical model, disjunctive programming formulation is adapted to flexible job shop problems by adding a new constraint to find optimal schedule solution.Keywords: scheduling, flexible job shop problem, shifting bottleneck heuristic, mathematical modelling
Procedia PDF Downloads 40120631 God, The Master Programmer: The Relationship Between God and Computers
Authors: Mohammad Sabbagh
Abstract:
Anyone who reads the Torah or the Quran learns that GOD created everything that is around us, seen and unseen, in six days. Within HIS plan of creation, HE placed for us a key proof of HIS existence which is essentially computers and the ability to program them. Digital computer programming began with binary instructions, which eventually evolved to what is known as high-level programming languages. Any programmer in our modern time can attest that you are essentially giving the computer commands by words and when the program is compiled, whatever is processed as output is limited to what the computer was given as an ability and furthermore as an instruction. So one can deduce that GOD created everything around us with HIS words, programming everything around in six days, just like how we can program a virtual world on the computer. GOD did mention in the Quran that one day where GOD’s throne is, is 1000 years of what we count; therefore, one might understand that GOD spoke non-stop for 6000 years of what we count, and gave everything it’s the function, attributes, class, methods and interactions. Similar to what we do in object-oriented programming. Of course, GOD has the higher example, and what HE created is much more than OOP. So when GOD said that everything is already predetermined, it is because any input, whether physical, spiritual or by thought, is outputted by any of HIS creatures, the answer has already been programmed. Any path, any thought, any idea has already been laid out with a reaction to any decision an inputter makes. Exalted is GOD!. GOD refers to HIMSELF as The Fastest Accountant in The Quran; the Arabic word that was used is close to processor or calculator. If you create a 3D simulation of a supernova explosion to understand how GOD produces certain elements and fuses protons together to spread more of HIS blessings around HIS skies; in 2022 you are going to require one of the strongest, fastest, most capable supercomputers of the world that has a theoretical speed of 50 petaFLOPS to accomplish that. In other words, the ability to perform one quadrillion (1015) floating-point operations per second. A number a human cannot even fathom. To put in more of a perspective, GOD is calculating when the computer is going through those 50 petaFLOPS calculations per second and HE is also calculating all the physics of every atom and what is smaller than that in all the actual explosion, and it’s all in truth. When GOD said HE created the world in truth, one of the meanings a person can understand is that when certain things occur around you, whether how a car crashes or how a tree grows; there is a science and a way to understand it, and whatever programming or science you deduce from whatever event you observed, it can relate to other similar events. That is why GOD might have said in The Quran that it is the people of knowledge, scholars, or scientist that fears GOD the most! One thing that is essential for us to keep up with what the computer is doing and for us to track our progress along with any errors is we incorporate logging mechanisms and backups. GOD in The Quran said that ‘WE used to copy what you used to do’. Essentially as the world is running, think of it as an interactive movie that is being played out in front of you, in a full-immersive non-virtual reality setting. GOD is recording it, from every angle to every thought, to every action. This brings the idea of how scary the Day of Judgment will be when one might realize that it’s going to be a fully immersive video when we would be getting and reading our book.Keywords: programming, the Quran, object orientation, computers and humans, GOD
Procedia PDF Downloads 10720630 Impact of the Electricity Market Prices during the COVID-19 Pandemic on Energy Storage Operation
Authors: Marin Mandić, Elis Sutlović, Tonći Modrić, Luka Stanić
Abstract:
With the restructuring and deregulation of the power system, storage owners, generation companies or private producers can offer their multiple services on various power markets and earn income in different types of markets, such as the day-ahead, real-time, ancillary services market, etc. During the COVID-19 pandemic, electricity prices, as well as ancillary services prices, increased significantly. The optimization of the energy storage operation was performed using a suitable model for simulating the operation of a pumped storage hydropower plant under market conditions. The objective function maximizes the income earned through energy arbitration, regulation-up, regulation-down and spinning reserve services. The optimization technique used for solving the objective function is mixed integer linear programming (MILP). In numerical examples, the pumped storage hydropower plant operation has been optimized considering the already achieved hourly electricity market prices from Nord Pool for the pre-pandemic (2019) and the pandemic (2020 and 2021) years. The impact of the electricity market prices during the COVID-19 pandemic on energy storage operation is shown through the analysis of income, operating hours, reserved capacity and consumed energy for each service. The results indicate the role of energy storage during a significant fluctuation in electricity and services prices.Keywords: electrical market prices, electricity market, energy storage optimization, mixed integer linear programming (MILP) optimization
Procedia PDF Downloads 17320629 An Earth Mover’s Distance Algorithm Based DDoS Detection Mechanism in SDN
Authors: Yang Zhou, Kangfeng Zheng, Wei Ni, Ren Ping Liu
Abstract:
Software-defined networking (SDN) provides a solution for scalable network framework with decoupled control and data plane. However, this architecture also induces a particular distributed denial-of-service (DDoS) attack that can affect or even overwhelm the SDN network. DDoS attack detection problem has to date been mostly researched as entropy comparison problem. However, this problem lacks the utilization of SDN, and the results are not accurate. In this paper, we propose a DDoS attack detection method, which interprets DDoS detection as a signature matching problem and is formulated as Earth Mover’s Distance (EMD) model. Considering the feasibility and accuracy, we further propose to define the cost function of EMD to be a generalized Kullback-Leibler divergence. Simulation results show that our proposed method can detect DDoS attacks by comparing EMD values with the ones computed in the case without attacks. Moreover, our method can significantly increase the true positive rate of detection.Keywords: DDoS detection, EMD, relative entropy, SDN
Procedia PDF Downloads 33820628 Optimizing the Insertion of Renewables in the Colombian Power Sector
Authors: Felipe Henao, Yeny Rodriguez, Juan P. Viteri, Isaac Dyner
Abstract:
Colombia is rich in natural resources and greatly focuses on the exploitation of water for hydroelectricity purposes. Alternative cleaner energy sources, such as solar and wind power, have been largely neglected despite: a) its abundance, b) the complementarities between hydro, solar and wind power, and c) the cost competitiveness of renewable technologies. The current limited mix of energy sources creates considerable weaknesses for the system, particularly when facing extreme dry weather conditions, such as El Niño event. In the past, El Niño have exposed the truly consequences of a system heavily dependent on hydropower, i.e. loss of power supply, high energy production costs, and loss of overall competitiveness for the country. Nonetheless, it is expected that the participation of hydroelectricity will increase in the near future. In this context, this paper proposes a stochastic lineal programming model to optimize the insertion of renewable energy systems (RES) into the Colombian electricity sector. The model considers cost-based generation competition between traditional energy technologies and alternative RES. This work evaluates the financial, environmental, and technical implications of different combinations of technologies. Various scenarios regarding the future evolution of costs of the technologies are considered to conduct sensitivity analysis of the solutions – to assess the extent of the participation of the RES in the Colombian power sector. Optimization results indicate that, even in the worst case scenario, where costs remain constant, the Colombian power sector should diversify its portfolio of technologies and invest strongly in solar and wind power technologies. The diversification through RES will contribute to make the system less vulnerable to extreme weather conditions, reduce the overall system costs, cut CO2 emissions, and decrease the chances of having national blackout events in the future. In contrast, the business as usual scenario indicates that the system will turn more costly and less reliable.Keywords: energy policy and planning, stochastic programming, sustainable development, water management
Procedia PDF Downloads 29620627 Efficient DNN Training on Heterogeneous Clusters with Pipeline Parallelism
Abstract:
Pipeline parallelism has been widely used to accelerate distributed deep learning to alleviate GPU memory bottlenecks and to ensure that models can be trained and deployed smoothly under limited graphics memory conditions. However, in highly heterogeneous distributed clusters, traditional model partitioning methods are not able to achieve load balancing. The overlap of communication and computation is also a big challenge. In this paper, HePipe is proposed, an efficient pipeline parallel training method for highly heterogeneous clusters. According to the characteristics of the neural network model pipeline training task, oriented to the 2-level heterogeneous cluster computing topology, a training method based on the 2-level stage division of neural network modeling and partitioning is designed to improve the parallelism. Additionally, a multi-forward 1F1B scheduling strategy is designed to accelerate the training time of each stage by executing the computation units in advance to maximize the overlap between the forward propagation communication and backward propagation computation. Finally, a dynamic recomputation strategy based on task memory requirement prediction is proposed to improve the fitness ratio of task and memory, which improves the throughput of the cluster and solves the memory shortfall problem caused by memory differences in heterogeneous clusters. The empirical results show that HePipe improves the training speed by 1.6×−2.2× over the existing asynchronous pipeline baselines.Keywords: pipeline parallelism, heterogeneous cluster, model training, 2-level stage partitioning
Procedia PDF Downloads 1820626 Effects of Merging Personal and Social Responsibility with Sports Education Model on Students' Game Performance and Responsibility
Authors: Yi-Hsiang Pan, Chen-Hui Huang, Wei-Ting Hsu
Abstract:
The purposes of the study were to understand these topics as follows: 1. To explore the effect of merging teaching personal and social responsibility (TPSR) with sports education model on students' game performance and responsibility. 2. To explore the effect of sports education model on students' game performance and responsibility. 3. To compare the difference between "merging TPSR with sports education model" and "sports education model" on students' game performance and responsibility. The participants include three high school physical education teachers and six physical education classes. Every teacher teaches an experimental group and a control group. The participants had 121 students, including 65 students in the experimental group and 56 students in the control group. The research methods had game performance assessment, questionnaire investigation, interview, focus group meeting. The research instruments include personal and social responsibility questionnaire and game performance assessment instrument. Paired t-test test and MANCOVA were used to test the difference between "merging TPSR with sports education model" and "sports education model" on students' learning performance. 1) "Merging TPSR with sports education model" showed significant improvements in students' game performance, and responsibilities with self-direction, helping others, cooperation. 2) "Sports education model" also had significant improvements in students' game performance, and responsibilities with effort, self-direction, helping others. 3.) There was no significant difference in game performance and responsibilities between "merging TPSR with sports education model" and "sports education model". 4)."Merging TPSR with sports education model" significantly improve learning atmosphere and peer relationships, it may be developed in the physical education curriculum. The conclusions were as follows: Both "Merging TPSR with sports education model" and "sports education model" can help improve students' responsibility and game performance. However, "Merging TPSR with sports education model" can reduce the competitive atmosphere in highly intensive games between students. The curricular projects of hybrid TPSR-Sport Education model is a good approach for moral character education.Keywords: curriculum and teaching model, sports self-efficacy, sport enthusiastic, character education
Procedia PDF Downloads 31320625 Analyzing Changes in Runoff Patterns Due to Urbanization Using SWAT Models
Authors: Asawari Ajay Avhad
Abstract:
The Soil and Water Assessment Tool (SWAT) is a hydrological model designed to predict the complex interactions within natural and human-altered watersheds. This research applies the SWAT model to the Ulhas River basin, a small watershed undergoing urbanization and characterized by bowl-like topography. Three simulation scenarios (LC17, LC22, and LC27) are investigated, each representing different land use and land cover (LULC) configurations, to assess the impact of urbanization on runoff. The LULC for the year 2027 is generated using the MOLUSCE Plugin of QGIS, incorporating various spatial factors such as DEM, Distance from Road, Distance from River, Slope, and distance from settlements. Future climate data is simulated within the SWAT model using historical data spanning 30 years. A susceptibility map for runoff across the basin is created, classifying runoff into five susceptibility levels ranging from very low to very high. Sub-basins corresponding to major urban settlements are identified as highly susceptible to runoff. With consideration of future climate projections, a slight increase in runoff is forecasted. The reliability of the methodology was validated through the identification of sub-basins known for experiencing severe flood events, which were determined to be highly susceptible to runoff. The susceptibility map successfully pinpointed these sub-basins with a track record of extreme flood occurrences, thus reinforcing the credibility of the assessment methodology. This study suggests that the methodology employed could serve as a valuable tool in flood management planning.Keywords: future land use impact, flood management, run off prediction, ArcSWAT
Procedia PDF Downloads 4620624 Parameters Identification and Sensitivity Study for Abrasive WaterJet Milling Model
Authors: Didier Auroux, Vladimir Groza
Abstract:
This work is part of STEEP Marie-Curie ITN project, and it focuses on the identification of unknown parameters of the proposed generic Abrasive WaterJet Milling (AWJM) PDE model, that appears as an ill-posed inverse problem. The necessity of studying this problem comes from the industrial milling applications where the possibility to predict and model the final surface with high accuracy is one of the primary tasks in the absence of any knowledge of the model parameters that should be used. In this framework, we propose the identification of model parameters by minimizing a cost function, measuring the difference between experimental and numerical solutions. The adjoint approach based on corresponding Lagrangian gives the opportunity to find out the unknowns of the AWJM model and their optimal values that could be used to reproduce the required trench profile. Due to the complexity of the nonlinear problem and a large number of model parameters, we use an automatic differentiation software tool (TAPENADE) for the adjoint computations. By adding noise to the artificial data, we show that in fact the parameter identification problem is highly unstable and strictly depends on input measurements. Regularization terms could be effectively used to deal with the presence of data noise and to improve the identification correctness. Based on this approach we present results in 2D and 3D of the identification of the model parameters and of the surface prediction both with self-generated data and measurements obtained from the real production. Considering different types of model and measurement errors allows us to obtain acceptable results for manufacturing and to expect the proper identification of unknowns. This approach also gives us the ability to distribute the research on more complex cases and consider different types of model and measurement errors as well as 3D time-dependent model with variations of the jet feed speed.Keywords: Abrasive Waterjet Milling, inverse problem, model parameters identification, regularization
Procedia PDF Downloads 31620623 Green Supply Chain Design: A Mathematical Modeling Approach
Authors: Nusrat T. Chowdhury
Abstract:
Green Supply Chain Management (GSCM) is becoming a key to success for profitable businesses. The various activities contributing to carbon emissions in a supply chain are transportation, ordering and holding of inventory. This research work develops a mixed-integer nonlinear programming (MINLP) model that considers the scenario of a supply chain with multiple periods, multiple products and multiple suppliers. The model assumes that the demand is deterministic, the buyer has a limited storage space in each period, the buyer is responsible for the transportation cost, a supplier-dependent ordering cost applies for each period in which an order is placed on a supplier and inventory shortage is permissible. The model provides an optimal decision regarding what products to order, in what quantities, with which suppliers, and in which periods in order to maximize the profit. For the purpose of evaluating the carbon emissions, three different carbon regulating policies i.e., carbon cap-and-trade, the strict cap on carbon emission and carbon tax on emissions, have been considered. The proposed MINLP has been validated using a randomly generated data set.Keywords: green supply chain, carbon emission, mixed integer non-linear program, inventory shortage, carbon cap-and-trade
Procedia PDF Downloads 23920622 Heuristic for Scheduling Correlated Parallel Machine to Minimize Maximum Lateness and Total Weighed Completion Time
Authors: Yang-Kuei Lin, Yun-Xi Zhang
Abstract:
This research focuses on the bicriteria correlated parallel machine scheduling problem. The two objective functions considered in this problem are to minimize maximum lateness and total weighted completion time. We first present a mixed integer programming (MIP) model that can find the entire efficient frontier for the studied problem. Next, we have proposed a bicriteria heuristic that can find non-dominated solutions for the studied problem. The performance of the proposed bicriteria heuristic is compared with the efficient frontier generated by solving the MIP model. Computational results indicate that the proposed bicriteria heuristic can solve the problem efficiently and find a set of diverse solutions that are uniformly distributed along the efficient frontier.Keywords: bicriteria, correlated parallel machines, heuristic, scheduling
Procedia PDF Downloads 14120621 On the Role of Cutting Conditions on Surface Roughness in High-Speed Thread Milling of Brass C3600
Authors: Amir Mahyar Khorasani, Ian Gibson, Moshe Goldberg, Mohammad Masoud Movahedi, Guy Littlefair
Abstract:
One of the important factors in manufacturing processes especially machining operations is surface quality. Improving this parameter results in improving fatigue strength, corrosion resistance, creep life and surface friction. The reliability and clearance of removable joints such as thread and nuts are highly related to the surface roughness. In this work, the effect of different cutting parameters such as cutting fluid pressure, feed rate and cutting speed on the surface quality of the crest of thread in the high-speed milling of Brass C3600 have been determined. Two popular neural networks containing MLP and RBF coupling with Taguchi L32 have been used to model surface roughness which was shown to be highly adept for such tasks. The contribution of this work is modelling surface roughness on the crest of the thread by using precise profilometer with nanoscale resolution. Experimental tests have been carried out for validation and approved suitable accuracy of the proposed model. Also analysing the interaction of parameters two by two showed that the most effective cutting parameter on the surface value is feed rate followed by cutting speed and cutting fluid pressure.Keywords: artificial neural networks, cutting conditions, high-speed machining, surface roughness, thread milling
Procedia PDF Downloads 37720620 A Robust Software for Advanced Analysis of Space Steel Frames
Authors: Viet-Hung Truong, Seung-Eock Kim
Abstract:
This paper presents a robust software package for practical advanced analysis of space steel framed structures. The pre- and post-processors of the presented software package are coded in the C++ programming language while the solver is written by using the FORTRAN programming language. A user-friendly graphical interface of the presented software is developed to facilitate the modeling process and result interpretation of the problem. The solver employs the stability functions for capturing the second-order effects to minimize modeling and computational time. Both the plastic-hinge and fiber-hinge beam-column elements are available in the presented software. The generalized displacement control method is adopted to solve the nonlinear equilibrium equations.Keywords: advanced analysis, beam-column, fiber-hinge, plastic hinge, steel frame
Procedia PDF Downloads 30720619 FisherONE: Employing Distinct Pedagogy through Technology Integration in Senior Secondary Education
Authors: J. Kontoleon, D.Gall, M.Pidskalny
Abstract:
FisherONE offers a distinct pedagogic model for senior secondary education that integrates advanced technology to meet the learning needs of Year 11 and 12 students across Catholic schools in Queensland. As a fully online platform, FisherONE employs pedagogy that combines flexibility with personalized, data-driven learning. The model leverages tools like the MaxHub hybrid interactive system and AI-powered learning assistants to create tailored learning pathways that promote student autonomy and engagement. This paper examines FisherONE’s success in employing pedagogic strategies through technology. Initial findings suggest that students benefit from the blended approach of virtual assessments and real-time support, even as AI-assisted tools remain in the proof-of-concept phase. The study outlines how FisherONE plans to continue refining its educational methods to better serve students in distance learning environments, specifically in challenging subjects like physics. The integration of technology in FisherONE enhances the effectiveness of teaching and learning, addressing common challenges in online education by offering scalable, individualized learning experiences. This approach demonstrates the future potential of technology in education and the role it can play in fostering meaningful student outcomes.Keywords: AI-assisted learning, innovative pedagogy, personalized learning, senior education, technology in education
Procedia PDF Downloads 1820618 Cold Spray Fabrication of Coating for Highly Corrosive Environment
Authors: Harminder Singh
Abstract:
Cold spray is a novel and emerging technology for the fabrication of coating. In this study, coating is successfully developed by this process on superalloy surface. The selected coating composition is already proved as corrosion resistant. The microstructure of the newly developed coating is examined by various characterization techniques, for testing its suitability for high temperature corrosive conditions of waste incinerator. The energy producing waste incinerators are still running at low efficiency, mainly due to their chlorine based highly corrosive conditions. The characterization results show that the developed cold sprayed coating structure is suitable for its further testing in highly aggressive conditions.Keywords: coating, cold spray, corrosion, microstructure
Procedia PDF Downloads 39420617 Development of a One-Window Services Model for Accessing Cancer Immunotherapies
Authors: Rizwan Arshad, Alessio Panza, Nimra Inayat, Syeda Mariam Batool Kazmi, Shawana Azmat
Abstract:
The rapidly expanding use of immunotherapy for a wide range of cancers from late to early stages has, predictably, been accompanied by evidence of inequities in access to these highly effective but costly treatments. In this survey-based case study, we aimed to develop a One-window services model (OWSM) based on Anderson’s behavioral model to enhance competence in accessing cancer medications, particularly immunotherapies, through the analysis of 20 patient surveys conducted in the Armed forces bone marrow transplant center of the district, Rawalpindi from November to December 2022. The purposive sampling technique was used. Cronbach’s alpha coefficient was found to be 0.71. It was analyzed using SPSS version 26 with descriptive analysis, and results showed that the majority of the cancer patients were non-competent to access their prescribed cancer immunotherapy because of individual-level, socioeconomic, and organizational barriers.Keywords: cancer immunotherapy, one-window services model, accessibility, competence
Procedia PDF Downloads 7620616 Seismic Bearing Capacity Estimation of Shallow Foundations on Dense Sand Underlain by Loose Sand Strata by Using Finite Elements Limit Analysis
Authors: Pragyan Paramita Das, Vishwas N. Khatri
Abstract:
By using the lower- and upper- bound finite elements to limit analysis in conjunction with second-order conic programming (SOCP), the effect of seismic forces on the bearing capacity of surface strip footing resting on dense sand underlain by loose sand deposit is explored. The soil is assumed to obey the Mohr-Coulomb’s yield criterion and an associated flow rule. The angle of internal friction (ϕ) of the top and the bottom layer is varied from 42° to 44° and 32° to 34° respectively. The coefficient of seismic acceleration is varied from 0 to 0.3. The variation of bearing capacity with different thickness of top layer for various seismic acceleration coefficients is generated. A comparison will be made with the available solutions from literature wherever applicable.Keywords: bearing capacity, conic programming, finite elements, seismic forces
Procedia PDF Downloads 17020615 Eco-Design of Construction Industrial Park in China with Selection of Candidate Tenants
Authors: Yang Zhou, Kaijian Li, Guiwen Liu
Abstract:
Offsite construction is an innovative alternative to conventional site-based construction, with wide-ranging benefits. It requires building components, elements or modules were prefabricated and pre-assembly before installed into their final locations. To improve efficiency and achieve synergies, in recent years, construction companies were clustered into construction industrial parks (CIPs) in China. A CIP is a community of construction manufacturing and service businesses located together on a common property. Companies involved in industrial clusters can obtain environment and economic benefits by sharing resources and information in a given region. Therefore, the concept of industrial symbiosis (IS) can be applied to the traditional CIP to achieve sustainable industrial development or redevelopment through the implementation of eco-industrial parks (EIP). However, before designing a symbiosis network between companies in a CIP, candidate support tenants need to be selected to complement the existing construction companies. In this study, an access indicator system and a linear programming model are established to select candidate tenants in a CIP while satisfying the degree of connectivity among the enterprises in the CIP, minimizing the environmental impact, and maximizing the annualized profit of the CIP. The access indicator system comprises three primary indicators and fifteen secondary indicators, is proposed from the perspective of park-based level. The fifteen indicators are classified as three primary indicators including industrial symbiosis, environment performance and economic benefit, according to the three dimensions of sustainability (environment, economic and social dimensions) and the three R's of the environment (reduce, reuse and recycle). The linear programming model is a method to assess the satisfactoriness of all the indicators and to make an optimal multi-objective selection among candidate tenants. This method provides a practical tool for planners of a CIP in evaluating which among the candidate tenants would best complement existing anchor construction tenants. The reasonability and validity of the indicator system and the method is worth further study in the future.Keywords: construction industrial park, China, industrial symbiosis, offsite construction, selection of support tenants
Procedia PDF Downloads 27420614 Utility Analysis of API Economy Based on Multi-Sided Platform Markets Model
Authors: Mami Sugiura, Shinichi Arakawa, Masayuki Murata, Satoshi Imai, Toru Katagiri, Motoyoshi Sekiya
Abstract:
API (Application Programming Interface) economy, where many participants join/interact and form the economy, is expected to increase collaboration between information services through API, and thereby, it is expected to increase market value from the service collaborations. In this paper, we introduce API evaluators, which are the activator of API economy by reviewing and/or evaluating APIs, and develop a multi-sided API economy model that formulates interactions among platform provider, API developers, consumers, and API evaluators. By obtaining the equilibrium that maximizes utility of all participants, the impact of API evaluators on the utility of participants in the API economy is revealed. Numerical results show that, with the existence of API evaluators, the number of developers and consumers increase by 1.5% and the utility of platformer increases by 2.3%. We also discuss the strategies of platform provider to maximize its utility under the existence of API evaluators.Keywords: API economy, multi-sided markets, API evaluator, platform, platform provider
Procedia PDF Downloads 18620613 Review of Theories and Applications of Genetic Programing in Sediment Yield Modeling
Authors: Adesoji Tunbosun Jaiyeola, Josiah Adeyemo
Abstract:
Sediment yield can be considered to be the total sediment load that leaves a drainage basin. The knowledge of the quantity of sediments present in a river at a particular time can lead to better flood capacity in reservoirs and consequently help to control over-bane flooding. Furthermore, as sediment accumulates in the reservoir, it gradually loses its ability to store water for the purposes for which it was built. The development of hydrological models to forecast the quantity of sediment present in a reservoir helps planners and managers of water resources systems, to understand the system better in terms of its problems and alternative ways to address them. The application of artificial intelligence models and technique to such real-life situations have proven to be an effective approach of solving complex problems. This paper makes an extensive review of literature relevant to the theories and applications of evolutionary algorithms, and most especially genetic programming. The successful applications of genetic programming as a soft computing technique were reviewed in sediment modelling and other branches of knowledge. Some fundamental issues such as benchmark, generalization ability, bloat and over-fitting and other open issues relating to the working principles of GP, which needs to be addressed by the GP community were also highlighted. This review aim to give GP theoreticians, researchers and the general community of GP enough research direction, valuable guide and also keep all stakeholders abreast of the issues which need attention during the next decade for the advancement of GP.Keywords: benchmark, bloat, generalization, genetic programming, over-fitting, sediment yield
Procedia PDF Downloads 44620612 Lightweight Sheet Molding Compound Composites by Coating Glass Fiber with Cellulose Nanocrystals
Authors: Amir Asadi, Karim Habib, Robert J. Moon, Kyriaki Kalaitzidou
Abstract:
There has been considerable interest in cellulose nanomaterials (CN) as polymer and polymer composites reinforcement due to their high specific modulus and strength, low density and toxicity, and accessible hydroxyl side groups that can be readily chemically modified. The focus of this study is making lightweight composites for better fuel efficiency and lower CO2 emission in auto industries with no compromise on mechanical performance using a scalable technique that can be easily integrated in sheet molding compound (SMC) manufacturing lines. Light weighting will be achieved by replacing part of the heavier components, i.e. glass fibers (GF), with a small amount of cellulose nanocrystals (CNC) in short GF/epoxy composites made using SMC. CNC will be introduced as coating of the GF rovings prior to their use in the SMC line. The employed coating method is similar to the fiber sizing technique commonly used and thus it can be easily scaled and integrated to industrial SMC lines. This will be an alternative route to the most techniques that involve dispersing CN in polymer matrix, in which the nanomaterials agglomeration limits the capability for scaling up in an industrial production. We have demonstrated that incorporating CNC as a coating on GF surface by immersing the GF in CNC aqueous suspensions, a simple and scalable technique, increases the interfacial shear strength (IFSS) by ~69% compared to the composites produced by uncoated GF, suggesting an enhancement of stress transfer across the GF/matrix interface. As a result of IFSS enhancement, incorporation of 0.17 wt% CNC in the composite results in increases of ~10% in both elastic modulus and tensile strength, and 40 % and 43 % in flexural modulus and strength respectively. We have also determined that dispersing 1.4 and 2 wt% CNC in the epoxy matrix of short GF/epoxy SMC composites by sonication allows removing 10 wt% GF with no penalty on tensile and flexural properties leading to 7.5% lighter composites. Although sonication is a scalable technique, it is not quite as simple and inexpensive as coating the GF by passing through an aqueous suspension of CNC. In this study, the above findings are integrated to 1) investigate the effect of CNC content on mechanical properties by passing the GF rovings through CNC aqueous suspension with various concentrations (0-5%) and 2) determine the optimum ratio of the added CNC to the removed GF to achieve the maximum possible weight reduction with no cost on mechanical performance of the SMC composites. The results of this study are of industrial relevance, providing a path toward producing high volume lightweight and mechanically enhanced SMC composites using cellulose nanomaterials.Keywords: cellulose nanocrystals, light weight polymer-matrix composites, mechanical properties, sheet molding compound (SMC)
Procedia PDF Downloads 22520611 Modeling of Compaction Curves for CCA-Cement Stabilized Lateritic Soils
Authors: O. Ahmed Apampa, Yinusa, A. Jimoh
Abstract:
The aim of this study was to develop an appropriate model for predicting the compaction behavior of lateritic soils and corn cob ash (CCA) stabilized lateritic soils. This was done by first adopting an equation earlier developed for fine-grained soils and subsequent adaptation by others and extending it to modified lateritic soil through the introduction of alpha and beta parameters which are polynomial functions of the CCA binder input. The polynomial equations were determined with MATLAB R2011 curve fitting tool, while the alpha and beta parameters were determined by standard linear programming techniques using the Solver function of Microsoft Excel 2010. The model so developed was a good fit with a correlation coefficient R2 value of 0.86. The paper concludes that it is possible to determine the optimum moisture content and the maximum dry density of CCA stabilized soils from the compaction test of the unmodified soil, and recommends that this procedure is extended to other binder stabilized lateritic soils to facilitate quick decision making in roadworks.Keywords: compaction, corn cob ash, lateritic soil, stabilization
Procedia PDF Downloads 53320610 BIM-Based Tool for Sustainability Assessment and Certification Documents Provision
Authors: Taki Eddine Seghier, Mohd Hamdan Ahmad, Yaik-Wah Lim, Samuel Opeyemi Williams
Abstract:
The assessment of building sustainability to achieve a specific green benchmark and the preparation of the required documents in order to receive a green building certification, both are considered as major challenging tasks for green building design team. However, this labor and time-consuming process can take advantage of the available Building Information Modeling (BIM) features such as material take-off and scheduling. Furthermore, the workflow can be automated in order to track potentially achievable credit points and provide rating feedback for several design options by using integrated Visual Programing (VP) to handle the stored parameters within the BIM model. Hence, this study proposes a BIM-based tool that uses Green Building Index (GBI) rating system requirements as a unique input case to evaluate the building sustainability in the design stage of the building project life cycle. The tool covers two key models for data extraction, firstly, a model for data extraction, calculation and the classification of achievable credit points in a green template, secondly, a model for the generation of the required documents for green building certification. The tool was validated on a BIM model of residential building and it serves as proof of concept that building sustainability assessment of GBI certification can be automatically evaluated and documented through BIM.Keywords: green building rating system, GBRS, building information modeling, BIM, visual programming, VP, sustainability assessment
Procedia PDF Downloads 32620609 Logistic Regression Model versus Additive Model for Recurrent Event Data
Authors: Entisar A. Elgmati
Abstract:
Recurrent infant diarrhea is studied using daily data collected in Salvador, Brazil over one year and three months. A logistic regression model is fitted instead of Aalen's additive model using the same covariates that were used in the analysis with the additive model. The model gives reasonably similar results to that using additive regression model. In addition, the problem with the estimated conditional probabilities not being constrained between zero and one in additive model is solved here. Also martingale residuals that have been used to judge the goodness of fit for the additive model are shown to be useful for judging the goodness of fit of the logistic model.Keywords: additive model, cumulative probabilities, infant diarrhoea, recurrent event
Procedia PDF Downloads 63520608 Formal Verification for Ethereum Smart Contract Using Coq
Authors: Xia Yang, Zheng Yang, Haiyong Sun, Yan Fang, Jingyu Liu, Jia Song
Abstract:
The smart contract in Ethereum is a unique program deployed on the Ethereum Virtual Machine (EVM) to help manage cryptocurrency. The security of this smart contract is critical to Ethereum’s operation and highly sensitive. In this paper, we present a formal model for smart contract, using the separated term-obligation (STO) strategy to formalize and verify the smart contract. We use the IBM smart sponsor contract (SSC) as an example to elaborate the detail of the formalizing process. We also propose a formal smart sponsor contract model (FSSCM) and verify SSC’s security properties with an interactive theorem prover Coq. We found the 'Unchecked-Send' vulnerability in the SSC, using our formal model and verification method. Finally, we demonstrate how we can formalize and verify other smart contracts with this approach, and our work indicates that this formal verification can effectively verify the correctness and security of smart contracts.Keywords: smart contract, formal verification, Ethereum, Coq
Procedia PDF Downloads 69120607 An Optimal Control Model to Determine Body Forces of Stokes Flow
Authors: Yuanhao Gao, Pin Lin, Kees Weijer
Abstract:
In this paper, we will determine the external body force distribution with analysis of stokes fluid motion using mathematical modelling and numerical approaching. The body force distribution is regarded as the unknown variable and could be determined by the idea of optimal control theory. The Stokes flow motion and its velocity are generated by given forces in a unit square domain. A regularized objective functional is built to match the numerical result of flow velocity with the generated velocity data. So that the force distribution could be determined by minimizing the value of objective functional, which is also the difference between the numerical and experimental velocity. Then after utilizing the Lagrange multiplier method, some partial differential equations are formulated consisting the optimal control system to solve. Finite element method and conjugate gradient method are used to discretize equations and deduce the iterative expression of target body force to compute the velocity numerically and body force distribution. Programming environment FreeFEM++ supports the implementation of this model.Keywords: optimal control model, Stokes equation, finite element method, conjugate gradient method
Procedia PDF Downloads 40520606 Estimation of Transition and Emission Probabilities
Authors: Aakansha Gupta, Neha Vadnere, Tapasvi Soni, M. Anbarsi
Abstract:
Protein secondary structure prediction is one of the most important goals pursued by bioinformatics and theoretical chemistry; it is highly important in medicine and biotechnology. Some aspects of protein functions and genome analysis can be predicted by secondary structure prediction. This is used to help annotate sequences, classify proteins, identify domains, and recognize functional motifs. In this paper, we represent protein secondary structure as a mathematical model. To extract and predict the protein secondary structure from the primary structure, we require a set of parameters. Any constants appearing in the model are specified by these parameters, which also provide a mechanism for efficient and accurate use of data. To estimate these model parameters there are many algorithms out of which the most popular one is the EM algorithm or called the Expectation Maximization Algorithm. These model parameters are estimated with the use of protein datasets like RS126 by using the Bayesian Probabilistic method (data set being categorical). This paper can then be extended into comparing the efficiency of EM algorithm to the other algorithms for estimating the model parameters, which will in turn lead to an efficient component for the Protein Secondary Structure Prediction. Further this paper provides a scope to use these parameters for predicting secondary structure of proteins using machine learning techniques like neural networks and fuzzy logic. The ultimate objective will be to obtain greater accuracy better than the previously achieved.Keywords: model parameters, expectation maximization algorithm, protein secondary structure prediction, bioinformatics
Procedia PDF Downloads 48020605 Context Detection in Spreadsheets Based on Automatically Inferred Table Schema
Authors: Alexander Wachtel, Michael T. Franzen, Walter F. Tichy
Abstract:
Programming requires years of training. With natural language and end user development methods, programming could become available to everyone. It enables end users to program their own devices and extend the functionality of the existing system without any knowledge of programming languages. In this paper, we describe an Interactive Spreadsheet Processing Module (ISPM), a natural language interface to spreadsheets that allows users to address ranges within the spreadsheet based on inferred table schema. Using the ISPM, end users are able to search for values in the schema of the table and to address the data in spreadsheets implicitly. Furthermore, it enables them to select and sort the spreadsheet data by using natural language. ISPM uses a machine learning technique to automatically infer areas within a spreadsheet, including different kinds of headers and data ranges. Since ranges can be identified from natural language queries, the end users can query the data using natural language. During the evaluation 12 undergraduate students were asked to perform operations (sum, sort, group and select) using the system and also Excel without ISPM interface, and the time taken for task completion was compared across the two systems. Only for the selection task did users take less time in Excel (since they directly selected the cells using the mouse) than in ISPM, by using natural language for end user software engineering, to overcome the present bottleneck of professional developers.Keywords: natural language processing, natural language interfaces, human computer interaction, end user development, dialog systems, data recognition, spreadsheet
Procedia PDF Downloads 31120604 Evolving Credit Scoring Models using Genetic Programming and Language Integrated Query Expression Trees
Authors: Alexandru-Ion Marinescu
Abstract:
There exist a plethora of methods in the scientific literature which tackle the well-established task of credit score evaluation. In its most abstract form, a credit scoring algorithm takes as input several credit applicant properties, such as age, marital status, employment status, loan duration, etc. and must output a binary response variable (i.e. “GOOD” or “BAD”) stating whether the client is susceptible to payment return delays. Data imbalance is a common occurrence among financial institution databases, with the majority being classified as “GOOD” clients (clients that respect the loan return calendar) alongside a small percentage of “BAD” clients. But it is the “BAD” clients we are interested in since accurately predicting their behavior is crucial in preventing unwanted loss for loan providers. We add to this whole context the constraint that the algorithm must yield an actual, tractable mathematical formula, which is friendlier towards financial analysts. To this end, we have turned to genetic algorithms and genetic programming, aiming to evolve actual mathematical expressions using specially tailored mutation and crossover operators. As far as data representation is concerned, we employ a very flexible mechanism – LINQ expression trees, readily available in the C# programming language, enabling us to construct executable pieces of code at runtime. As the title implies, they model trees, with intermediate nodes being operators (addition, subtraction, multiplication, division) or mathematical functions (sin, cos, abs, round, etc.) and leaf nodes storing either constants or variables. There is a one-to-one correspondence between the client properties and the formula variables. The mutation and crossover operators work on a flattened version of the tree, obtained via a pre-order traversal. A consequence of our chosen technique is that we can identify and discard client properties which do not take part in the final score evaluation, effectively acting as a dimensionality reduction scheme. We compare ourselves with state of the art approaches, such as support vector machines, Bayesian networks, and extreme learning machines, to name a few. The data sets we benchmark against amount to a total of 8, of which we mention the well-known Australian credit and German credit data sets, and the performance indicators are the following: percentage correctly classified, area under curve, partial Gini index, H-measure, Brier score and Kolmogorov-Smirnov statistic, respectively. Finally, we obtain encouraging results, which, although placing us in the lower half of the hierarchy, drive us to further refine the algorithm.Keywords: expression trees, financial credit scoring, genetic algorithm, genetic programming, symbolic evolution
Procedia PDF Downloads 11720603 Evaluation of Model-Based Code Generation for Embedded Systems–Mature Approach for Development in Evolution
Authors: Nikolay P. Brayanov, Anna V. Stoynova
Abstract:
Model-based development approach is gaining more support and acceptance. Its higher abstraction level brings simplification of systems’ description that allows domain experts to do their best without particular knowledge in programming. The different levels of simulation support the rapid prototyping, verifying and validating the product even before it exists physically. Nowadays model-based approach is beneficial for modelling of complex embedded systems as well as a generation of code for many different hardware platforms. Moreover, it is possible to be applied in safety-relevant industries like automotive, which brings extra automation of the expensive device certification process and especially in the software qualification. Using it, some companies report about cost savings and quality improvements, but there are others claiming no major changes or even about cost increases. This publication demonstrates the level of maturity and autonomy of model-based approach for code generation. It is based on a real live automotive seat heater (ASH) module, developed using The Mathworks, Inc. tools. The model, created with Simulink, Stateflow and Matlab is used for automatic generation of C code with Embedded Coder. To prove the maturity of the process, Code generation advisor is used for automatic configuration. All additional configuration parameters are set to auto, when applicable, leaving the generation process to function autonomously. As a result of the investigation, the publication compares the quality of generated embedded code and a manually developed one. The measurements show that generally, the code generated by automatic approach is not worse than the manual one. A deeper analysis of the technical parameters enumerates the disadvantages, part of them identified as topics for our future work.Keywords: embedded code generation, embedded C code quality, embedded systems, model-based development
Procedia PDF Downloads 244