Search results for: regularization parameter search
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 3876

Search results for: regularization parameter search

1986 An Optimized Method for 3D Magnetic Navigation of Nanoparticles inside Human Arteries

Authors: Evangelos G. Karvelas, Christos Liosis, Andreas Theodorakakos, Theodoros E. Karakasidis

Abstract:

In the present work, a numerical method for the estimation of the appropriate gradient magnetic fields for optimum driving of the particles into the desired area inside the human body is presented. The proposed method combines Computational Fluid Dynamics (CFD), Discrete Element Method (DEM) and Covariance Matrix Adaptation (CMA) evolution strategy for the magnetic navigation of nanoparticles. It is based on an iteration procedure that intents to eliminate the deviation of the nanoparticles from a desired path. Hence, the gradient magnetic field is constantly adjusted in a suitable way so that the particles’ follow as close as possible to a desired trajectory. Using the proposed method, it is obvious that the diameter of particles is crucial parameter for an efficient navigation. In addition, increase of particles' diameter decreases their deviation from the desired path. Moreover, the navigation method can navigate nanoparticles into the desired areas with efficiency approximately 99%.

Keywords: computational fluid dynamics, CFD, covariance matrix adaptation evolution strategy, discrete element method, DEM, magnetic navigation, spherical particles

Procedia PDF Downloads 124
1985 Human Kinetics Education and the Computer Operations, Effects and Merits

Authors: Kehinde Adeyeye Adelabu

Abstract:

Computer applications has completely revolutionized the way of life of people which does not exclude the field of sport education. There are computer technologies which help to enhance teaching in every field of education. Invention of computers has done great to the field of education. This study was therefore carried out to examine the effects and merits of computer operations in Human Kinetics Education and Sports. The study was able to identify the component of computer, uses of computer in Human Kinetics education (sports), computer applications in some branches of human kinetics education. A qualitative research method was employed by the author in gathering experts’ views and used to analyze the effects and merits of computer applications in the field of human kinetics education. No experiment was performed in the cause of carrying out the study. The source of information for the study was text-books, journal, articles, past project reports, internet i.e. Google search engine. Computer has significantly helped to improve Education (Human Kinetic), it has complemented the basic physical fitness testing and gave a more scientific basis to the testing. The use of the software and packages has made cost projections, database applications, inventory control, management of events, word processing, electronic mailing and record keeping easier than the pasts.

Keywords: application, computer operation, education, human kinetics

Procedia PDF Downloads 163
1984 Case Study Analysis for Driver's Company in the Transport Sector with the Help of Data Mining

Authors: Diana Katherine Gonzalez Galindo, David Rolando Suarez Mora

Abstract:

With this study, we used data mining as a new alternative of the solution to evaluate the comments of the customers in order to find a pattern that helps us to determine some behaviors to reduce the deactivation of the partners of the LEVEL app. In one of the greatest business created in the last times, the partners are being affected due to an internal process that compensates the customer for a bad experience, but these comments could be false towards the driver, that’s why we made an investigation to collect information to restructure this process, many partners have been disassociated due to this internal process and many of them refuse the comments given by the customer. The main methodology used in this case study is the observation, we recollect information in real time what gave us the opportunity to see the most common issues to get the most accurate solution. With this new process helped by data mining, we could get a prediction based on the behaviors of the customer and some basic data recollected such as the age, the gender, and others; this could help us in future to improve another process. This investigation gives more opportunities to the partner to keep his account active even if the customer writes a message through the app. The term is trying to avoid a recession of drivers in the future offering improving in the processes, at the same time we are in search of stablishing a strategy which benefits both the app’s managers and the associated driver.

Keywords: agent, driver, deactivation, rider

Procedia PDF Downloads 266
1983 Adsorption of Cd(II) and Pb(II) from Aqueous Solutions by Using Pods of Acacia Karoo

Authors: Gulshan Kumar Jawa, Sandeep Mohan Ahuja

Abstract:

With the increase in industrialization, the presence of heavy metals in wastewater streams has turned into a serious concern for the ecosystem. The metals diffuse through the food chains, causing various health hazards. Conventional methods used to remove these heavy metals from water have some limitations, such as cost, secondary pollution due to sludge formation, recovery of metal, economic viability at low metal concentrations, etc. Many of the biomaterials have been investigated by researchers for the adsorption of heavy metals from water solutions as an alternative technique for the last two decades and have found promising results. In this paper, the batch study on the use of pods of acacia karoo for the adsorption of Cd(II) and Pb(II) from aqueous solutions has been reported. The effect of various parameters on the removal of metal ions, such as pH, contact time, stirring speed, initial metal ion concentration, adsorbent dose, and temperature, have been established to find the optimum parameters through one parameter optimization. Further, kinetic, equilibrium, and thermodynamic studies have been conducted. The pods of acacia karoo have shown great potential for adsorption of Cd(II) and Pb(II) from aqueous solutions and have proven to be a better and more economical alternative for the purpose.

Keywords: adsorption, heavy metals, biomaterials, Cadmium(II), Lead(II), pods of acacia karoo

Procedia PDF Downloads 21
1982 An Analysis of Economical Drivers and Technical Challenges for Large-Scale Biohydrogen Deployment

Authors: Rouzbeh Jafari, Joe Nava

Abstract:

This study includes learnings from an engineering practice normally performed on large scale biohydrogen processes. If properly scale-up is done, biohydrogen can be a reliable pathway for biowaste valorization. Most of the studies on biohydrogen process development have used model feedstock to investigate process key performance indicators (KPIs). This study does not intend to compare different technologies with model feedstock. However, it reports economic drivers and technical challenges which help in developing a road map for expanding biohydrogen economy deployment in Canada. BBA is a consulting firm responsible for the design of hydrogen production projects. Through executing these projects, activity has been performed to identify, register and mitigate technical drawbacks of large-scale hydrogen production. Those learnings, in this study, have been applied to the biohydrogen process. Through data collected by a comprehensive literature review, a base case has been considered as a reference, and several case studies have been performed. Critical parameters of the process were identified and through common engineering practice (process design, simulation, cost estimate, and life cycle assessment) impact of these parameters on the commercialization risk matrix and class 5 cost estimations were reported. The process considered in this study is food waste and woody biomass dark fermentation. To propose a reliable road map to develop a sustainable biohydrogen production process impact of critical parameters was studied on the end-to-end process. These parameters were 1) feedstock composition, 2) feedstock pre-treatment, 3) unit operation selection, and 4) multi-product concept. A couple of emerging technologies also were assessed such as photo-fermentation, integrated dark fermentation, and using ultrasound and microwave to break-down feedstock`s complex matrix and increase overall hydrogen yield. To properly report the impact of each parameter KPIs were identified as 1) Hydrogen yield, 2) energy consumption, 3) secondary waste generated, 4) CO2 footprint, 5) Product profile, 6) $/kg-H2 and 5) environmental impact. The feedstock is the main parameter defining the economic viability of biohydrogen production. Through parametric studies, it was found that biohydrogen production favors feedstock with higher carbohydrates. The feedstock composition was varied, by increasing one critical element (such as carbohydrate) and monitoring KPIs evolution. Different cases were studied with diverse feedstock, such as energy crops, wastewater slug, and lignocellulosic waste. The base case process was applied to have reference KPIs values and modifications such as pretreatment and feedstock mix-and-match were implemented to investigate KPIs changes. The complexity of the feedstock is the main bottleneck in the successful commercial deployment of the biohydrogen process as a reliable pathway for waste valorization. Hydrogen yield, reaction kinetics, and performance of key unit operations highly impacted as feedstock composition fluctuates during the lifetime of the process or from one case to another. In this case, concept of multi-product becomes more reliable. In this concept, the process is not designed to produce only one target product such as biohydrogen but will have two or multiple products (biohydrogen and biomethane or biochemicals). This new approach is being investigated by the BBA team and the results will be shared in another scientific contribution.

Keywords: biohydrogen, process scale-up, economic evaluation, commercialization uncertainties, hydrogen economy

Procedia PDF Downloads 85
1981 Deployment of Electronic Healthcare Records and Development of Big Data Analytics Capabilities in the Healthcare Industry: A Systematic Literature Review

Authors: Tigabu Dagne Akal

Abstract:

Electronic health records (EHRs) can help to store, maintain, and make the appropriate handling of patient histories for proper treatment and decision. Merging the EHRs with big data analytics (BDA) capabilities enable healthcare stakeholders to provide effective and efficient treatments for chronic diseases. Though there are huge opportunities and efforts that exist in the deployment of EMRs and the development of BDA, there are challenges in addressing resources and organizational capabilities that are required to achieve the competitive advantage and sustainability of EHRs and BDA. The resource-based view (RBV), information system (IS), and non- IS theories should be extended to examine organizational capabilities and resources which are required for successful data analytics in the healthcare industries. The main purpose of this study is to develop a conceptual framework for the development of healthcare BDA capabilities based on past works so that researchers can extend. The research question was formulated for the search strategy as a research methodology. The study selection was made at the end. Based on the study selection, the conceptual framework for the development of BDA capabilities in the healthcare settings was formulated.

Keywords: EHR, EMR, Big data, Big data analytics, resource-based view

Procedia PDF Downloads 120
1980 Relation between Pavement Roughness and Distress Parameters for Highways

Authors: Suryapeta Harini

Abstract:

Road surface roughness is one of the essential aspects of the road's functional condition, indicating riding comfort in both the transverse and longitudinal directions. The government of India has made maintaining good surface evenness a prerequisite for all highway projects. Pavement distress data was collected with a Network Survey Vehicle (NSV) on a National Highway. It determines the smoothness and frictional qualities of the pavement surface, which are related to driving safety and ease. Based on the data obtained in the field, a regression equation was created with the IRI value and the visual distresses. The suggested system can use wireless acceleration sensors and GPS to gather vehicle status and location data, as well as calculate the international roughness index (IRI). Potholes, raveling, rut depth, cracked area, and repair work are all affected by pavement roughness, according to the current study. The study was carried out in one location. Data collected through using Bump integrator was used for the validation. The bump integrator (BI) obtained using deflection from the network survey vehicle was correlated with the distress parameter to establish an equation.

Keywords: roughness index, network survey vehicle, regression, correlation

Procedia PDF Downloads 160
1979 Analyzing the Feasibility of Low-Cost Composite Wind Turbine Blades for Residential Energy Production

Authors: Aravindhan Nepolean, Chidamabaranathan Bibin, Rajesh K., Gopinath S., Ashok Kumar R., Arun Kumar S., Sadasivan N.

Abstract:

Wind turbine blades are an important parameter for surging renewable energy production. Optimizing blade profiles and developing new materials for wind turbine blades take a lot of time and effort. Even though many standards for wind turbine blades have been developed for large-scale applications, they are not more effective in small-scale applications. We used acrylonitrile-butadiene-styrene to make small-scale wind turbine blades in this study (ABS). We chose the material because it is inexpensive and easy to machine into the desired form. They also have outstanding chemical, stress, and creep resistance. The blade measures 332 mm in length and has a 664 mm rotor diameter. A modal study of blades is carried out, as well as a comparison with current e-glass fiber. They were able to balance the output with less vibration, according to the findings. Q blade software is used to simulate rotating output. The modal analysis testing and prototype validation of wind turbine blades were used for experimental validation.

Keywords: acrylonitrile-butadiene-styrene, e-glass fiber, modal, renewable energy, q-blade

Procedia PDF Downloads 141
1978 Fuzzy Data, Random Drift, and a Theoretical Model for the Sequential Emergence of Religious Capacity in Genus Homo

Authors: Margaret Boone Rappaport, Christopher J. Corbally

Abstract:

The ancient ape ancestral population from which living great ape and human species evolved had demographic features affecting their evolution. The population was large, had great genetic variability, and natural selection was effective at honing adaptations. The emerging populations of chimpanzees and humans were affected more by founder effects and genetic drift because they were smaller. Natural selection did not disappear, but it was not as strong. Consequences of the 'population crash' and the human effective population size are introduced briefly. The history of the ancient apes is written in the genomes of living humans and great apes. The expansion of the brain began before the human line emerged. Coalescence times for some genes are very old – up to several million years, long before Homo sapiens. The mismatch between gene trees and species trees highlights the anthropoid speciation processes, and gives the human genome history a fuzzy, probabilistic quality. However, it suggests traits that might form a foundation for capacities emerging later. A theoretical model is presented in which the genomes of early ape populations provide the substructure for the emergence of religious capacity later on the human line. The model does not search for religion, but its foundations. It suggests a course by which an evolutionary line that began with prosimians eventually produced a human species with biologically based religious capacity. The model of the sequential emergence of religious capacity relies on cognitive science, neuroscience, paleoneurology, primate field studies, cognitive archaeology, genomics, and population genetics. And, it emphasizes five trait types: (1) Documented, positive selection of sensory capabilities on the human line may have favored survival, but also eventually enriched human religious experience. (2) The bonobo model suggests a possible down-regulation of aggression and increase in tolerance while feeding, as well as paedomorphism – but, in a human species that remains cognitively sharp (unlike the bonobo). The two species emerged from the same ancient ape population, so it is logical to search for shared traits. (3) An up-regulation of emotional sensitivity and compassion seems to have occurred on the human line. This finds support in modern genetic studies. (4) The authors’ published model of morality's emergence in Homo erectus encompasses a cognitively based, decision-making capacity that was hypothetically overtaken, in part, by religious capacity. Together, they produced a strong, variable, biocultural capability to support human sociability. (5) The full flowering of human religious capacity came with the parietal expansion and smaller face (klinorhynchy) found only in Homo sapiens. Details from paleoneurology suggest the stage was set for human theologies. Larger parietal lobes allowed humans to imagine inner spaces, processes, and beings, and, with the frontal lobe, led to the first theologies composed of structured and integrated theories of the relationships between humans and the supernatural. The model leads to the evolution of a small population of African hominins that was ready to emerge with religious capacity when the species Homo sapiens evolved two hundred thousand years ago. By 50-60,000 years ago, when human ancestors left Africa, they were fully enabled.

Keywords: genetic drift, genomics, parietal expansion, religious capacity

Procedia PDF Downloads 325
1977 Resource Creation Using Natural Language Processing Techniques for Malay Translated Qur'an

Authors: Nor Diana Ahmad, Eric Atwell, Brandon Bennett

Abstract:

Text processing techniques for English have been developed for several decades. But for the Malay language, text processing methods are still far behind. Moreover, there are limited resources, tools for computational linguistic analysis available for the Malay language. Therefore, this research presents the use of natural language processing (NLP) in processing Malay translated Qur’an text. As the result, a new language resource for Malay translated Qur’an was created. This resource will help other researchers to build the necessary processing tools for the Malay language. This research also develops a simple question-answer prototype to demonstrate the use of the Malay Qur’an resource for text processing. This prototype has been developed using Python. The prototype pre-processes the Malay Qur’an and an input query using a stemming algorithm and then searches for occurrences of the query word stem. The result produced shows improved matching likelihood between user query and its answer. A POS-tagging algorithm has also been produced. The stemming and tagging algorithms can be used as tools for research related to other Malay texts and can be used to support applications such as information retrieval, question answering systems, ontology-based search and other text analysis tasks.

Keywords: language resource, Malay translated Qur'an, natural language processing (NLP), text processing

Procedia PDF Downloads 302
1976 Application of Support Vector Machines in Forecasting Non-Residential

Authors: Wiwat Kittinaraporn, Napat Harnpornchai, Sutja Boonyachut

Abstract:

This paper deals with the application of a novel neural network technique, so-called Support Vector Machine (SVM). The objective of this study is to explore the variable and parameter of forecasting factors in the construction industry to build up forecasting model for construction quantity in Thailand. The scope of the research is to study the non-residential construction quantity in Thailand. There are 44 sets of yearly data available, ranging from 1965 to 2009. The correlation between economic indicators and construction demand with the lag of one year was developed by Apichat Buakla. The selected variables are used to develop SVM models to forecast the non-residential construction quantity in Thailand. The parameters are selected by using ten-fold cross-validation method. The results are indicated in term of Mean Absolute Percentage Error (MAPE). The MAPE value for the non-residential construction quantity predicted by Epsilon-SVR in corporation with Radial Basis Function (RBF) of kernel function type is 5.90. Analysis of the experimental results show that the support vector machine modelling technique can be applied to forecast construction quantity time series which is useful for decision planning and management purpose.

Keywords: forecasting, non-residential, construction, support vector machines

Procedia PDF Downloads 420
1975 Integrating Evidence Into Health Policy: Navigating Cross-Sector and Interdisciplinary Collaboration

Authors: Tessa Heeren

Abstract:

The following proposal pertains to the complex process of successfully implementing health policies that are based on public health research. A systematic review was conducted by myself and faculty at the Cluj School of Public Health in Romania. The reviewed articles covered a wide range of topics, such as barriers and facilitators to multi-sector collaboration, differences in professional cultures, and systemic obstacles. The reviewed literature identified communication, collaboration, user-friendly dissemination, and documentation of processes in the execution of applied research as important themes for the promotion of evidence in the public health decision-making process. This proposal fits into the Academy Health National Health Policy conference because it identifies and examines differences between the worlds of research and politics. Implications and new insights for federal and/or state health policy: Recommendations made based on the findings of this research include using politically relevant levers to promote research (e.g. campaign donors, lobbies, established parties, etc.), modernizing dissemination practices, and reforms in which the involvement of external stakeholders is facilitated without relying on invitations from individual policy makers. Description of how evidence and/or data was or could be used: The reviewed articles illustrated shortcomings and areas for improvement in policy research processes and collaborative development. In general, the evidence base in the field of integrating research into policy lacks critical details of the actual process of developing evidence based policy. This shortcoming in logistical details creates a barrier for potential replication of collaborative efforts described in studies. Potential impact of the presentation for health policy: The reviewed articles focused on identifying barriers and facilitators that arise in cross sector collaboration, rather than the process and impact of integrating evidence into policy. In addition, the type of evidence used in policy was rarely specified, and widely varying interpretations of the definition of evidence complicated overall conclusions. Background: Using evidence to inform public health decision making processes has been proven effective; however, it is not clear how research is applied in practice. Aims: The objectives of the current study were to assess the extent to which evidence is used in public health decision-making process. Methods: To identify eligible studies, seven bibliographic databases, specifically, PubMed, Scopus, Cochrane Library, Science Direct, Web of Science, ClinicalKey, Health and Safety Science Abstract were screened (search dates: 1990 – September 2015); a general internet search was also conducted. Primary research and systematic reviews about the use of evidence in public health policy in Europe were included. The studies considered for inclusion were assessed by two reviewers, along with extracted data on objective, methods, population, and results. Data were synthetized as a narrative review. Results: Of 2564 articles initially identified, 2525 titles and abstracts were screened. Ultimately, 30 articles fit the research criteria by describing how or why evidence is used/not used in public health policy. The majority of included studies involved interviews and surveys (N=17). Study participants were policy makers, health care professionals, researchers, community members, service users, experts in public health.

Keywords: cross-sector, dissemination, health policy, policy implementation

Procedia PDF Downloads 209
1974 Controlling the Expense of Political Contests Using a Modified N-Players Tullock’s Model

Authors: C. Cohen, O. Levi

Abstract:

This work introduces a generalization of the classical Tullock’s model of one-stage contests under complete information with multiple unlimited numbers of contestants. In classical Tullock’s model, the contest winner is not necessarily the highest bidder. Instead, the winner is determined according to a draw in which the winning probabilities are the relative contestants’ efforts. The Tullock modeling fits well political contests, in which the winner is not necessarily the highest effort contestant. This work presents a modified model which uses a simple non-discriminating rule, namely, a parameter to influence the total costs planned for an election, for example, the contest designer can control the contestants' efforts. The winner pays a fee, and the losers are reimbursed the same amount. Our proposed model includes a mechanism that controls the efforts exerted and balances competition, creating a tighter, less predictable and more interesting contest. Additionally, the proposed model follows the fairness criterion in the sense that it does not alter the contestants' probabilities of winning compared to the classic Tullock’s model. We provide an analytic solution for the contestant's optimal effort and expected reward.

Keywords: contests, Tullock's model, political elections, control expenses

Procedia PDF Downloads 130
1973 Steepest Descent Method with New Step Sizes

Authors: Bib Paruhum Silalahi, Djihad Wungguli, Sugi Guritman

Abstract:

Steepest descent method is a simple gradient method for optimization. This method has a slow convergence in heading to the optimal solution, which occurs because of the zigzag form of the steps. Barzilai and Borwein modified this algorithm so that it performs well for problems with large dimensions. Barzilai and Borwein method results have sparked a lot of research on the method of steepest descent, including alternate minimization gradient method and Yuan method. Inspired by previous works, we modified the step size of the steepest descent method. We then compare the modification results against the Barzilai and Borwein method, alternate minimization gradient method and Yuan method for quadratic function cases in terms of the iterations number and the running time. The average results indicate that the steepest descent method with the new step sizes provide good results for small dimensions and able to compete with the results of Barzilai and Borwein method and the alternate minimization gradient method for large dimensions. The new step sizes have faster convergence compared to the other methods, especially for cases with large dimensions.

Keywords: steepest descent, line search, iteration, running time, unconstrained optimization, convergence

Procedia PDF Downloads 530
1972 Calculating the Carbon Footprint of Laser Cutting Machines from Cradle to Grave and Examination the Effect of the Use of the Machine on the Carbon Footprint

Authors: Melike Yaylacı, Tuğba Bilgin

Abstract:

Against the climate crisis, an increasing number of countries are working on green energy, carbon emission measurement, calculation and reduction. The work of industrial organizations with the highest carbon emissions on these issues is increasing. Aim of this paper is calculating carbon emissions of laser cutting machine with cradle-to-grave approach and discuss the potential affects of usage condisions, such as laser power, gas type, gas pressure, on carbon footprint. In particular, this study includes consumption of electricity used in production, laser cutting machine raw materials, and disposal of the machine. In the process of raw material supplying, machine procesing and shipping, all calculations were studied using the Tier1 approach. Laser cutting machines require a specified cutting parameter set for each different material in different thickneses, this parameters are a combination of laser power, gas type, cutting speed, gas pressure and focus point, The another purpose of this study is examine the potential affect of different cutting parameters for the same material in same thickness on carbon footprint.

Keywords: life cycle assessment, carbon emission, laser cutting machine, cutting parameters

Procedia PDF Downloads 80
1971 Application and Assessment of Artificial Neural Networks for Biodiesel Iodine Value Prediction

Authors: Raquel M. De sousa, Sofiane Labidi, Allan Kardec D. Barros, Alex O. Barradas Filho, Aldalea L. B. Marques

Abstract:

Several parameters are established in order to measure biodiesel quality. One of them is the iodine value, which is an important parameter that measures the total unsaturation within a mixture of fatty acids. Limitation of unsaturated fatty acids is necessary since warming of a higher quantity of these ones ends in either formation of deposits inside the motor or damage of lubricant. Determination of iodine value by official procedure tends to be very laborious, with high costs and toxicity of the reagents, this study uses an artificial neural network (ANN) in order to predict the iodine value property as an alternative to these problems. The methodology of development of networks used 13 esters of fatty acids in the input with convergence algorithms of backpropagation type were optimized in order to get an architecture of prediction of iodine value. This study allowed us to demonstrate the neural networks’ ability to learn the correlation between biodiesel quality properties, in this case iodine value, and the molecular structures that make it up. The model developed in the study reached a correlation coefficient (R) of 0.99 for both network validation and network simulation, with Levenberg-Maquardt algorithm.

Keywords: artificial neural networks, biodiesel, iodine value, prediction

Procedia PDF Downloads 588
1970 Use of Regression Analysis in Determining the Length of Plastic Hinge in Reinforced Concrete Columns

Authors: Mehmet Alpaslan Köroğlu, Musa Hakan Arslan, Muslu Kazım Körez

Abstract:

Basic objective of this study is to create a regression analysis method that can estimate the length of a plastic hinge which is an important design parameter, by making use of the outcomes of (lateral load-lateral displacement hysteretic curves) the experimental studies conducted for the reinforced square concrete columns. For this aim, 170 different square reinforced concrete column tests results have been collected from the existing literature. The parameters which are thought affecting the plastic hinge length such as cross-section properties, features of material used, axial loading level, confinement of the column, longitudinal reinforcement bars in the columns etc. have been obtained from these 170 different square reinforced concrete column tests. In the study, when determining the length of plastic hinge, using the experimental test results, a regression analysis have been separately tested and compared with each other. In addition, the outcome of mentioned methods on determination of plastic hinge length of the reinforced concrete columns has been compared to other methods available in the literature.

Keywords: columns, plastic hinge length, regression analysis, reinforced concrete

Procedia PDF Downloads 458
1969 Optimum Design of Alkali Activated Slag Concretes for Low Chloride Ion Permeability and Water Absorption Capacity

Authors: Müzeyyen Balçikanli, Erdoğan Özbay, Hakan Tacettin Türker, Okan Karahan, Cengiz Duran Atiş

Abstract:

In this research, effect of curing time (TC), curing temperature (CT), sodium concentration (SC) and silicate modules (SM) on the compressive strength, chloride ion permeability, and water absorption capacity of alkali activated slag (AAS) concretes were investigated. For maximization of compressive strength while for minimization of chloride ion permeability and water absorption capacity of AAS concretes, best possible combination of CT, CTime, SC and SM were determined. An experimental program was conducted by using the central composite design method. Alkali solution-slag ratio was kept constant at 0.53 in all mixture. The effects of the independent parameters were characterized and analyzed by using statistically significant quadratic regression models on the measured properties (dependent parameters). The proposed regression models are valid for AAS concretes with the SC from 0.1% to 7.5%, SM from 0.4 to 3.2, CT from 20 °C to 94 °C and TC from 1.2 hours to 25 hours. The results of test and analysis indicate that the most effective parameter for the compressive strength, chloride ion permeability and water absorption capacity is the sodium concentration.

Keywords: alkali activation, slag, rapid chloride permeability, water absorption capacity

Procedia PDF Downloads 300
1968 Valorization of Marine Seaweed Biomass: Furanic Platform Chemicals and Beyond

Authors: Sanjay Kumar, Saikat Dutta, Devendra S. Rawat, Jitendra K. Pandey, Pankaj Kumar

Abstract:

Exploding demand for various types of fuels and gradually growing impacts of atmospheric carbon dioxide have forced the researchers to search biofuels in general and algae-based biofuels in particular. However, strain identification in terms of fuel productivity and over all economics of fuel generation remains a debatable challenge. Utilization of marine biomass, especially the ones important in the Indian subcontinent, in forming furanic fuels and specialty chemicals would likely to be a better value-addition pathway. Seaweed species e.g. Ulva, Sarconema, and Gracilaria species have been found more productive than land-based biomass sources due to their higher growth rate. Additionally, non-recalcitrant nature of marine biomass unlike lignocellulosics has attracted much attention in recent years towards producing bioethanol. Here we report the production of renewable, biomass-derived platform molecules such as furfural and 5-(chloromethyl) furfural (CMF) from a seaweed species which are abundant marine biomass. These products have high potential for synthetic upgradation into various classes of value-added compounds such as fuels, fuel-additives, and monomers for polymers, solvents, agrochemicals, and pharmaceuticals.

Keywords: seaweeds, Ulva, CMF, furan

Procedia PDF Downloads 432
1967 Taguchi Method for Analyzing a Flexible Integrated Logistics Network

Authors: E. Behmanesh, J. Pannek

Abstract:

Logistics network design is known as one of the strategic decision problems. As these kinds of problems belong to the category of NP-hard problems, traditional ways are failed to find an optimal solution in short time. In this study, we attempt to involve reverse flow through an integrated design of forward/reverse supply chain network that formulated into a mixed integer linear programming. This Integrated, multi-stages model is enriched by three different delivery path which makes the problem more complex. To tackle with such an NP-hard problem a revised random path direct encoding method based memetic algorithm is considered as the solution methodology. Each algorithm has some parameters that need to be investigate to reveal the best performance. In this regard, Taguchi method is adapted to identify the optimum operating condition of the proposed memetic algorithm to improve the results. In this study, four factors namely, population size, crossover rate, local search iteration and a number of iteration are considered. Analyzing the parameters and improvement in results are the outlook of this research.

Keywords: integrated logistics network, flexible path, memetic algorithm, Taguchi method

Procedia PDF Downloads 175
1966 Minimization of Seepage in Sandy Soil Using Different Grouting Types

Authors: Eng. M. Ahmed, A. Ibrahim, M. Ashour

Abstract:

One of the major concerns facing dam is the repair of their structures to prevent the seepage under them. In previous years, many existing dams have been treated by grouting, but with varying degrees of success. One of the major reasons for this erratic performance is the unsuitable selection of the grouting materials to reduce the seepage. Grouting is an effective way to improve the engineering properties of the soil and strengthen of the permeability of the soil to reduce the seepage. The purpose of this paper is to focus on the efficiency of current available grouting materials and techniques from construction, environmental and economical point of view. The seepage reduction usually accomplished by either chemical grouting or cementious grouting using ultrafine cement. In addition, the study shows a comparison between grouting materials according to their degree of permeability reduction and cost. The application of seepage reduction is based on the permeation grouting using grout curtain installation. The computer program (SEEP/W) is employed to model a dam rested on sandy soil, using grout curtain to reduce seepage quantity and hydraulic gradient by different grouting materials. This study presents a relationship that takes into account the permeability of the soil, grout curtain spacing and a new performance parameter that can be used to predict the best selection of grouting materials for seepage reduction.

Keywords: seepage, sandy soil, grouting, permeability

Procedia PDF Downloads 349
1965 The Effect of Soil Binder and Gypsum to the Changes of the Expansive Soil Shear Strength Parameters

Authors: Yulia Hastuti, Ratna Dewi, Muhammad Sandi

Abstract:

Many methods of soil stabilization that can be done such as by mixing chemicals. In this research, stabilization by mixing the soil using two types of chemical admixture, those are gypsum with a variation of 5%, 10%, and 15% and Soil binder with a concentration of 20 gr / lot of water, 25 gr / lot of water, and 30 gr / lot of water aimed to determine the effect on the soil plasticity index values and comparing the value of shear strength parameters of the mixture with the original soil conditions using a Triaxial UU test. Based on research done shows that with increasing variations in the mix, then the value of plasticity index decreased, which was originally 42% (very high degree of swelling) becomes worth 11.24% (lower Swelling degree) when a mixture of gypsum 15% and 30 gr / Lt water soil binder. As for the value shear, strength parameters increased in all variations of mixture. Admixture with the highest shear strength parameter's value is at 15% the mixture of gypsum and 20 gr / litre of water of soil binder with the 14 day treatment period, which has enhanced the cohesion value of 559.01%, the friction angle by 1157.14%. And a shear strength value of 568.49%. It can be concluded that the admixture of gypsum and soil binder correctly, can increase the value of shear strength parameters significantly and decrease the value of plasticity index of the soil.

Keywords: expansive soil, gypsum, soil binder, shear strength

Procedia PDF Downloads 457
1964 Multi-Objective Four-Dimensional Traveling Salesman Problem in an IoT-Based Transport System

Authors: Arindam Roy, Madhushree Das, Apurba Manna, Samir Maity

Abstract:

In this research paper, an algorithmic approach is developed to solve a novel multi-objective four-dimensional traveling salesman problem (MO4DTSP) where different paths with various numbers of conveyances are available to travel between two cities. NSGA-II and Decomposition algorithms are modified to solve MO4DTSP in an IoT-based transport system. This IoT-based transport system can be widely observed, analyzed, and controlled by an extensive distribution of traffic networks consisting of various types of sensors and actuators. Due to urbanization, most of the cities are connected using an intelligent traffic management system. Practically, for a traveler, multiple routes and vehicles are available to travel between any two cities. Thus, the classical TSP is reformulated as multi-route and multi-vehicle i.e., 4DTSP. The proposed MO4DTSP is designed with traveling cost, time, and customer satisfaction as objectives. In reality, customer satisfaction is an important parameter that depends on travel costs and time reflects in the present model.

Keywords: multi-objective four-dimensional traveling salesman problem (MO4DTSP), decomposition, NSGA-II, IoT-based transport system, customer satisfaction

Procedia PDF Downloads 97
1963 Comparative Analysis of Dissimilarity Detection between Binary Images Based on Equivalency and Non-Equivalency of Image Inversion

Authors: Adnan A. Y. Mustafa

Abstract:

Image matching is a fundamental problem that arises frequently in many aspects of robot and computer vision. It can become a time-consuming process when matching images to a database consisting of hundreds of images, especially if the images are big. One approach to reducing the time complexity of the matching process is to reduce the search space in a pre-matching stage, by simply removing dissimilar images quickly. The Probabilistic Matching Model for Binary Images (PMMBI) showed that dissimilarity detection between binary images can be accomplished quickly by random pixel mapping and is size invariant. The model is based on the gamma binary similarity distance that recognizes an image and its inverse as containing the same scene and hence considers them to be the same image. However, in many applications, an image and its inverse are not treated as being the same but rather dissimilar. In this paper, we present a comparative analysis of dissimilarity detection between PMMBI based on the gamma binary similarity distance and a modified PMMBI model based on a similarity distance that does distinguish between an image and its inverse as being dissimilar.

Keywords: binary image, dissimilarity detection, probabilistic matching model for binary images, image mapping

Procedia PDF Downloads 138
1962 Discovering Traditional Plants Used by Indigenous People in the Tropical Rainforest of Malaysia for the Treatment of Malaria

Authors: Izdihar Ismail, Alona C. Linatoc, Maryati Mohamed

Abstract:

The tropical rainforest of Malaysia is known for its rich biological diversity and high endemicity. The potential for these forests to hold the cure for many diseases and illnesses is high and much is yet to be discovered. This study explores the richness of the tropical rainforest of Endau-Rompin National Park in Johor, Malaysia in search of plants traditionally used by the indigenous people in the treatment of malaria and malaria-like symptoms. Seven species of plants were evaluated and tested for antiplasmodial activities. Different plant parts were subjected to methanolic and aqueous extractions. A total of 24 extracts were evaluated by histidine-rich protein II (HRP2) assay against K1 strain of Plasmodium falciparum chloroquine-resistant. Ten extracts showed significant inhibition of the growth of P. falciparum. Phytochemical screening of the same extracts revealed the presence of alkaloids, flavonoids, terpenoids and anthraquinones. This study affirms that tropical rainforests may still hold undiscovered cures for many diseases and illnesses that have inflicted millions of people worldwide. The species studied herein have not known to have been studied elsewhere before.

Keywords: Endau-Rompin, malaria, Malaysia, tropical rainforest, traditional knowledge

Procedia PDF Downloads 255
1961 Study on the Transition to Pacemaker of Two Coupled Neurons

Authors: Sun Zhe, Ruggero Micheletto

Abstract:

The research of neural network is very important for the development of advanced next generation intelligent devices and the medical treatment. The most important part of the neural network research is the learning. The process of learning in our brain is essentially several adjustment processes of connection strength between neurons. It is very difficult to figure out how this mechanism works in the complex network and how the connection strength influences brain functions. For this reason, we made a model with only two coupled neurons and studied the influence of connection strength between them. To emulate the neuronal activity of realistic neurons, we prefer to use the Izhikevich neuron model. This model can simulate the neuron variables accurately and it’s simplicity is very suitable to implement on computers. In this research, the parameter ρ is used to estimate the correlation coefficient between spike train of two coupling neurons.We think the results is very important for figuring out the mechanism between synchronization of coupling neurons and synaptic plasticity. The result also presented the importance of the spike frequency adaptation in complex systems.

Keywords: neural networks, noise, stochastic processes, coupled neurons, correlation coefficient, synchronization, pacemaker, synaptic plasticity

Procedia PDF Downloads 266
1960 Optimization of Geometric Parameters of Microfluidic Channels for Flow-Based Studies

Authors: Parth Gupta, Ujjawal Singh, Shashank Kumar, Mansi Chandra, Arnab Sarkar

Abstract:

Microfluidic devices have emerged as indispensable tools across various scientific disciplines, offering precise control and manipulation of fluids at the microscale. Their efficacy in flow-based research, spanning engineering, chemistry, and biology, relies heavily on the geometric design of microfluidic channels. This work introduces a novel approach to optimise these channels through Response Surface Methodology (RSM), departing from the conventional practice of addressing one parameter at a time. Traditionally, optimising microfluidic channels involved isolated adjustments to individual parameters, limiting the comprehensive understanding of their combined effects. In contrast, our approach considers the simultaneous impact of multiple parameters, employing RSM to efficiently explore the complex design space. The outcome is an innovative microfluidic channel that consumes an optimal sample volume and minimises flow time, enhancing overall efficiency. The relevance of geometric parameter optimization in microfluidic channels extends significantly in biomedical engineering. The flow characteristics of porous materials within these channels depend on many factors, including fluid viscosity, environmental conditions (such as temperature and humidity), and specific design parameters like sample volume, channel width, channel length, and substrate porosity. This intricate interplay directly influences the performance and efficacy of microfluidic devices, which, if not optimized, can lead to increased costs and errors in disease testing and analysis. In the context of biomedical applications, the proposed approach addresses the critical need for precision in fluid flow. it mitigate manufacturing costs associated with trial-and-error methodologies by optimising multiple geometric parameters concurrently. The resulting microfluidic channels offer enhanced performance and contribute to a streamlined, cost-effective process for testing and analyzing diseases. A key highlight of our methodology is its consideration of the interconnected nature of geometric parameters. For instance, the volume of the sample, when optimized alongside channel width, length, and substrate porosity, creates a synergistic effect that minimizes errors and maximizes efficiency. This holistic optimization approach ensures that microfluidic devices operate at their peak performance, delivering reliable results in disease testing. A key highlight of our methodology is its consideration of the interconnected nature of geometric parameters. For instance, the volume of the sample, when optimized alongside channel width, length, and substrate porosity, creates a synergistic effect that minimizes errors and maximizes efficiency. This holistic optimization approach ensures that microfluidic devices operate at their peak performance, delivering reliable results in disease testing. A key highlight of our methodology is its consideration of the interconnected nature of geometric parameters. For instance, the volume of the sample, when optimized alongside channel width, length, and substrate porosity, creates a synergistic effect that minimizes errors and maximizes efficiency. This holistic optimization approach ensures that microfluidic devices operate at their peak performance, delivering reliable results in disease testing.

Keywords: microfluidic device, minitab, statistical optimization, response surface methodology

Procedia PDF Downloads 39
1959 Thermal Resistance of Special Garments Exposed to a Radiant Heat

Authors: Jana Pichova, Lubos Hes, Vladimir Bajzik

Abstract:

Protective clothing is designed to keep a wearer save in hazardous conditions or enable perform short time working operation without being injured or feeling discomfort. Firefighters or other related workers are exposed to abnormal heat which can be conductive, convective or radiant type. Their garment is proposed to resist this conditions and prevent burn injuries or dead of human. However thermal comfort of firefighter exposed to high heat source have not been studied yet. Thermal resistance is the best representative parameter of thermal comfort. In this study a new method of testing of thermal resistance of special clothing exposed to high radiation heat source was designed. This method simulates human body wearing single or multi-layered garment which is exposed to radiative heat. Setup of this method enables measuring of radiative heat flow in time without effect of convection. The new testing method is verified on chosen group of textiles for firefighters.

Keywords: protective clothing, radiative heat, thermal comfort of firefighters, thermal resistance of special garments

Procedia PDF Downloads 359
1958 Plastic Strain Accumulation Due to Asymmetric Cyclic Loading of Zircaloy-2 at 400°C

Authors: R. S. Rajpurohit, N. C. Santhi Srinivas, Vakil Singh

Abstract:

Asymmetric stress cycling leads to accumulation of plastic strain which is called as ratcheting strain. The problem is generally associated with nuclear fuel cladding materials used in nuclear power plants and pressurized pipelines. In the present investigation, asymmetric stress controlled fatigue tests were conducted with three different parameters namely, mean stress, stress amplitude and stress rate (keeping two parameters constant and varying third parameter) to see the plastic strain accumulation and its effect on fatigue life and deformation behavior of Zircaloy-2 at 400°C. The tests were conducted with variable mean stress (45-70 MPa), stress amplitude (95-120 MPa) and stress rate (30-750 MPa/s) and tested specimens were characterized using transmission and scanning electron microscopy. The experimental results show that with the increase in mean stress and stress amplitude, the ratcheting strain accumulation increases with reduction in fatigue life. However, increase in stress rate leads to improvement in fatigue life of the material due to small ratcheting strain accumulation. Fractographs showed a decrease in area fraction of fatigue failed region.

Keywords: asymmetric cyclic loading, ratcheting fatigue, mean stress, stress amplitude, stress rate, plastic strain

Procedia PDF Downloads 248
1957 Comparison of Different k-NN Models for Speed Prediction in an Urban Traffic Network

Authors: Seyoung Kim, Jeongmin Kim, Kwang Ryel Ryu

Abstract:

A database that records average traffic speeds measured at five-minute intervals for all the links in the traffic network of a metropolitan city. While learning from this data the models that can predict future traffic speed would be beneficial for the applications such as the car navigation system, building predictive models for every link becomes a nontrivial job if the number of links in a given network is huge. An advantage of adopting k-nearest neighbor (k-NN) as predictive models is that it does not require any explicit model building. Instead, k-NN takes a long time to make a prediction because it needs to search for the k-nearest neighbors in the database at prediction time. In this paper, we investigate how much we can speed up k-NN in making traffic speed predictions by reducing the amount of data to be searched for without a significant sacrifice of prediction accuracy. The rationale behind this is that we had a better look at only the recent data because the traffic patterns not only repeat daily or weekly but also change over time. In our experiments, we build several different k-NN models employing different sets of features which are the current and past traffic speeds of the target link and the neighbor links in its up/down-stream. The performances of these models are compared by measuring the average prediction accuracy and the average time taken to make a prediction using various amounts of data.

Keywords: big data, k-NN, machine learning, traffic speed prediction

Procedia PDF Downloads 343