Search results for: H∞ methods.
2865 Evaluation of Urban Development Proposals An ANP Approach
Authors: T. Gómez-Navarro, M. García-Melón, D. Díaz-Martín, S. Acuna-Dutra,
Abstract:
In this paper a new approach to prioritize urban planning projects in an efficient and reliable way is presented. It is based on environmental pressure indices and multicriteria decision methods. The paper introduces a rigorous method with acceptable complexity of rank ordering urban development proposals according to their environmental pressure. The technique combines the use of Environmental Pressure Indicators, the aggregation of indicators in an Environmental Pressure Index by means of the Analytic Network Process method and interpreting the information obtained from the experts during the decision-making process. The ANP method allows the aggregation of the experts- judgments on each of the indicators into one Environmental Pressure Index. In addition, ANP is based on utility ratio functions which are the most appropriate for the analysis of uncertain data, like experts- estimations. Finally, unlike the other multicriteria techniques, ANP allows the decision problem to be modelled using the relationships among dependent criteria. The method has been applied to the proposal for urban development of La Carlota airport in Caracas (Venezuela). The Venezuelan Government would like to see a recreational project develop on the abandoned area and mean a significant improvement for the capital. There are currently three options on their table which are currently under evaluation. They include a Health Club, a Residential area and a Theme Park. The participating experts coincided in the appreciation that the method proposed in this paper is useful and an improvement from traditional techniques such as environmental impact studies, lifecycle analysis, etc. They find the results obtained coherent, the process seems sufficiently rigorous and precise, and the use of resources is significantly less than in other methods.
Keywords: Environmental pressure indicators, multicriteria decision analysis, analytic network process.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 18032864 Effects of Centrifugation, Encapsulation Method and Different Coating Materials on the Total Antioxidant Activity of the Microcapsules of Powdered Cherry Laurels
Authors: B. Cilek Tatar, G. Sumnu, M. Oztop, E. Ayaz
Abstract:
Encapsulation protects sensitive food ingredients against heat, oxygen, moisture and pH until they are released to the system. It can mask the unwanted taste of nutrients that are added to the foods for fortification purposes. Cherry laurels (Prunus laurocerasus) contain phenolic compounds which decrease the proneness to several chronic diseases such as types of cancer and cardiovascular diseases. The objective of this research was to study the effects of centrifugation, different coating materials and homogenization methods on microencapsulation of powders obtained from cherry laurels. In this study, maltodextrin and mixture of maltodextrin:whey protein with a ratio of 1:3 (w/w) were chosen as coating materials. Total solid content of coating materials was kept constant as 10% (w/w). Capsules were obtained from powders of freeze-dried cherry laurels through encapsulation process by silent crusher homogenizer or microfluidization. Freeze-dried cherry laurels were core materials and core to coating ratio was chosen as 1:10 by weight. To homogenize the mixture, high speed homogenizer was used at 4000 rpm for 5 min. Then, silent crusher or microfluidizer was used to complete encapsulation process. The mixtures were treated either by silent crusher for 1 min at 75000 rpm or microfluidizer at 50 MPa for 3 passes. Freeze drying for 48 hours was applied to emulsions to obtain capsules in powder form. After these steps, dry capsules were grounded manually into a fine powder. The microcapsules were analyzed for total antioxidant activity with DPPH (1,1-diphenyl-2-picrylhydrazyl) radical scavenging method. Prior to high speed homogenization, the samples were centrifuged (4000 rpm, 1 min). Centrifugation was found to have positive effect on total antioxidant activity of capsules. Microcapsules treated by microfluidizer were found to have higher total antioxidant activities than those treated by silent crusher. It was found that increasing whey protein concentration in coating material (using maltodextrin:whey protein 1:3 mixture) had positive effect on total antioxidant activity for both silent crusher and microfluidization methods. Therefore, capsules prepared by microfluidization of centrifuged mixtures can be selected as the best conditions for encapsulation of cherry laurel powder by considering their total antioxidant activity. In this study, it was shown that capsules prepared by these methods can be recommended to be incorporated into foods in order to enhance their functionality by increasing antioxidant activity.
Keywords: Antioxidant activity, cherry laurel, microencapsulation, microfluidization.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 13092863 A Quantitative Study on the Effects of School Development on Character Development
Authors: Merve Gücen
Abstract:
One of the aims of education is to educate individuals who have embraced universal moral principles and transform universal moral principles into moral values. Character education aims to educate behaviors of individuals in their mental activities to transform moral principles into moral values in their lives. As the result of this education, individuals are expected to develop positive character traits and become morally indifferent individuals. What are the characteristics of the factors that influence character education at this stage? How should character education help individuals develop positive character traits? Which methods are more effective? These questions come to mind when studying character education. Our research was developed within the framework of these questions. The aim of our study is to provide the most effective use of the education factor that affects character. In this context, we tried to explain character definition, character development, character education and the factors affecting character education using qualitative research methods. At this stage, character education programs applied in various countries were examined and a character education program consisting of Islamic values was prepared and implemented in an International Imam Hatip High School in Istanbul. Our application was carried out with the collaboration of school and families. Various seminars were organized in the school and participation of families was ensured. In the last phase of our study, we worked with the students and their families on the effectiveness of the events held during the program. In this study, it was found that activities such as storytelling and theater in character education programs were effective in recognizing wrong behaviors in individuals. It was determined that our program had a positive effect on the quality of education. It was seen that applications of this educational program affected the behavior of the employees in the educational institution.
Keywords: Character development, values education, family activities, education program.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 10422862 Citizen Participation in Informal Settlements; Potentials & Obstacles - The Case of Iran, Shiraz, Saadi Community
Authors: Hamid Mohammadi Makerani
Abstract:
In recent years, “Bottom-up Planning Approach" has been widely accepted and expanded from planning theorists. Citizen participation becomes more important in decision-making in informal settlements. Many of previous projects and strategies due to ignorance of citizen participation, have been failed facing with informal settlements and in some cases lead physical expansion of these neighbourhoods. According to recent experiences, the new participatory approach was in somehow successful. This paper focuses on local experiences in Iran. A considerable amount of people live in informal settlements in Iran. With the previous methods, the government could not solve the problems of these settlements. It is time to examine new methods such as empowerment of the local citizens and involve them to solve the current physical, social, and economic problems. The paper aims to address the previous and new strategies facing with informal settlements, the conditions under which citizens could be involved in planning process, limits and potentials of this process, the main actors and issues and finally motivations that are able to promote citizen participation. Documentary studies, observation, interview and questionnaire have been used to achieve the above mentioned objectives. Nearly 80 percent of responder in Saadi Community are ready to participate in regularising their neighbourhoods, if pre-conditions of citizen involvement are being provided. These pre-conditions include kind of problem and its severity, the importance of issue, existence of a short-term solution, etc. Moreover, confirmation of dweller-s ownership can promote the citizen engagement in participatory projects.
Keywords: Bottom-up Planning, Citizen Participation, Informal Settlements, Local Resources.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 25012861 QCM-D Study of E-casein Adsorption on Bimodal PEG Brushes
Authors: N. Ngadi, J. Abrahamson, C. Fee, K. Morison
Abstract:
Adsorption of proteins onto a solid surface is believed to be the initial and controlling step in biofouling. A better knowledge of the fouling process can be obtained by controlling the formation of the first protein layer at a solid surface. A number of methods have been investigated to inhibit adsorption of proteins. In this study, the adsorption kinetics of
Keywords: E-casein, QCM-D, stainless steel, bimodal brush, PEG
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 13792860 Novel Adaptive Channel Equalization Algorithms by Statistical Sampling
Authors: János Levendovszky, András Oláh
Abstract:
In this paper, novel statistical sampling based equalization techniques and CNN based detection are proposed to increase the spectral efficiency of multiuser communication systems over fading channels. Multiuser communication combined with selective fading can result in interferences which severely deteriorate the quality of service in wireless data transmission (e.g. CDMA in mobile communication). The paper introduces new equalization methods to combat interferences by minimizing the Bit Error Rate (BER) as a function of the equalizer coefficients. This provides higher performance than the traditional Minimum Mean Square Error equalization. Since the calculation of BER as a function of the equalizer coefficients is of exponential complexity, statistical sampling methods are proposed to approximate the gradient which yields fast equalization and superior performance to the traditional algorithms. Efficient estimation of the gradient is achieved by using stratified sampling and the Li-Silvester bounds. A simple mechanism is derived to identify the dominant samples in real-time, for the sake of efficient estimation. The equalizer weights are adapted recursively by minimizing the estimated BER. The near-optimal performance of the new algorithms is also demonstrated by extensive simulations. The paper has also developed a (Cellular Neural Network) CNN based approach to detection. In this case fast quadratic optimization has been carried out by t, whereas the task of equalizer is to ensure the required template structure (sparseness) for the CNN. The performance of the method has also been analyzed by simulations.
Keywords: Cellular Neural Network, channel equalization, communication over fading channels, multiuser communication, spectral efficiency, statistical sampling.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 15202859 Computational Methods in Official Statistics with an Example on Calculating and Predicting Diabetes Mellitus [DM] Prevalence in Different Age Groups within Australia in Future Years, in Light of the Aging Population
Authors: D. Hilton
Abstract:
An analysis of the Australian Diabetes Screening Study estimated undiagnosed diabetes mellitus [DM] prevalence in a high risk general practice based cohort. DM prevalence varied from 9.4% to 18.1% depending upon the diagnostic criteria utilised with age being a highly significant risk factor. Utilising the gold standard oral glucose tolerance test, the prevalence of DM was 22-23% in those aged >= 70 years and <15% in those aged 40-59 years. Opportunistic screening in Australian general practice potentially can identify many persons with undiagnosed type 2 DM. An Australian Bureau of Statistics document published three years ago, reported the highest rate of DM in men aged 65-74 years [19%] whereas the rate for women was highest in those over 75 years [13%]. If you consider that the Australian Bureau of Statistics report in 2007 found that 13% of the population was over 65 years of age and that this will increase to 23-25% by 2056 with a further projected increase to 25-28% by 2101, obviously this information has to be factored into the equation when age related diabetes prevalence predictions are calculated. This 10-15% proportional increase of elderly persons within the population demographics has dramatic implications for the estimated number of elderly persons with DM in these age groupings. Computational methodology showing the age related demographic changes reported in these official statistical documents will be done showing estimates for 2056 and 2101 for different age groups. This has relevance for future diabetes prevalence rates and shows that along with many countries worldwide Australia is facing an increasing pandemic. In contrast Japan is expected to have a decrease in the next twenty years in the number of persons with diabetes.
Keywords: Epidemiological methods, aging, prevalence.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 19542858 Numerical Treatment of Matrix Differential Models Using Matrix Splines
Authors: Kholod M. Abualnaja
Abstract:
This paper consider the solution of the matrix differential models using quadratic, cubic, quartic, and quintic splines. Also using the Taylor’s and Picard’s matrix methods, one illustrative example is included.
Keywords: Matrix Splines, Cubic Splines, Quartic Splines.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 17032857 Quantification of Soft Tissue Artefacts Using Motion Capture Data and Ultrasound Depth Measurements
Authors: Azadeh Rouhandeh, Chris Joslin, Zhen Qu, Yuu Ono
Abstract:
The centre of rotation of the hip joint is needed for an accurate simulation of the joint performance in many applications such as pre-operative planning simulation, human gait analysis, and hip joint disorders. In human movement analysis, the hip joint center can be estimated using a functional method based on the relative motion of the femur to pelvis measured using reflective markers attached to the skin surface. The principal source of errors in estimation of hip joint centre location using functional methods is soft tissue artefacts due to the relative motion between the markers and bone. One of the main objectives in human movement analysis is the assessment of soft tissue artefact as the accuracy of functional methods depends upon it. Various studies have described the movement of soft tissue artefact invasively, such as intra-cortical pins, external fixators, percutaneous skeletal trackers, and Roentgen photogrammetry. The goal of this study is to present a non-invasive method to assess the displacements of the markers relative to the underlying bone using optical motion capture data and tissue thickness from ultrasound measurements during flexion, extension, and abduction (all with knee extended) of the hip joint. Results show that the artefact skin marker displacements are non-linear and larger in areas closer to the hip joint. Also marker displacements are dependent on the movement type and relatively larger in abduction movement. The quantification of soft tissue artefacts can be used as a basis for a correction procedure for hip joint kinematics.
Keywords: Hip joint centre, motion capture, soft tissue artefact, ultrasound depth measurement.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 28612856 The Effect of a Three-Month Training Program on the Back Kyphosis of Former Male Addicts
Authors: M. J. Pourvaghar, Sh. Khoshemehry
Abstract:
Adopting inappropriate body posture during addiction can cause muscular and skeletal deformities. This study is aimed at investigating the effects of a program of the selected corrective exercises on the kyphosis of addicted male patients. Materials and methods: This was a quasi-experimental study. This study has been carried out using the semi-experimental method. The subjects of the present study included 104 addicted men between 25 to 45 years of age. In 2014, these men were referred to one of the NA (Narcotic Anonymous) centres in Kashan in 2015. A total of 24 people suffering from drug withdrawal, who had abnormal kyphosis, were purposefully selected as a sample. The sample was randomly divided into two groups, experimental and control; each group consisted of 12 people. The experimental group participated in a training program for 12 weeks consisting of three 60 minute sessions per week. That includes strengthening, stretching and PNF exercises (deep stretching of the muscle). The control group did no exercise or corrective activity. The Kolmogorov-Smirnov test was used to assess normal distribution of data; and a paired t-test and covariance analysis test were used to assess the effectiveness of the exercises, with a significance level of P≤0.05 by using SPSS18. The results showed that three months of the selected corrective exercises had a significant effect (P≤ 0.005) on the correction of the kyphosis of the addicted male patients after three months of rehabilitation (drug withdrawal) in the experimental group, while there was no significant difference recorded in the control group (P≥0.05). The results show that exercise and corrective activities can be used as non-invasive and non-pharmacological methods to rehabilitate kyphosis abnormalities after drug withdrawal and treatment for addiction.
Keywords: Kyphosis, corrective exercises, addict, drug withdrawal.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 17902855 A Survey on Principal Aspects of Secure Image Transmission
Authors: Ali Soleymani, Zulkarnain Md Ali, Md Jan Nordin
Abstract:
This paper is a review on the aspects and approaches of design an image cryptosystem. First a general introduction given for cryptography and images encryption and followed by different techniques in image encryption and related works for each technique surveyed. Finally, general security analysis methods for encrypted images are mentioned.
Keywords: Image, cryptography, encryption, security, analysis.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 23852854 Evaluation of Underground Water Flow into Tabriz Metro Tunnel First Line by Hydro-Mechanical Coupling Analysis
Authors: L. Nikakhtar, S. Zare
Abstract:
One of the main practical difficulties attended with tunnel construction is related to underground water. Uncontrolled water behavior may cause extra loads on the lining, mechanical instability, and unfavorable environmental problems. Estimating underground water inflow rate to the tunnels is a complex skill. The common calculation methods are: empirical methods, analytical solutions, numerical solutions based on the equivalent continuous porous media. In this research the rate of underground water inflow to the Tabriz metro first line tunnel has been investigated by numerical finite difference method using FLAC2D software. Comparing results of Heuer analytical method and numerical simulation showed good agreement with each other. Fully coupled and one-way coupled hydro mechanical states as well as water-free conditions in the soil around the tunnel are used in numerical models and these models have been applied to evaluate the loading value on the tunnel support system. Results showed that the fully coupled hydro mechanical analysis estimated more axial forces, moments and shear forces in linings, so this type of analysis is more conservative and reliable method for design of tunnel lining system. As sensitivity analysis, inflow water rates into the tunnel were evaluated in different soil permeability, underground water levels and depths of the tunnel. Result demonstrated that water level in constant depth of the tunnel is more sensitive factor for water inflow rate to the tunnel in comparison of other parameters investigated in the sensitivity analysis.
Keywords: Coupled hydro mechanical analysis, FLAC2D, Tabriz Metro, inflow rate.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 10482853 Using Game Engines in Lightning Shielding: The Application of the Rolling Spheres Method on Virtual As-Built Power Substations
Authors: Yuri A. Gruber, Matheus Rosendo, Ulisses G. A. Casemiro, Klaus de Geus, Rafael T. Bee
Abstract:
Lightning strikes can cause severe negative impacts to the electrical sector causing direct damage to equipment as well as shutdowns, especially when occurring in power substations. In order to mitigate this problem, a meticulous planning of the power substation protection system is of vital importance. A critical part of this is the distribution of shielding wires through the substation, which creates a 3D imaginary protection mesh similar to a circus tarpaulin. Equipment enclosed in the volume defined by that 3D mesh is considered protected against lightning strikes. The use of traditional methods of longitudinal cutting analysis based on 2D CAD tools makes the process laborious and the results obtained may not guarantee satisfactory protection of electrical equipment. This work describes the application of a Game Engine to the problem of lightning protection of power substations providing the visualization of the 3D protection mesh, the amount of protected components and the highlight of equipment which remain unprotected. In addition, aspects regarding the implementation and the advantages of approaching the problem using Unreal® Engine 4 are described. In order to validate results, a comparison with traditional 2D methods is applied to the same case study to which the proposed technique has been applied. Finally, a comparative study involving different levels of protection using the technique developed in this work is presented, showing that modern game engines can be a powerful accessory for simulations in several areas of engineering.
Keywords: Game engine, rolling spheres method, substation protection, UE4, Unreal® Engine 4.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 12412852 Development of an Ensemble Classification Model Based on Hybrid Filter-Wrapper Feature Selection for Email Phishing Detection
Authors: R. B. Ibrahim, M. S. Argungu, I. M. Mungadi
Abstract:
It is obvious in this present time, internet has become an indispensable part of human life since its inception. The Internet has provided diverse opportunities to make life so easy for human beings, through the adoption of various channels. Among these channels are email, internet banking, video conferencing, and the like. Email is one of the easiest means of communication hugely accepted among individuals and organizations globally. But over decades the security integrity of this platform has been challenged with malicious activities like Phishing. Email phishing is designed by phishers to fool the recipient into handing over sensitive personal information such as passwords, credit card numbers, account credentials, social security numbers, etc. This activity has caused a lot of financial damage to email users globally which has resulted in bankruptcy, sudden death of victims, and other health-related sicknesses. Although many methods have been proposed to detect email phishing, in this research, the results of multiple machine-learning methods for predicting email phishing have been compared with the use of filter-wrapper feature selection. It is worth noting that all three models performed substantially but one outperformed the other. The dataset used for these models is obtained from Kaggle online data repository, while three classifiers: decision tree, Naïve Bayes, and Logistic regression are ensemble (Bagging) respectively. Results from the study show that the Decision Tree (CART) bagging ensemble recorded the highest accuracy of 98.13% using PEF (Phishing Essential Features). This result further demonstrates the dependability of the proposed model.
Keywords: Ensemble, hybrid, filter-wrapper, phishing.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1792851 Development of a Tunisian Measurement Scale for Patient Satisfaction: Study case in Tunisian Private Clinics
Authors: M. Daoud-Marrakchi, S. Fendri-Elouze, Ch. Ill, B. Bejar-Ghadhab
Abstract:
The aim of this research is to propose a Measurement Scale for Patient Satisfaction (MSPS) in the context of Tunisian private clinics. This scale is developed using value management methods and is validated by statistic tools with SPSS.Keywords: Functional analysis, Patient satisfaction, Questionnaire, Reliability, Validity.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 19582850 Reducing the Imbalance Penalty through Artificial Intelligence Methods Geothermal Production Forecasting: A Case Study for Turkey
Abstract:
In addition to being rich in renewable energy resources, Turkey is one of the countries that promise potential in geothermal energy production with its high installed power, cheapness, and sustainability. Increasing imbalance penalties become an economic burden for organizations, since the geothermal generation plants cannot maintain the balance of supply and demand due to the inadequacy of the production forecasts given in the day-ahead market. A better production forecast reduces the imbalance penalties of market participants and provides a better imbalance in the day ahead market. In this study, using machine learning, deep learning and time series methods, the total generation of the power plants belonging to Zorlu Doğal Electricity Generation, which has a high installed capacity in terms of geothermal, was predicted for the first one-week and first two-weeks of March, then the imbalance penalties were calculated with these estimates and compared with the real values. These modeling operations were carried out on two datasets, the basic dataset and the dataset created by extracting new features from this dataset with the feature engineering method. According to the results, Support Vector Regression from traditional machine learning models outperformed other models and exhibited the best performance. In addition, the estimation results in the feature engineering dataset showed lower error rates than the basic dataset. It has been concluded that the estimated imbalance penalty calculated for the selected organization is lower than the actual imbalance penalty, optimum and profitable accounts.
Keywords: Machine learning, deep learning, time series models, feature engineering, geothermal energy production forecasting.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2042849 Importance of Risk Assessment in Managers´ Decision-Making Process
Authors: Mária Hudáková, Vladimír Míka, Katarína Hollá
Abstract:
Making decisions is the core of management and a result of conscious activities which is under way in a particular environment and concrete conditions. The managers decide about the goals, procedures and about the methods how to respond to the changes and to the problems which developed. Their decisions affect the effectiveness, quality, economy and the overall successfulness in every organisation. In spite of this fact, they do not pay sufficient attention to the individual steps of the decision-making process. They emphasise more how to cope with the individual methods and techniques of making decisions and forget about the way how to cope with analysing the problem or assessing the individual solution variants. In many cases, the underestimating of the analytical phase can lead to an incorrect assessment of the problem and this can then negatively influence its further solution. Based on our analysis of the theoretical solutions by individual authors who are dealing with this area and the realised research in Slovakia and also abroad we can recognise an insufficient interest of the managers to assess the risks in the decision-making process. The goal of this paper is to assess the risks in the managers´ decision-making process relating to the conditions of the environment, to the subject’s activity (the manager’s personality), to the insufficient assessment of individual variants for solving the problems but also to situations when the arisen problem is not solved. The benefit of this paper is the effort to increase the need of the managers to deal with the risks during the decision-making process. It is important for every manager to assess the risks in his/her decision-making process and to make efforts to take such decisions which reflect the basic conditions, states and development of the environment in the best way and especially for the managers´ decisions to contribute to achieving the determined goals of the organisation as effectively as possible.
Keywords: Risk, decision-making, manager, process, analysis, source of risk.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 17992848 Development of a Catchment Water Quality Model for Continuous Simulations of Pollutants Build-up and Wash-off
Authors: Iqbal Hossain, Dr. Monzur Imteaz, Dr. Shirley Gato-Trinidad, Prof. Abdallah Shanableh
Abstract:
Estimation of runoff water quality parameters is required to determine appropriate water quality management options. Various models are used to estimate runoff water quality parameters. However, most models provide event-based estimates of water quality parameters for specific sites. The work presented in this paper describes the development of a model that continuously simulates the accumulation and wash-off of water quality pollutants in a catchment. The model allows estimation of pollutants build-up during dry periods and pollutants wash-off during storm events. The model was developed by integrating two individual models; rainfall-runoff model, and catchment water quality model. The rainfall-runoff model is based on the time-area runoff estimation method. The model allows users to estimate the time of concentration using a range of established methods. The model also allows estimation of the continuing runoff losses using any of the available estimation methods (i.e., constant, linearly varying or exponentially varying). Pollutants build-up in a catchment was represented by one of three pre-defined functions; power, exponential, or saturation. Similarly, pollutants wash-off was represented by one of three different functions; power, rating-curve, or exponential. The developed runoff water quality model was set-up to simulate the build-up and wash-off of total suspended solids (TSS), total phosphorus (TP) and total nitrogen (TN). The application of the model was demonstrated using available runoff and TSS field data from road and roof surfaces in the Gold Coast, Australia. The model provided excellent representation of the field data demonstrating the simplicity yet effectiveness of the proposed model.
Keywords: Catchment, continuous pollutants build-up, pollutants wash-off, runoff, runoff water quality model.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 31352847 Simulating Human Behavior in (Un)Built Environments: Using an Actor Profiling Method
Authors: Hadas Sopher, Davide Schaumann, Yehuda E. Kalay
Abstract:
This paper addresses the shortcomings of architectural computation tools in representing human behavior in built environments, prior to construction and occupancy of those environments. Evaluating whether a design fits the needs of its future users is currently done solely post construction, or is based on the knowledge and intuition of the designer. This issue is of high importance when designing complex buildings such as hospitals, where the quality of treatment as well as patient and staff satisfaction are of major concern. Existing computational pre-occupancy human behavior evaluation methods are geared mainly to test ergonomic issues, such as wheelchair accessibility, emergency egress, etc. As such, they rely on Agent Based Modeling (ABM) techniques, which emphasize the individual user. Yet we know that most human activities are social, and involve a number of actors working together, which ABM methods cannot handle. Therefore, we present an event-based model that manages the interaction between multiple Actors, Spaces, and Activities, to describe dynamically how people use spaces. This approach requires expanding the computational representation of Actors beyond their physical description, to include psychological, social, cultural, and other parameters. The model presented in this paper includes cognitive abilities and rules that describe the response of actors to their physical and social surroundings, based on the actors’ internal status. The model has been applied in a simulation of hospital wards, and showed adaptability to a wide variety of situated behaviors and interactions.Keywords: Agent based modeling, architectural design evaluation, event modeling, human behavior simulation, spatial cognition.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 10722846 Climate Change in Albania and Its Effect on Cereal Yield
Abstract:
This study is focused on analyzing climate change in Albania and its potential effects on cereal yields. Initially, monthly temperature and rainfalls in Albania were studied for the period 1960-2021. Climacteric variables are important variables when trying to model cereal yield behavior, especially when significant changes in weather conditions are observed. For this purpose, in the second part of the study, linear and nonlinear models explaining cereal yield are constructed for the same period, 1960-2021. The multiple linear regression analysis and lasso regression method are applied to the data between cereal yield and each independent variable: average temperature, average rainfall, fertilizer consumption, arable land, land under cereal production, and nitrous oxide emissions. In our regression model, heteroscedasticity is not observed, data follow a normal distribution, and there is a low correlation between factors, so we do not have the problem of multicollinearity. Machine learning methods, such as Random Forest (RF), are used to predict cereal yield responses to climacteric and other variables. RF showed high accuracy compared to the other statistical models in the prediction of cereal yield. We found that changes in average temperature negatively affect cereal yield. The coefficients of fertilizer consumption, arable land, and land under cereal production are positively affecting production. Our results show that the RF method is an effective and versatile machine-learning method for cereal yield prediction compared to the other two methods: multiple linear regression and lasso regression method.
Keywords: Cereal yield, climate change, machine learning, multiple regression model, random forest.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2492845 Neuro-Fuzzy Networks for Identification of Mathematical Model Parameters of Geofield
Authors: A. Pashayev, R. Sadiqov, C. Ardil, F. Ildiz , H. Karabork
Abstract:
The new technology of fuzzy neural networks for identification of parameters for mathematical models of geofields is proposed and checked. The effectiveness of that soft computing technology is demonstrated, especially in the early stage of modeling, when the information is uncertain and limited.
Keywords: Identification, interpolation methods, neuro-fuzzy networks, geofield.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 13322844 The Necessity of Biomass Application for Developing Combined Heat and Power (CHP)with Biogas Fuel: Case Study
Authors: F. Amin Salehi, L. Sharp, M. A. Abdoli, D.E.Cotton, K.Rezapour
Abstract:
The daily increase of organic waste materials resulting from different activities in the country is one of the main factors for the pollution of environment. Today, with regard to the low level of the output of using traditional methods, the high cost of disposal waste materials and environmental pollutions, the use of modern methods such as anaerobic digestion for the production of biogas has been prevailing. The collected biogas from the process of anaerobic digestion, as a renewable energy source similar to natural gas but with a less methane and heating value is usable. Today, with the help of technologies of filtration and proper preparation, access to biogas with features fully similar to natural gas has become possible. At present biogas is one of the main sources of supplying electrical and thermal energy and also an appropriate option to be used in four stroke engine, diesel engine, sterling engine, gas turbine, gas micro turbine and fuel cell to produce electricity. The use of biogas for different reasons which returns to socio-economic and environmental advantages has been noticed in CHP for the production of energy in the world. The production of biogas from the technology of anaerobic digestion and its application in CHP power plants in Iran can not only supply part of the energy demands in the country, but it can materialize moving in line with the sustainable development. In this article, the necessity of the development of CHP plants with biogas fuels in the country will be dealt based on studies performed from the economic, environmental and social aspects. Also to prove the importance of the establishment of these kinds of power plants from the economic point of view, necessary calculations has been done as a case study for a CHP power plant with a biogas fuel.Keywords: Anaerobic Digestion, Biogas, CHP, Organic Wastes
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 19382843 Using Smartphones as an Instrument of Early Warning and Emergency Localization
Authors: David Kubát
Abstract:
This paper suggests using smartphones and community GPS application to make alerts more accurate and therefore positively influence the entire warning process. The paper is based on formerly published paper describing a Radio-HELP system. It summarizes existing methods and lists the advantages of proposed solution. The paper analyzes the advantages and disadvantages of each possible input, processing and output of the warning system.
Keywords: e-Call, warning, information, Radio-Help, WAZE
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 19332842 Effects of Virtual Reality on the Upper Extremity Spasticity and Motor Function in Patients with Stroke: A Single Blinded Randomized Controlled Trial
Authors: K. Afsahi, M. Soheilifar, S. H. Hosseini, O. S. Esmaeili, R. Kezemi, N. Mehrbod, N. Vahed, T. Hajiahmad, N. N. Ansari
Abstract:
Background: Stroke is a disabling neurological disease. Rehabilitative therapies are important treatment methods. This clinical trial was done to compare the effects of virtual reality (VR) beside conventional rehabilitation versus conventional rehabilitation alone on the spasticity and motor function in stroke patients. Materials and methods: In this open-label randomized controlled clinical trial, 40 consecutive patients with stable first-ever ischemic stroke in the past three to 12 months that were referred to a rehabilitation clinic in Tehran, Iran in 2020 were enrolled. After signing the informed written consent form, subjects were randomly assigned by block randomization of five in each block as cases with 1:1 into two groups of 20 cases; conventional plus VR therapy group: 45-minute conventional therapy session plus 15-minute VR therapy, and conventional group: 60-minute conventional therapy session. VR rehabilitation is designed and developed with different stages. Outcomes were Modified Ashworth scale, Recovery Stage score for motor function, range of motion (ROM) of shoulder abduction/wrist extension, and patients’ satisfaction rate. Data were compared after study termination. Results: The satisfaction rate among the patients was significantly better in combination group (P = 0.003). Only wrist extension was varied between groups and was better in combination group. The variables generally had statistically significant difference (P < 0.05). Conclusion: VR plus conventional rehabilitation therapy is superior versus conventional rehabilitation alone on the wrist and elbow spasticity and motor function in patients with stroke.
Keywords: Stroke, virtual therapy, efficacy, rehabilitation.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 7542841 Integrated Subset Split for Balancing Network Utilization and Quality of Routing
Authors: S. V. Kasmir Raja, P. Herbert Raj
Abstract:
The overlay approach has been widely used by many service providers for Traffic Engineering (TE) in large Internet backbones. In the overlay approach, logical connections are set up between edge nodes to form a full mesh virtual network on top of the physical topology. IP routing is then run over the virtual network. Traffic engineering objectives are achieved through carefully routing logical connections over the physical links. Although the overlay approach has been implemented in many operational networks, it has a number of well-known scaling issues. This paper proposes a new approach to achieve traffic engineering without full-mesh overlaying with the help of integrated approach and equal subset split method. Traffic engineering needs to determine the optimal routing of traffic over the existing network infrastructure by efficiently allocating resource in order to optimize traffic performance on an IP network. Even though constraint-based routing [1] of Multi-Protocol Label Switching (MPLS) is developed to address this need, since it is not widely tested or debugged, Internet Service Providers (ISPs) resort to TE methods under Open Shortest Path First (OSPF), which is the most commonly used intra-domain routing protocol. Determining OSPF link weights for optimal network performance is an NP-hard problem. As it is not possible to solve this problem, we present a subset split method to improve the efficiency and performance by minimizing the maximum link utilization in the network via a small number of link weight modifications. The results of this method are compared against results of MPLS architecture [9] and other heuristic methods.
Keywords: Constraint based routing, Link Utilization, Subsetsplit method and Traffic Engineering.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 13962840 Statistical Texture Analysis
Authors: G. N. Srinivasan, G. Shobha
Abstract:
This paper presents an overview of the methodologies and algorithms for statistical texture analysis of 2D images. Methods for digital-image texture analysis are reviewed based on available literature and research work either carried out or supervised by the authors.Keywords: Image Texture, Texture Analysis, Statistical Approaches, Structural approaches, spectral approaches, Morphological approaches, Fractals, Fourier Transforms, Gabor Filters, Wavelet transforms.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 9382839 A Hybrid Mesh Free Local RBF- Cartesian FD Scheme for Incompressible Flow around Solid Bodies
Authors: A. Javed, K. Djidjeli, J. T. Xing, S. J. Cox
Abstract:
A method for simulating flow around the solid bodies has been presented using hybrid meshfree and mesh-based schemes. The presented scheme optimizes the computational efficiency by combining the advantages of both meshfree and mesh-based methods. In this approach, a cloud of meshfree nodes has been used in the domain around the solid body. These meshfree nodes have the ability to efficiently adapt to complex geometrical shapes. In the rest of the domain, conventional Cartesian grid has been used beyond the meshfree cloud. Complex geometrical shapes can therefore be dealt efficiently by using meshfree nodal cloud and computational efficiency is maintained through the use of conventional mesh-based scheme on Cartesian grid in the larger part of the domain. Spatial discretization of meshfree nodes has been achieved through local radial basis functions in finite difference mode (RBF-FD). Conventional finite difference scheme has been used in the Cartesian ‘meshed’ domain. Accuracy tests of the hybrid scheme have been conducted to establish the order of accuracy. Numerical tests have been performed by simulating two dimensional steady and unsteady incompressible flows around cylindrical object. Steady flow cases have been run at Reynolds numbers of 10, 20 and 40 and unsteady flow problems have been studied at Reynolds numbers of 100 and 200. Flow Parameters including lift, drag, vortex shedding, and vorticity contours are calculated. Numerical results have been found to be in good agreement with computational and experimental results available in the literature.
Keywords: CFD, Meshfree particle methods, Hybrid grid, Incompressible Navier Strokes equations, RBF-FD.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 29052838 Competitors’ Influence Analysis of a Retailer by Using Customer Value and Huff’s Gravity Model
Authors: Yepeng Cheng, Yasuhiko Morimoto
Abstract:
Customer relationship analysis is vital for retail stores, especially for supermarkets. The point of sale (POS) systems make it possible to record the daily purchasing behaviors of customers as an identification point of sale (ID-POS) database, which can be used to analyze customer behaviors of a supermarket. The customer value is an indicator based on ID-POS database for detecting the customer loyalty of a store. In general, there are many supermarkets in a city, and other nearby competitor supermarkets significantly affect the customer value of customers of a supermarket. However, it is impossible to get detailed ID-POS databases of competitor supermarkets. This study firstly focused on the customer value and distance between a customer's home and supermarkets in a city, and then constructed the models based on logistic regression analysis to analyze correlations between distance and purchasing behaviors only from a POS database of a supermarket chain. During the modeling process, there are three primary problems existed, including the incomparable problem of customer values, the multicollinearity problem among customer value and distance data, and the number of valid partial regression coefficients. The improved customer value, Huff’s gravity model, and inverse attractiveness frequency are considered to solve these problems. This paper presents three types of models based on these three methods for loyal customer classification and competitors’ influence analysis. In numerical experiments, all types of models are useful for loyal customer classification. The type of model, including all three methods, is the most superior one for evaluating the influence of the other nearby supermarkets on customers' purchasing of a supermarket chain from the viewpoint of valid partial regression coefficients and accuracy.Keywords: Customer value, Huff's Gravity Model, POS, retailer.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 6122837 Communication Engineering Curriculum (Past, Present and the Future)
Authors: Abdurazzag Ali Aburas, Indira Rustempasic, Indira Muhic, Busra Gheith Yildiz
Abstract:
At present time, competition, unpredictable fluctuations have made communication engineering education in the global sphere really difficult. Confront with new situation in the engineering education sector. Communication engineering education has to be reformed and ready to use more advanced technologies. We realized that one of the general problems of student`s education is that after graduating from their universities, they are not prepared to face the real life challenges and full skilled to work in industry. They are prepared only to think like engineers and professionals but they also need to possess some others non-technical skills. In today-s environment, technical competence alone is not sufficient for career success. Employers want employees (graduate engineers) who have good oral and written communication (soft) skills. It does require for team work, business awareness, organization, management skills, responsibility, initiative, problem solving and IT competency. This proposed curriculum brings interactive, creative, interesting, effective learning methods, which includes online education, virtual labs, practical work, problem-based learning (PBL), and lectures given by industry experts. Giving short assignments, presentations, reports, research papers and projects students can significantly improve their non-technical skills. Also, we noticed the importance of using ICT technologies in engineering education which used by students and teachers, and included that into proposed teaching and learning methods. We added collaborative learning between students through team work which builds theirs skills besides course materials. The prospective on this research that we intent to update communication engineering curriculum in order to get fully constructed engineer students to ready for real industry work.
Keywords: communication engineering, curriculum education, ICT, industry
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 18982836 Predictive Semi-Empirical NOx Model for Diesel Engine
Authors: Saurabh Sharma, Yong Sun, Bruce Vernham
Abstract:
Accurate prediction of NOx emission is a continuous challenge in the field of diesel engine-out emission modeling. Performing experiments for each conditions and scenario cost significant amount of money and man hours, therefore model-based development strategy has been implemented in order to solve that issue. NOx formation is highly dependent on the burn gas temperature and the O2 concentration inside the cylinder. The current empirical models are developed by calibrating the parameters representing the engine operating conditions with respect to the measured NOx. This makes the prediction of purely empirical models limited to the region where it has been calibrated. An alternative solution to that is presented in this paper, which focus on the utilization of in-cylinder combustion parameters to form a predictive semi-empirical NOx model. The result of this work is shown by developing a fast and predictive NOx model by using the physical parameters and empirical correlation. The model is developed based on the steady state data collected at entire operating region of the engine and the predictive combustion model, which is developed in Gamma Technology (GT)-Power by using Direct Injected (DI)-Pulse combustion object. In this approach, temperature in both burned and unburnt zone is considered during the combustion period i.e. from Intake Valve Closing (IVC) to Exhaust Valve Opening (EVO). Also, the oxygen concentration consumed in burnt zone and trapped fuel mass is also considered while developing the reported model. Several statistical methods are used to construct the model, including individual machine learning methods and ensemble machine learning methods. A detailed validation of the model on multiple diesel engines is reported in this work. Substantial numbers of cases are tested for different engine configurations over a large span of speed and load points. Different sweeps of operating conditions such as Exhaust Gas Recirculation (EGR), injection timing and Variable Valve Timing (VVT) are also considered for the validation. Model shows a very good predictability and robustness at both sea level and altitude condition with different ambient conditions. The various advantages such as high accuracy and robustness at different operating conditions, low computational time and lower number of data points requires for the calibration establishes the platform where the model-based approach can be used for the engine calibration and development process. Moreover, the focus of this work is towards establishing a framework for the future model development for other various targets such as soot, Combustion Noise Level (CNL), NO2/NOx ratio etc.
Keywords: Diesel engine, machine learning, NOx emission, semi-empirical.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 855