Search results for: (G'/G) expansion method.
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 8332

Search results for: (G'/G) expansion method.

1192 Machine Vision System for Automatic Weeding Strategy in Oil Palm Plantation using Image Filtering Technique

Authors: Kamarul Hawari Ghazali, Mohd. Marzuki Mustafa, Aini Hussain

Abstract:

Machine vision is an application of computer vision to automate conventional work in industry, manufacturing or any other field. Nowadays, people in agriculture industry have embarked into research on implementation of engineering technology in their farming activities. One of the precision farming activities that involve machine vision system is automatic weeding strategy. Automatic weeding strategy in oil palm plantation could minimize the volume of herbicides that is sprayed to the fields. This paper discusses an automatic weeding strategy in oil palm plantation using machine vision system for the detection and differential spraying of weeds. The implementation of vision system involved the used of image processing technique to analyze weed images in order to recognized and distinguished its types. Image filtering technique has been used to process the images as well as a feature extraction method to classify the type of weed images. As a result, the image processing technique contributes a promising result of classification to be implemented in machine vision system for automated weeding strategy.

Keywords: Machine vision, Automatic Weeding Strategy, filter, feature extraction

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1870
1191 A General Framework for Knowledge Discovery Using High Performance Machine Learning Algorithms

Authors: S. Nandagopalan, N. Pradeep

Abstract:

The aim of this paper is to propose a general framework for storing, analyzing, and extracting knowledge from two-dimensional echocardiographic images, color Doppler images, non-medical images, and general data sets. A number of high performance data mining algorithms have been used to carry out this task. Our framework encompasses four layers namely physical storage, object identification, knowledge discovery, user level. Techniques such as active contour model to identify the cardiac chambers, pixel classification to segment the color Doppler echo image, universal model for image retrieval, Bayesian method for classification, parallel algorithms for image segmentation, etc., were employed. Using the feature vector database that have been efficiently constructed, one can perform various data mining tasks like clustering, classification, etc. with efficient algorithms along with image mining given a query image. All these facilities are included in the framework that is supported by state-of-the-art user interface (UI). The algorithms were tested with actual patient data and Coral image database and the results show that their performance is better than the results reported already.

Keywords: Active Contour, Bayesian, Echocardiographic image, Feature vector.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1718
1190 Towards the Use of Software Product Metrics as an Indicator for Measuring Mobile Applications Power Consumption

Authors: Ching Kin Keong, Koh Tieng Wei, Abdul Azim Abd. Ghani, Khaironi Yatim Sharif

Abstract:

Maintaining factory default battery endurance rate over time in supporting huge amount of running applications on energy-restricted mobile devices has created a new challenge for mobile applications developer. While delivering customers’ unlimited expectations, developers are barely aware of efficient use of energy from the application itself. Thus, developers need a set of valid energy consumption indicators in assisting them to develop energy saving applications. In this paper, we present a few software product metrics that can be used as an indicator to measure energy consumption of Android-based mobile applications in the early of design stage. In particular, Trepn Profiler (Power profiling tool for Qualcomm processor) has used to collect the data of mobile application power consumption, and then analyzed for the 23 software metrics in this preliminary study. The results show that McCabe cyclomatic complexity, number of parameters, nested block depth, number of methods, weighted methods per class, number of classes, total lines of code and method lines have direct relationship with power consumption of mobile application.

Keywords: Battery endurance, software metrics, mobile application, power consumption.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1945
1189 Perceptions of Greenhouse Vegetable Growers Regarding Use of Biological Control Practices: A Case Study in Jiroft County, Iran

Authors: Hossein Shabanali Fami, Omid Sharifi, Javad Ghasemi, Mahtab Pouratashi, Mona Sadat Moghadasian

Abstract:

The main purpose of this study was to investigate perception of greenhouse vegetable growers regarding use of biological control practices during the growing season. The statistical population of the study included greenhouse vegetable growers in Jiroft county (N=1862). A sample of 137 vegetable growers was selected, using random sampling method. Data were collected via a questionnaire. The validity of the instrument was obtained by the faculty members of the Department of Agricultural Development and Management in the University of Tehran. Cronbach’s alpha was applied to estimate the reliability which showed a high reliability for the instrument. Data was analyzed using SPSS/Windows 13.5. The results revealed that greenhouse vegetable growers had moderate level of perception regarding biological control practices. Levels of vegetable growers’ perceptions regarding biological control practices were different on the basis of their academic qualifications as well as educational level and job. In addition, the results indicated that about 54.1% of variations in vegetable growers’ perceptions could be explained by variables such as awareness of biological control practices, knowledge on pests, annual production and age.

Keywords: Greenhouse, biological control, biological agents, perception, vegetable grower.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1709
1188 Role of Organic Wastewater Constituents in Iron Redox Cycling for Ferric Sludge Reuse in the Fenton-Based Treatment

Authors: J. Bolobajev, M. Trapido, A. Goi

Abstract:

The practical application of the Fenton-based treatment method for organic compounds-contaminated water purification is limited mainly because of the large amount of ferric sludge formed during the treatment, where ferrous iron (Fe(II)) is used as the activator of the hydrogen peroxide oxidation processes. Reuse of ferric sludge collected from clarifiers to substitute Fe(II) salts allows reducing the total cost of Fenton-type treatment technologies and minimizing the accumulation of hazardous ferric waste. Dissolution of ferric iron (Fe(III)) from the sludge to liquid phase at acidic pH and autocatalytic transformation of Fe(III) to Fe(II) by phenolic compounds (tannic acid, lignin, phenol, catechol, pyrogallol and hydroquinone) added or present as water/wastewater constituents were found to be essentially involved in the Fenton-based oxidation mechanism. Observed enhanced formation of highly reactive species, hydroxyl radicals, resulted in a substantial organic contaminant degradation increase. Sludge reuse at acidic pH and in the presence of ferric iron reductants is a novel strategy in the Fenton-based treatment application for organic compounds-contaminated water purification.

Keywords: Ferric sludge reuse, ferric iron reductant, water treatment, organic pollutant.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1672
1187 Lean Environmental Management Integration System (LEMIS) Framework Development

Authors: Puvanasvaran, A. P., Suresh V., N. Norazlin

Abstract:

The Lean Environmental Management Integration System (LEMIS) framework development is integration between lean core element and ISO 14001. The curiosity on the relationship between continuous improvement and sustainability of lean implementation has influenced this study toward LEMIS. Characteristic of ISO 14001 standard clauses and core elements of lean principles are explored from past studies and literature reviews. Survey was carried out on ISO 14001 certified companies to examine continual improvement by implementing the ISO 14001 standard. The study found that there is a significant and positive relationship between Lean Principles: value, value stream, flow, pull and perfection with the ISO 14001 requirements. LEMIS is significant to support the continuous improvement and sustainability. The integration system can be implemented to any manufacturing company. It gives awareness on the importance on why organizations need to sustain its environmental management system. In the meantime, the lean principle can be adapted in order to streamline daily activities of the company. Throughout the study, it had proven that there is no sacrifice or trade-off between lean principles with ISO 14001 requirements. The framework developed in the study can be further simplified in the future, especially the method of crossing each sub requirements of ISO 14001 standard with the core elements of Lean principles in this study.

Keywords: LEMIS, ISO 14001, integration, framework.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2388
1186 Modeling and Simulating Reaction-Diffusion Systems with State-Dependent Diffusion Coefficients

Authors: Paola Lecca, Lorenzo Dematte, Corrado Priami

Abstract:

The present models and simulation algorithms of intracellular stochastic kinetics are usually based on the premise that diffusion is so fast that the concentrations of all the involved species are homogeneous in space. However, recents experimental measurements of intracellular diffusion constants indicate that the assumption of a homogeneous well-stirred cytosol is not necessarily valid even for small prokaryotic cells. In this work a mathematical treatment of diffusion that can be incorporated in a stochastic algorithm simulating the dynamics of a reaction-diffusion system is presented. The movement of a molecule A from a region i to a region j of the space is represented as a first order reaction Ai k- ! Aj , where the rate constant k depends on the diffusion coefficient. The diffusion coefficients are modeled as function of the local concentration of the solutes, their intrinsic viscosities, their frictional coefficients and the temperature of the system. The stochastic time evolution of the system is given by the occurrence of diffusion events and chemical reaction events. At each time step an event (reaction or diffusion) is selected from a probability distribution of waiting times determined by the intrinsic reaction kinetics and diffusion dynamics. To demonstrate the method the simulation results of the reaction-diffusion system of chaperoneassisted protein folding in cytoplasm are shown.

Keywords: Reaction-diffusion systems, diffusion coefficient, stochastic simulation algorithm.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1529
1185 Building Information Modeling-Based Approach for Automatic Quantity Take-off and Cost Estimation

Authors: Lo Kar Yin, Law Ka Mei

Abstract:

Architectural, engineering, construction and operations (AECO) industry practitioners have been well adapting to the dynamic construction market from the fundamental training of its disciplines. As further triggered by the pandemic since 2019, great steps are taken in virtual environment and the best collaboration is strived with project teams without boundaries. With adoption of Building Information Modeling-based approach and qualitative analysis, this paper is to review quantity take-off (QTO) and cost estimation process through modeling techniques in liaison with suppliers, fabricators, subcontractors, contractors, designers, consultants and services providers in the construction industry value chain for automatic project cost budgeting, project cost control and cost evaluation on design options of in-situ reinforced-concrete construction and Modular Integrated Construction (MiC) at design stage, variation of works and cash flow/spending analysis at construction stage as far as practicable, with a view to sharing the findings for enhancing mutual trust and co-operation among AECO industry practitioners. It is to foster development through a common prototype of design and build project delivery method in NEC4 Engineering and Construction Contract (ECC) Options A and C.

Keywords: Building Information Modeling, cost estimation, quantity take-off, modeling techniques.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 719
1184 A Simulation Method to Find the Optimal Design of Photovoltaic Home System in Malaysia, Case Study: A Building Integrated Photovoltaic in Putra Jaya

Authors: Riza Muhida, Maisarah Ali, Puteri Shireen Jahn Kassim, Muhammad Abu Eusuf, Agus G.E. Sutjipto, Afzeri

Abstract:

Over recent years, the number of building integrated photovoltaic (BIPV) installations for home systems have been increasing in Malaysia. The paper concerns an analysis - as part of current Research and Development (R&D) efforts - to integrate photovoltaics as an architectural feature of a detached house in the new satellite township of Putrajaya, Malaysia. The analysis was undertaken using calculation and simulation tools to optimize performance of BIPV home system. In this study, a the simulation analysis was undertaken for selected bungalow units based on a long term recorded weather data for city of Kuala Lumpur. The simulation and calculation was done with consideration of a PV panels' tilt and direction, shading effect and economical considerations. A simulation of the performance of a grid connected BIPV house in Kuala Lumpur was undertaken. This case study uses a 60 PV modules with power output of 2.7 kW giving an average of PV electricity output is 255 kWh/month..

Keywords: Building integrated photovoltaic, Malaysia, Simulation, panels' tilt and direction.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2296
1183 The Effect of Board Composition and Ownership Concentration on Earnings Management: Evidence from IRAN

Authors: F. Rahnamay Roodposhti, S. A. Nabavi Chashmi

Abstract:

The role of corporate governance is to reduce the divergence of interests between shareholders and managers. The role of corporate governance is more useful when managers have an incentive to deviate from shareholders- interests. One example of management-s deviation from shareholders- interests is the management of earnings through the use of accounting accruals. This paper examines the association between corporate governance internal mechanisms ownership concentration, board independence, the existence of CEO-Chairman duality and earnings management. Firm size and leverage are control variables. The population used in this study comprises firms listed on the Tehran Stock Exchange (TSE) between 2004 and 2008, the sample comprises 196 firms. Panel Data method is employed as technique to estimate the model. We find that there is negative significant association between ownership concentration and board independence manage earnings with earnings management, there is negative significant association between the existence of CEO-Chairman duality and earnings management. This study also found a positive significant association between control variable (firm size and leverage) and earnings management.

Keywords: Earnings management, board independence, ownership concentration, corporate governance.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3922
1182 Characterization and Geochemical Modeling of Cu and Zn Sorption Using Mixed Mineral Systems Injected with Iron Sulfide under Sulfidic-Anoxic Conditions I: Case Study of Cwmheidol Mine Waste Water, Wales, United Kingdom

Authors: D. E. Egirani, J. E. Andrews, A. R. Baker

Abstract:

This study investigates sorption of Cu and Zn contained in natural mine wastewater, using mixed mineral systems in sulfidic-anoxic condition. The mine wastewater was obtained from disused mine workings at Cwmheidol in Wales, United Kingdom. These contaminants flow into water courses. These water courses include River Rheidol. In this River fishing activities exist. In an attempt to reduce Cu-Zn levels of fish intake in the watercourses, single mineral systems and 1:1 mixed mineral systems of clay and goethite were tested with the mine waste water for copper and zinc removal at variable pH. Modelling of hydroxyl complexes was carried out using phreeqc method. Reactions using batch mode technique was conducted at room temperature. There was significant differences in the behaviour of copper and zinc removal using mixed mineral systems when compared  to single mineral systems. All mixed mineral systems sorb more Cu than Zn when tested with mine wastewater.

Keywords: Cu- Zn, hydroxyl complexes, kinetics, mixed mineral systems, reactivity

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 928
1181 Effect of Rubber Treatment on Compressive Strength and Modulus of Elasticity of Self-Compacting Rubberized Concrete

Authors: I. Miličević, M. Hadzima Nyarko, R. Bušić, J. Simonović Radosavljević, M. Prokopijević, K. Vojisavljević

Abstract:

This paper investigates the effects of different treatment methods of rubber aggregates for self-compacting concrete (SCC) on compressive strength and modulus of elasticity. SCC mixtures with 10% replacement of fine aggregate with crumb rubber by total aggregate volume and with different aggregate treatment methods were investigated. The rubber aggregate was treated in three different methods: dry process, water-soaking, and NaOH treatment plus water soaking. Properties of SCC in a fresh and hardened state were tested and evaluated. Scanning electron microscope (SEM) analysis of three different SCC patches were made and discussed. It was observed that applying the proposed NaOH plus water soaking method resulted in the improvement of fresh and hardened concrete properties. It resulted in a more uniform distribution of rubber particles in the cement matrix, a better bond between rubber particles and the cement matrix, and higher compressive strength of SCC rubberized concrete.

Keywords: Compressive strength, modulus of elasticity, NaOH treatment, rubber aggregate, self-compacting rubberized concrete, scanning electron microscope analysis.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 647
1180 Strongly Coupled Finite Element Formulation of Electromechanical Systems with Integrated Mesh Morphing using Radial Basis Functions

Authors: D. Kriebel, J. E. Mehner

Abstract:

The paper introduces a method to efficiently simulate nonlinear changing electrostatic fields occurring in micro-electromechanical systems (MEMS). Large deflections of the capacitor electrodes usually introduce nonlinear electromechanical forces on the mechanical system. Traditional finite element methods require a time-consuming remeshing process to capture exact results for this physical domain interaction. In order to accelerate the simulation process and eliminate the remeshing process, a formulation of a strongly coupled electromechanical transducer element will be introduced which uses a combination of finite-element with an advanced mesh morphing technique using radial basis functions (RBF). The RBF allows large geometrical changes of the electric field domain while retain high element quality of the deformed mesh. Coupling effects between mechanical and electrical domains are directly included within the element formulation. Fringing field effects are described accurate by using traditional arbitrary shape functions.

Keywords: electromechanical, electric field, transducer, simulation, modeling, finite-element, mesh morphing, radial basis function

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 522
1179 Investigating Daylight Quality in Malaysian Government Office Buildings Through Daylight Factor and Surface Luminance

Authors: Mohd Zin Kandar, Mohd Sabere Sulaiman, Yong Razidah Rashid, Dilshan Remaz Ossen, Aminatuzuhariah MAbdullah, Lim Yaik Wah, Mansour Nikpour

Abstract:

In recent years, there has been an increasing interest in using daylight to save energy in buildings. In tropical regions, daylighting is always an energy saver. On the other hand, daylight provides visual comfort. According to standards, it shows that many criteria should be taken into consideration in order to have daylight utilization and visual comfort. The current standard in Malaysia, MS 1525 does not provide sufficient guideline. Hence, more research is needed on daylight performance. If architects do not consider daylight design, it not only causes inconvenience in working spaces but also causes more energy consumption as well as environmental pollution. This research had surveyed daylight performance in 5 selected office buildings from different area of Malaysian through experimental method. Several parameters of daylight quality such as daylight factor, surface luminance and surface luminance ratio were measured in different rooms in each building. The result of this research demonstrated that most of the buildings were not designed for daylight utilization. Therefore, it is very important that architects follow the daylight design recommendation to reduce consumption of electric power for artificial lighting while the sufficient quality of daylight is available.

Keywords: Daylight factor, Field measurement, Daylighting quality, Tropical

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3487
1178 Contribution of Electrochemical Treatment in Treating Textile Dye Wastewater

Authors: Usha N. Murthy, Rekha H. B., Mahaveer Devoor

Abstract:

The introduction of more stringent pollution regulations, in relation to financial and social pressures for sustainable development, has pressed toward limiting the volumes of industrial and domestic effluents discharged into the environment - as well as to increase the efforts within research and development of new or more efficient wastewater treatment technologies. Considering both discharge volume and effluent composition, wastewater generated by the textile industry is rated as the most polluting among all industrial sectors. The pollution load is mainly due to spent dye baths, which are composed of unreacted dyes, dispersing agents, surfactants, salts and organics. In the present investigation, the textile dye wastewater was characterized by high color, chemical oxygen demand (COD), total dissolved solids (TDS) and pH. Electrochemical oxidation process for four plate electrodes was carried out at five different current intensities, out of which 0.14A has achieved maximum percentage removal of COD with 75% and 83% of color. The COD removal rate in kg COD/h/m2 decreases with increase in the current intensity. The energy consumption increases with increase in the current intensity. Hence, textile dye wastewater can be effectively pretreated by electrochemical oxidation method where the process limits objectionable color while leaving the COD associated with organics left for natural degradation thus causing a sustainable reduction in pollution load.

Keywords: Electrochemical treatment, COD, color.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2397
1177 Implementation-Oriented Discussion for Historical and Cultural Villages’ Conservation Planning

Authors: Xing Zhang

Abstract:

Since the State Council of China issued the Regulations on the Conservation of Historical Cultural Towns and Villages in 2008, formulation of conservation planning has been carried out in national, provincial and municipal historical and cultural villages for protection needs, which provides a legal basis for inheritance of historical culture and protection of historical resources. Although the quantity and content of the conservation planning are continually increasing, the implementation and application are still ambiguous. To solve the aforementioned problems, this paper explores methods to enhance the implementation of conservation planning from the perspective of planning formulation. Specifically, the technical framework of "overall objectives planning - sub-objectives planning - zoning guidelines - implementation by stages" is proposed to implement the planning objectives in different classifications and stages. Then combined with details of the Qiqiao historical and cultural village conservation planning project in Ningbo, five sub-objectives are set, which are implemented through the village zoning guidelines. At the same time, the key points and specific projects in the near-term, medium-term and long-term work are clarified, and the spatial planning is transformed into the action plan with time scale. The proposed framework and method provide a reference for the implementation and management of the conservation planning of historical and cultural villages in the future.

Keywords: Conservation planning, planning by stages, planning implementation, zoning guidelines.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 784
1176 Optimizing Dialogue Strategy Learning Using Learning Automata

Authors: G. Kumaravelan, R. Sivakumar

Abstract:

Modeling the behavior of the dialogue management in the design of a spoken dialogue system using statistical methodologies is currently a growing research area. This paper presents a work on developing an adaptive learning approach to optimize dialogue strategy. At the core of our system is a method formalizing dialogue management as a sequential decision making under uncertainty whose underlying probabilistic structure has a Markov Chain. Researchers have mostly focused on model-free algorithms for automating the design of dialogue management using machine learning techniques such as reinforcement learning. But in model-free algorithms there exist a dilemma in engaging the type of exploration versus exploitation. Hence we present a model-based online policy learning algorithm using interconnected learning automata for optimizing dialogue strategy. The proposed algorithm is capable of deriving an optimal policy that prescribes what action should be taken in various states of conversation so as to maximize the expected total reward to attain the goal and incorporates good exploration and exploitation in its updates to improve the naturalness of humancomputer interaction. We test the proposed approach using the most sophisticated evaluation framework PARADISE for accessing to the railway information system.

Keywords: Dialogue management, Learning automata, Reinforcement learning, Spoken dialogue system

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1613
1175 An Authentic Algorithm for Ciphering and Deciphering Called Latin Djokovic

Authors: Diogen Babuc

Abstract:

The question that is a motivation of writing is how many devote themselves to discovering something in the world of science where much is discerned and revealed, but at the same time, much is unknown. The insightful elements of this algorithm are the ciphering and deciphering algorithms of Playfair, Caesar, and Vigen`ere. Only a few of their main properties are taken and modified, with the aim of forming a specific functionality of the algorithm called Latin Djokovic. Specifically, a string is entered as input data. A key k is given, with a random value between the values a and b = a+3. The obtained value is stored in a variable with the aim of being constant during the run of the algorithm. In correlation to the given key, the string is divided into several groups of substrings, and each substring has a length of k characters. The next step involves encoding each substring from the list of existing substrings. Encoding is performed using the basis of Caesar algorithm, i.e. shifting with k characters. However, that k is incremented by 1 when moving to the next substring in that list. When the value of k becomes greater than b + 1, it will return to its initial value. The algorithm is executed, following the same procedure, until the last substring in the list is traversed. Using this polyalphabetic method, ciphering and deciphering of strings are achieved. The algorithm also works for a 100-character string. The x character is not used when the number of characters in a substring is incompatible with the expected length. The algorithm is simple to implement, but it is questionable if it works better than the other methods, from the point of view of execution time and storage space.

Keywords: Ciphering and deciphering, Authentic Algorithm, Polyalphabetic Cipher, Random Key, methods comparison.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 198
1174 Investigating the Effective Parameters in Determining the Type of Traffic Congestion Pricing Schemes in Urban Streets

Authors: Saeed Sayyad Hagh Shomar

Abstract:

Traffic congestion pricing – as a strategy in travel demand management in urban areas to reduce traffic congestion, air pollution and noise pollution – has drawn many attentions towards itself. Unlike the satisfying findings in this method, there are still problems in determining the best functional congestion pricing scheme with regard to the situation. The so-called problems in this process will result in further complications and even the scheme failure. That is why having proper knowledge of the significance of congestion pricing schemes and the effective factors in choosing them can lead to the success of this strategy. In this study, first, a variety of traffic congestion pricing schemes and their components are introduced; then, their functional usage is discussed. Next, by analyzing and comparing the barriers, limitations and advantages, the selection criteria of pricing schemes are described. The results, accordingly, show that the selection of the best scheme depends on various parameters. Finally, based on examining the effective parameters, it is concluded that the implementation of area-based schemes (cordon and zonal) has been more successful in non-diversion of traffic. That is considering the topology of the cities and the fact that traffic congestion is often created in the city centers, area-based schemes would be notably functional and appropriate.

Keywords: Congestion pricing, demand management, flat toll, variable toll.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 613
1173 Early Diagnosis of Alzheimer's Disease Using a Combination of Images Processing and Brain Signals

Authors: E. Irankhah, M. Zarif, E. Mazrooei Rad, K. Ghandehari

Abstract:

Alzheimer's prevalence is on the rise, and the disease comes with problems like cessation of treatment, high cost of treatment, and the lack of early detection methods. The pathology of this disease causes the formation of protein deposits in the brain of patients called plaque amyloid. Generally, the diagnosis of this disease is done by performing tests such as a cerebrospinal fluid, CT scan, MRI, and spinal cord fluid testing, or mental testing tests and eye tracing tests. In this paper, we tried to use the Medial Temporal Atrophy (MTA) method and the Leave One Out (LOO) cycle to extract the statistical properties of the three Fz, Pz, and Cz channels of ERP signals for early diagnosis of this disease. In the process of CT scan images, the accuracy of the results is 81% for the healthy person and 88% for the severe patient. After the process of ERP signaling, the accuracy of the results for a healthy person in the delta band in the Cz channel is 81% and in the alpha band the Pz channel is 90%. In the results obtained from the signal processing, the results of the severe patient in the delta band of the Cz channel were 89% and in the alpha band Pz channel 92%.

Keywords: Alzheimer's disease, image and signal processing, medial temporal atrophy, LOO Cycle.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2062
1172 Optimization of Hemp Fiber Reinforced Concrete for Mix Design Method

Authors: Zoe Chang, Max Williams, Gautham Das

Abstract:

The purpose of this study is to evaluate the incorporation of hemp fibers (HF) in concrete. Hemp fiber reinforced concrete (HFRC) is becoming more popular as an alternative for regular mix designs. This study was done to evaluate the compressive strength of HFRC regarding mix procedure. HF were obtained from the manufacturer and hand processed to ensure uniformity in width and length. The fibers were added to concrete as both wet and dry mix to investigate and optimize the mix design process. Results indicated that the dry mix had a compressive strength of 1157 psi compared to the wet mix of 985 psi. This dry mix compressive strength was within range of the standard mix compressive strength of 1533 psi. The statistical analysis revealed that the mix design process needs further optimization and uniformity concerning the addition of HF. Regression analysis revealed that the standard mix design had a coefficient of 0.9 as compared to the dry mix of 0.375 indicating a variation in the mixing process. While completing the dry mix, the addition of plain HF caused them to intertwine creating lumps and inconsistency. However, during the wet mixing process, combining water and HF before incorporation allows the fibers to uniformly disperse within the mix hence the regression analysis indicated a better coefficient of 0.55. This study concludes that HRFC is a viable alternative to regular mixes however more research surrounding its characteristics needs to be conducted.

Keywords: hemp fibers, hemp reinforced concrete, wet and dry, freeze thaw testing, compressive strength

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 563
1171 Blind Image Deconvolution by Neural Recursive Function Approximation

Authors: Jiann-Ming Wu, Hsiao-Chang Chen, Chun-Chang Wu, Pei-Hsun Hsu

Abstract:

This work explores blind image deconvolution by recursive function approximation based on supervised learning of neural networks, under the assumption that a degraded image is linear convolution of an original source image through a linear shift-invariant (LSI) blurring matrix. Supervised learning of neural networks of radial basis functions (RBF) is employed to construct an embedded recursive function within a blurring image, try to extract non-deterministic component of an original source image, and use them to estimate hyper parameters of a linear image degradation model. Based on the estimated blurring matrix, reconstruction of an original source image from a blurred image is further resolved by an annealed Hopfield neural network. By numerical simulations, the proposed novel method is shown effective for faithful estimation of an unknown blurring matrix and restoration of an original source image.

Keywords: Blind image deconvolution, linear shift-invariant(LSI), linear image degradation model, radial basis functions (rbf), recursive function, annealed Hopfield neural networks.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2065
1170 Modelling and Simulation of Cascaded H-Bridge Multilevel Single Source Inverter Using PSIM

Authors: Gaddafi S. Shehu, T. Yalcinoz, Abdullahi B. Kunya

Abstract:

Multilevel inverters such as flying capacitor, diodeclamped, and cascaded H-bridge inverters are very popular particularly in medium and high power applications. This paper focuses on a cascaded H-bridge module using a single direct current (DC) source in order to generate an 11-level output voltage. The noble approach reduces the number of switches and gate drivers, in comparison with a conventional method. The anticipated topology produces more accurate result with an isolation transformer at high switching frequency. Different modulation techniques can be used for the multilevel inverter, but this work features modulation techniques known as selective harmonic elimination (SHE).This modulation approach reduces the number of carriers with reduction in Switching Losses, Total Harmonic Distortion (THD), and thereby increasing Power Quality (PQ). Based on the simulation result obtained, it appears SHE has the ability to eliminate selected harmonics by chopping off the fundamental output component. The performance evaluation of the proposed cascaded multilevel inverter is performed using PSIM simulation package and THD of 0.94% is obtained.

Keywords: Cascaded H-bridge Multilevel Inverter, Power Quality, Selective Harmonic Elimination.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 5098
1169 Attention Based Fully Convolutional Neural Network for Simultaneous Detection and Segmentation of Optic Disc in Retinal Fundus Images

Authors: Sandip Sadhukhan, Arpita Sarkar, Debprasad Sinha, Goutam Kumar Ghorai, Gautam Sarkar, Ashis K. Dhara

Abstract:

Accurate segmentation of the optic disc is very important for computer-aided diagnosis of several ocular diseases such as glaucoma, diabetic retinopathy, and hypertensive retinopathy. The paper presents an accurate and fast optic disc detection and segmentation method using an attention based fully convolutional network. The network is trained from scratch using the fundus images of extended MESSIDOR database and the trained model is used for segmentation of optic disc. The false positives are removed based on morphological operation and shape features. The result is evaluated using three-fold cross-validation on six public fundus image databases such as DIARETDB0, DIARETDB1, DRIVE, AV-INSPIRE, CHASE DB1 and MESSIDOR. The attention based fully convolutional network is robust and effective for detection and segmentation of optic disc in the images affected by diabetic retinopathy and it outperforms existing techniques.

Keywords: Ocular diseases, retinal fundus image, optic disc detection and segmentation, fully convolutional network, overlap measure.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 784
1168 Use of Fuzzy Logic in the Corporate Reputation Assessment: Stock Market Investors’ Perspective

Authors: Tomasz L. Nawrocki, Danuta Szwajca

Abstract:

The growing importance of reputation in building enterprise value and achieving long-term competitive advantage creates the need for its measurement and evaluation for the management purposes (effective reputation and its risk management). The paper presents practical application of self-developed corporate reputation assessment model from the viewpoint of stock market investors. The model has a pioneer character and example analysis performed for selected industry is a form of specific test for this tool. In the proposed solution, three aspects - informational, financial and development, as well as social ones - were considered. It was also assumed that the individual sub-criteria will be based on public sources of information, and as the calculation apparatus, capable of obtaining synthetic final assessment, fuzzy logic will be used. The main reason for developing this model was to fulfill the gap in the scope of synthetic measure of corporate reputation that would provide higher degree of objectivity by relying on "hard" (not from surveys) and publicly available data. It should be also noted that results obtained on the basis of proposed corporate reputation assessment method give possibilities of various internal as well as inter-branch comparisons and analysis of corporate reputation impact.

Keywords: Corporate reputation, fuzzy logic, fuzzy model, stock market investors.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1373
1167 Model Canvas and Process for Educational Game Design in Outcome-Based Education

Authors: Ratima Damkham, Natasha Dejdumrong, Priyakorn Pusawiro

Abstract:

This paper explored the solution in game design to help game designers in the educational game designing using digital educational game model canvas (DEGMC) and digital educational game form (DEGF) based on Outcome-based Education program. DEGMC and DEGF can help designers develop an overview of the game while designing and planning their own game. The way to clearly assess players’ ability from learning outcomes and support their game learning design is by using the tools. Designers can balance educational content and entertainment in designing a game by using the strategies of the Business Model Canvas and design the gameplay and players’ ability assessment from learning outcomes they need by referring to the Constructive Alignment. Furthermore, they can use their design plan in this research to write their Game Design Document (GDD). The success of the research was evaluated by four experts’ perspectives in the education and computer field. From the experiments, the canvas and form helped the game designers model their game according to the learning outcomes and analysis of their own game elements. This method can be a path to research an educational game design in the future.

Keywords: Constructive alignment, constructivist theory, educational game, outcome-based education.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 860
1166 Riemannian Manifolds for Brain Extraction on Multi-modal Resonance Magnetic Images

Authors: Mohamed Gouskir, Belaid Bouikhalene, Hicham Aissaoui, Benachir Elhadadi

Abstract:

In this paper, we present an application of Riemannian geometry for processing non-Euclidean image data. We consider the image as residing in a Riemannian manifold, for developing a new method to brain edge detection and brain extraction. Automating this process is a challenge due to the high diversity in appearance brain tissue, among different patients and sequences. The main contribution, in this paper, is the use of an edge-based anisotropic diffusion tensor for the segmentation task by integrating both image edge geometry and Riemannian manifold (geodesic, metric tensor) to regularize the convergence contour and extract complex anatomical structures. We check the accuracy of the segmentation results on simulated brain MRI scans of single T1-weighted, T2-weighted and Proton Density sequences. We validate our approach using two different databases: BrainWeb database, and MRI Multiple sclerosis Database (MRI MS DB). We have compared, qualitatively and quantitatively, our approach with the well-known brain extraction algorithms. We show that using a Riemannian manifolds to medical image analysis improves the efficient results to brain extraction, in real time, outperforming the results of the standard techniques.

Keywords: Riemannian manifolds, Riemannian Tensor, Brain Segmentation, Non-Euclidean data, Brain Extraction.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1666
1165 Solid Dispersions of Cefixime Using β-Cyclodextrin: Characterization and in vitro Evaluation

Authors: Nagasamy Venkatesh Dhandapani, Amged Awad El-Gied

Abstract:

Cefixime, a BCS class II drug, is insoluble in water but freely soluble in acetone and in alcohol. The aqueous solubility of cefixime in water is poor and exhibits exceptionally slow and intrinsic dissolution rate. In the present study, cefixime and β-Cyclodextrin (β-CD) solid dispersions were prepared with a view to study the effect and influence of β-CD on the solubility and dissolution rate of this poorly aqueous soluble drug. Phase solubility profile revealed that the solubility of cefixime was increased in the presence of β-CD and was classified as AL-type. Effect of variable, such as drug:carrier ratio, was studied. Physical characterization of the solid dispersion was characterized by Fourier transform infrared spectroscopy (FT-IR) and Differential scanning calorimetry (DSC). These studies revealed that a distinct loss of drug crystallinity in the solid molecular dispersions is ostensibly accounting for enhancement of dissolution rate in distilled water. The drug release from the prepared solid dispersion exhibited a first order kinetics. Solid dispersions of cefixime showed a 6.77 times fold increase in dissolution rate over the pure drug.

Keywords: Cefixime, β-Cyclodextrin, solid dispersions, kneading method, dissolution, release kinetics.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1613
1164 Impact of Zn/Cr Ratio on ZnCrOx-SAPO-34 Bifunctional Catalyst for Direct Conversion of Syngas to Light Olefins

Authors: Yuxuan Huang, Weixin Qian, Hongfang Ma, Haitao Zhang, Weiyong Ying

Abstract:

Light olefins are important building blocks for chemical industry. Direct conversion of syngas to light olefins has been investigated for decades. Meanwhile, the limit for light olefins selectivity described by Anderson-Schulz-Flory (ASF) distribution model is still a great challenge to conventional Fischer-Tropsch synthesis. The emerging strategy called oxide-zeolite concept (OX-ZEO) is a promising way to get rid of this limit. ZnCrOx was prepared by co-precipitation method and (NH4)2CO3 was used as precipitant. SAPO-34 was prepared by hydrothermal synthesis, and Tetraethylammonium hydroxide (TEAOH) was used as template, while silica sol, pseudo-boehmite, and phosphoric acid were Al, Si and P source, respectively. The bifunctional catalyst was prepared by mechanical mixing of ZnCrOx and SAPO-34. Catalytic reactions were carried out under H2/CO=2, 380 ℃, 1 MPa and 6000 mL·gcat-1·h-1 in a fixed-bed reactor with a quartz lining. Catalysts were characterized by XRD, N2 adsorption-desorption, NH3-TPD, H2-TPR, and CO-TPD. The addition of Al as structure promoter enhances CO conversion and selectivity to light olefins. Zn/Cr ratio, which decides the active component content and chemisorption property of the catalyst, influences CO conversion and selectivity to light olefins at the same time. C2-4= distribution of 86% among hydrocarbons at CO conversion of 14% was reached when Zn/Cr=1.5.

Keywords: Light olefins, OX-ZEO, syngas, ZnCrOx.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1027
1163 NDENet: End-to-End Nighttime Dehazing and Enhancement

Authors: H. Baskar, A. S. Chakravarthy, P. Garg, D. Goel, A. S. Raj, K. Kumar, Lakshya, R. Parvatham, V. Sushant, B. Kumar Rout

Abstract:

In this paper, we present a computer vision task called nighttime dehaze-enhancement. This task aims to jointly perform dehazing and lightness enhancement. Our task fundamentally differs from nighttime dehazing – our goal is to jointly dehaze and enhance scenes, while nighttime dehazing aims to dehaze scenes under a nighttime setting. In order to facilitate further research on this task, we release a benchmark dataset called Reside-β Night dataset, consisting of 4122 nighttime hazed images from 2061 scenes and 2061 ground truth images. Moreover, we also propose a network called NDENet (Nighttime Dehaze-Enhancement Network), which jointly performs dehazing and low-light enhancement in an end-to-end manner. We evaluate our method on the proposed benchmark and achieve Structural Index Similarity (SSIM) of 0.8962 and Peak Signal to Noise Ratio (PSNR) of 26.25. We also compare our network with other baseline networks on our benchmark to demonstrate the effectiveness of our approach. We believe that nighttime dehaze-enhancement is an essential task particularly for autonomous navigation applications, and hope that our work will open up new frontiers in research. The code for our network is made publicly available.

Keywords: Dehazing, image enhancement, nighttime, computer vision.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 686