Search results for: some guidelines for implementation.
314 Resettlement and Livelihood Sustainability in Sub-Saharan Africa: The Case of Bui Hydro-Power Dam Project, Ghana
Authors: Francis Z. Naab, Abraham M. Nunbogu, Romanus D. Dinye, Alfred Dongzagla
Abstract:
The study assesses the effectiveness of the Bui Dam resettlement scheme in the Tain and the Bole districts in Ghana. The study adopted a mixed approach in its data collection and analyses. Of the eight communities affected by Bui hydropower project, and thus require resettlement, four were purposively selected for primary data collection. Primary data was gathered through questionnaire administration to 157 heads of resettled households, focus group discussions with men and women and in-depth interviews with key informants. The findings indicated that the affected people had been sufficiently contacted at all levels of their resettlement. In particular, the Ghana Dams Dialogue, which served as a liaison entity between the government and the resettlement communities came up for praise for its usefulness. Many tangible policies were put in place to address the socio-cultural differences of traditional authorities. The Bui Dam Authority also rigorously followed national and international laws and protocols in the design and implementation of the resettlement scheme. In assessing the effectiveness of the resettlement scheme, it was clear that there had been a great appreciation of the compensation regarding infrastructural development, but much more would have to be done to satisfy livelihood empowerment requirements. It was recommended that candid efforts be made to restore the lost identities of the communities resettled, and more dialogue is encouraged among communities living together.
Keywords: Resettlement, livelihood, hydro-power project, Bui Dam, Ghana.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1410313 Expert-Driving-Criteria Based on Fuzzy Logic Approach for Intelligent Driving Diagnosis
Authors: Andrés C. Cuervo Pinilla, Christian G. Quintero M., Chinthaka Premachandra
Abstract:
This paper considers people’s driving skills diagnosis under real driving conditions. In that sense, this research presents an approach that uses GPS signals which have a direct correlation with driving maneuvers. Besides, it is presented a novel expert-driving-criteria approximation using fuzzy logic which seeks to analyze GPS signals in order to issue an intelligent driving diagnosis. Based on above, this works presents in the first section the intelligent driving diagnosis system approach in terms of its own characteristics properties, explaining in detail significant considerations about how an expert-driving-criteria approximation must be developed. In the next section, the implementation of our developed system based on the proposed fuzzy logic approach is explained. Here, a proposed set of rules which corresponds to a quantitative abstraction of some traffics laws and driving secure techniques seeking to approach an expert-driving- criteria approximation is presented. Experimental testing has been performed in real driving conditions. The testing results show that the intelligent driving diagnosis system qualifies driver’s performance quantitatively with a high degree of reliability.Keywords: Driver support systems, intelligent transportation systems, fuzzy logic, real time data processing.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1201312 Design and Implementation of a Software Platform Based on Artificial Intelligence for Product Recommendation
Authors: G. Settanni, A. Panarese, R. Vaira, A. Galiano
Abstract:
Nowadays, artificial intelligence is used successfully in the field of e-commerce for its ability to learn from a large amount of data. In this research study, a prototype software platform was designed and implemented in order to suggest to users the most suitable products for their needs. The platform includes a recommender system based on artificial intelligence algorithms that provide suggestions and decision support to the customer. Specifically, support vector machine algorithms have been implemented combined with natural language processing techniques that allow the user to interact with the system, express their requests and receive suggestions. The interested user can access the web platform on the internet using a computer, tablet or mobile phone, register, provide the necessary information and view the products that the system deems them the most appropriate. The platform also integrates a dashboard that allows the use of the various functions, which the platform is equipped with, in an intuitive and simple way. Also, Long Short-Term Memory algorithms have been implemented and trained on historical data in order to predict customer scores of the different items. Items with the highest scores are recommended to customers.
Keywords: Deep Learning, Long Short-Term Memory, Machine Learning, Recommender Systems, Support Vector Machine.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 325311 Concept to Enhance the Project Success and Promote the Implementation of Success Factors in Infrastructure Projects
Abstract:
Infrastructure projects are often subjected to delays and cost overruns and mistakenly described as unsuccessful projects. These projects have many peculiarities such as public attention, impact on the environment, subjected to special regulations, etc. They also deal with several stakeholders with different motivations and face unique risks. With this in mind we need to reconsider our approach to manage them, define their success factors and implement these success factors. Infrastructure projects are not only lacking a unified meaning of project success or a definition of success factors, but also a clear method to implement these factors. This paper investigates this gap and introduces a concept to implement success factors in an efficient way, taking into consideration the specific characteristics of infrastructure projects. This concept consists of six enablers such as project organization, project team, project management workflow, contract management, communication and knowledge transfer and project documentations. These enablers allow other success factors to be efficiently implemented in projects. In conclusion, this paper provides project managers as well as company managers with a tool to define and implement success factors efficiently in their projects, along with upgrading their assets for the coming projects. This tool consists of processes and validated checklists to ensure the best use of company resources and knowledge. Due to the special features of infrastructure projects this tool will be tested in the German infrastructure market. However, it is meant to be adaptable to other markets and industries.
Keywords: Infrastructure projects, enablers, project success, success factors, transportation projects.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 993310 EGCL: An Extended G-Code Language with Flow Control, Functions and Mnemonic Variables
Authors: Oscar E. Ruiz, S. Arroyave, J. F. Cardona
Abstract:
In the context of computer numerical control (CNC) and computer aided manufacturing (CAM), the capabilities of programming languages such as symbolic and intuitive programming, program portability and geometrical portfolio have special importance. They allow to save time and to avoid errors during part programming and permit code re-usage. Our updated literature review indicates that the current state of art presents voids in parametric programming, program portability and programming flexibility. In response to this situation, this article presents a compiler implementation for EGCL (Extended G-code Language), a new, enriched CNC programming language which allows the use of descriptive variable names, geometrical functions and flow-control statements (if-then-else, while). Our compiler produces low-level generic, elementary ISO-compliant Gcode, thus allowing for flexibility in the choice of the executing CNC machine and in portability. Our results show that readable variable names and flow control statements allow a simplified and intuitive part programming and permit re-usage of the programs. Future work includes allowing the programmer to define own functions in terms of EGCL, in contrast to the current status of having them as library built-in functions.
Keywords: CNC Programming, Compiler, G-code Language, Numerically Controlled Machine-Tools.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2622309 Assessment of Energy Demand Considering Different Model Simulations in a Low Energy Demand House
Authors: M. Cañada-Soriano, C. Aparicio-Fernández, P. Sebastián Ferrer Gisbert, M. Val Field, J.-L. Vivancos-Bono
Abstract:
The lack of insulation along with the existence of air leakages constitute a meaningful impact on the energy performance of buildings. Both of them lead to increases in the energy demand through additional heating and/or cooling loads. Additionally, they cause thermal discomfort. In order to quantify these uncontrolled air currents, the Blower Door test can be used. It is a standardized procedure that determines the airtightness of a space by characterizing the rate of air leakages through the envelope surface. In this sense, the low-energy buildings complying with the Passive House design criteria are required to achieve high levels of airtightness. Due to the invisible nature of air leakages, additional tools are often considered to identify where the infiltrations take place such as the infrared thermography. The aim of this study is to assess the airtightness of a typical Mediterranean dwelling house, refurbished under the Passive House standard, using the Blower Door test. Moreover, the building energy performance modelling tools TRNSYS (TRaNsient System Simulation program) and TRNFlow (TRaNsient Flow) have been used to estimate the energy demand in different scenarios. In this sense, a sequential implementation of three different energy improvement measures (insulation thickness, glazing type and infiltrations) have been analyzed.
Keywords: Airtightness, blower door, TRNSYS, infrared thermography, energy demand.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 221308 Assessing the Adaptive Re-Use Potential of Buildings as Part of the Disaster Management Process
Authors: A. Esra İdemen, Sinan M. Şener, Emrah Acar
Abstract:
The technological paradigm of the disaster management field, especially in the case of governmental intervention strategies, is generally based on rapid and flexible accommodation solutions. From various technical solution patterns used to address the immediate housing needs of disaster victims, the adaptive re-use of existing buildings can be considered to be both low-cost and practical. However, there is a scarcity of analytical methods to screen, select and adapt buildings to help decision makers in cases of emergency. Following an extensive literature review, this paper aims to highlight key points and problem areas associated with the adaptive re-use of buildings within the disaster management context. In other disciplines such as real estate management, the adaptive re-use potential (ARP) of existing buildings is typically based on the prioritization of a set of technical and non-technical criteria which are then weighted to arrive at an economically viable investment decision. After a disaster, however, the assessment of the ARP of buildings requires consideration of different/additional layers of analysis which stem from general disaster management principles and the peculiarities of different types of disasters, as well as of their victims. In this paper, a discussion of the development of an adaptive re-use potential (ARP) assessment model is presented. It is thought that governmental and non-governmental decision makers who are required to take quick decisions to accommodate displaced masses following disasters are likely to benefit from the implementation of such a model.Keywords: Adaptive re-use of buildings, assessment model, disaster management, temporary housing.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1687307 An Implementation of Fuzzy Logic Technique for Prediction of the Power Transformer Faults
Authors: Omar M. Elmabrouk., Roaa Y. Taha., Najat M. Ebrahim, Sabbreen A. Mohammed
Abstract:
Power transformers are the most crucial part of power electrical system, distribution and transmission grid. This part is maintained using predictive or condition-based maintenance approach. The diagnosis of power transformer condition is performed based on Dissolved Gas Analysis (DGA). There are five main methods utilized for analyzing these gases. These methods are International Electrotechnical Commission (IEC) gas ratio, Key Gas, Roger gas ratio, Doernenburg, and Duval Triangle. Moreover, due to the importance of the transformers, there is a need for an accurate technique to diagnose and hence predict the transformer condition. The main objective of this technique is to avoid the transformer faults and hence to maintain the power electrical system, distribution and transmission grid. In this paper, the DGA was utilized based on the data collected from the transformer records available in the General Electricity Company of Libya (GECOL) which is located in Benghazi-Libya. The Fuzzy Logic (FL) technique was implemented as a diagnostic approach based on IEC gas ratio method. The FL technique gave better results and approved to be used as an accurate prediction technique for power transformer faults. Also, this technique is approved to be a quite interesting for the readers and the concern researchers in the area of FL mathematics and power transformer.
Keywords: Fuzzy logic, dissolved gas-in-oil analysis, DGA, prediction, power transformer.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1357306 ANN Based Currency Recognition System using Compressed Gray Scale and Application for Sri Lankan Currency Notes - SLCRec
Authors: D. A. K. S. Gunaratna, N. D. Kodikara, H. L. Premaratne
Abstract:
Automatic currency note recognition invariably depends on the currency note characteristics of a particular country and the extraction of features directly affects the recognition ability. Sri Lanka has not been involved in any kind of research or implementation of this kind. The proposed system “SLCRec" comes up with a solution focusing on minimizing false rejection of notes. Sri Lankan currency notes undergo severe changes in image quality in usage. Hence a special linear transformation function is adapted to wipe out noise patterns from backgrounds without affecting the notes- characteristic images and re-appear images of interest. The transformation maps the original gray scale range into a smaller range of 0 to 125. Applying Edge detection after the transformation provided better robustness for noise and fair representation of edges for new and old damaged notes. A three layer back propagation neural network is presented with the number of edges detected in row order of the notes and classification is accepted in four classes of interest which are 100, 500, 1000 and 2000 rupee notes. The experiments showed good classification results and proved that the proposed methodology has the capability of separating classes properly in varying image conditions.Keywords: Artificial intelligence, linear transformation and pattern recognition.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2833305 Performance Analysis of Traffic Classification with Machine Learning
Authors: Htay Htay Yi, Zin May Aye
Abstract:
Network security is role of the ICT environment because malicious users are continually growing that realm of education, business, and then related with ICT. The network security contravention is typically described and examined centrally based on a security event management system. The firewalls, Intrusion Detection System (IDS), and Intrusion Prevention System are becoming essential to monitor or prevent of potential violations, incidents attack, and imminent threats. In this system, the firewall rules are set only for where the system policies are needed. Dataset deployed in this system are derived from the testbed environment. The traffic as in DoS and PortScan traffics are applied in the testbed with firewall and IDS implementation. The network traffics are classified as normal or attacks in the existing testbed environment based on six machine learning classification methods applied in the system. It is required to be tested to get datasets and applied for DoS and PortScan. The dataset is based on CICIDS2017 and some features have been added. This system tested 26 features from the applied dataset. The system is to reduce false positive rates and to improve accuracy in the implemented testbed design. The system also proves good performance by selecting important features and comparing existing a dataset by machine learning classifiers.Keywords: False negative rate, intrusion detection system, machine learning methods, performance.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1070304 Performance of an Improved Fluidized System for Processing Green Tea
Authors: Nickson Kipng’etich Lang’at, Thomas Thoruwa, John Abraham, John Wanyoko
Abstract:
Green tea is made from the top two leaves and buds of a shrub, Camellia sinensis, of the family Theaceae and the order Theales. The green tea leaves are picked and immediately sent to be dried or steamed to prevent fermentation. Fluid bed drying technique is a common drying method used in drying green tea because of its ease in design and construction and fluidization of fine tea particles. Major problems in this method are significant loss of chemical content of the leaf and green appearance of tea, retention of high moisture content in the leaves and bed channeling and defluidization. The energy associated with the drying technology has been shown to be a vital factor in determining the quality of green tea. As part of the implementation, prototype dryer was built that facilitated sequence of operations involving steaming, cooling, pre-drying and final drying. The major findings of the project were in terms of quality characteristics of tea leaves and energy consumption during processing. The optimal design achieved a moisture content of 4.2 ± 0.84%. With the optimum drying temperature of 100 ºC, the specific energy consumption was 1697.8 kj.Kg-1 and evaporation rate of 4.272 x 10-4 Kg.m-2.s-1. The energy consumption in a fluidized system can be further reduced by focusing on energy saving designs.
Keywords: Evaporation rate, fluid bed dryer, maceration, specific energy consumption.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1700303 MARTI and MRSD: Newly Developed Isolation-Damping Devices with Adaptive Hardening for Seismic Protection of Structures
Authors: Murat Dicleli, Ali Salem Milani
Abstract:
In this paper, a summary of analytical and experimental studies into the behavior of a new hysteretic damper, designed for seismic protection of structures is presented. The Multidirectional Torsional Hysteretic Damper (MRSD) is a patented invention in which a symmetrical arrangement of identical cylindrical steel cores is so configured as to yield in torsion while the structure experiences planar movements due to earthquake shakings. The new device has certain desirable properties. Notably, it is characterized by a variable and controllable-via-design post-elastic stiffness. The mentioned property is a result of MRSD’s kinematic configuration which produces this geometric hardening, rather than being a secondary large-displacement effect. Additionally, the new system is capable of reaching high force and displacement capacities, shows high levels of damping, and very stable cyclic response. The device has gone through many stages of design refinement, multiple prototype verification tests and development of design guide-lines and computer codes to facilitate its implementation in practice. Practicality of the new device, as offspring of an academic sphere, is assured through extensive collaboration with industry in its final design stages, prototyping and verification test programs.Keywords: Seismic, isolation, damper, adaptive stiffness.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1998302 Performance Evaluation of Neural Network Prediction for Data Prefetching in Embedded Applications
Authors: Sofien Chtourou, Mohamed Chtourou, Omar Hammami
Abstract:
Embedded systems need to respect stringent real time constraints. Various hardware components included in such systems such as cache memories exhibit variability and therefore affect execution time. Indeed, a cache memory access from an embedded microprocessor might result in a cache hit where the data is available or a cache miss and the data need to be fetched with an additional delay from an external memory. It is therefore highly desirable to predict future memory accesses during execution in order to appropriately prefetch data without incurring delays. In this paper, we evaluate the potential of several artificial neural networks for the prediction of instruction memory addresses. Neural network have the potential to tackle the nonlinear behavior observed in memory accesses during program execution and their demonstrated numerous hardware implementation emphasize this choice over traditional forecasting techniques for their inclusion in embedded systems. However, embedded applications execute millions of instructions and therefore millions of addresses to be predicted. This very challenging problem of neural network based prediction of large time series is approached in this paper by evaluating various neural network architectures based on the recurrent neural network paradigm with pre-processing based on the Self Organizing Map (SOM) classification technique.Keywords: Address, data set, memory, prediction, recurrentneural network.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1675301 System for Monitoring Marine Turtles Using Unstructured Supplementary Service Data
Authors: Luís Pina
Abstract:
The conservation of marine biodiversity keeps ecosystems in balance and ensures the sustainable use of resources. In this context, technological resources have been used for monitoring marine species to allow biologists to obtain data in real-time. There are different mobile applications developed for data collection for monitoring purposes, but these systems are designed to be utilized only on third-generation (3G) phones or smartphones with Internet access and in rural parts of the developing countries, Internet services and smartphones are scarce. Thus, the objective of this work is to develop a system to monitor marine turtles using Unstructured Supplementary Service Data (USSD), which users can access through basic mobile phones. The system aims to improve the data collection mechanism and enhance the effectiveness of current systems in monitoring sea turtles using any type of mobile device without Internet access. The system will be able to report information related to the biological activities of marine turtles. Also, it will be used as a platform to assist marine conservation entities to receive reports of illegal sales of sea turtles. The system can also be utilized as an educational tool for communities, providing knowledge and allowing the inclusion of communities in the process of monitoring marine turtles. Therefore, this work may contribute with information to decision-making and implementation of contingency plans for marine conservation programs.
Keywords: GSM, marine biology, marine turtles, USSD.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 930300 Comparison between Separable and Irreducible Goppa Code in McEliece Cryptosystem
Authors: Thuraya M. Qaradaghi, Newroz N. Abdulrazaq
Abstract:
The McEliece cryptosystem is an asymmetric type of cryptography based on error correction code. The classical McEliece used irreducible binary Goppa code which considered unbreakable until now especially with parameter [1024, 524, and 101], but it is suffering from large public key matrix which leads to be difficult to be used practically. In this work Irreducible and Separable Goppa codes have been introduced. The Irreducible and Separable Goppa codes used are with flexible parameters and dynamic error vectors. A Comparison between Separable and Irreducible Goppa code in McEliece Cryptosystem has been done. For encryption stage, to get better result for comparison, two types of testing have been chosen; in the first one the random message is constant while the parameters of Goppa code have been changed. But for the second test, the parameters of Goppa code are constant (m=8 and t=10) while the random message have been changed. The results show that the time needed to calculate parity check matrix in separable are higher than the one for irreducible McEliece cryptosystem, which is considered expected results due to calculate extra parity check matrix in decryption process for g2(z) in separable type, and the time needed to execute error locator in decryption stage in separable type is better than the time needed to calculate it in irreducible type. The proposed implementation has been done by Visual studio C#.Keywords: McEliece cryptosystem, Goppa code, separable, irreducible.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2211299 Implementation of a Motion Detection System
Authors: Asif Ansari, T.C.Manjunath, C. Ardil
Abstract:
In today-s competitive environment, the security concerns have grown tremendously. In the modern world, possession is known to be 9/10-ths of the law. Hence, it is imperative for one to be able to safeguard one-s property from worldly harms such as thefts, destruction of property, people with malicious intent etc. Due to the advent of technology in the modern world, the methodologies used by thieves and robbers for stealing have been improving exponentially. Therefore, it is necessary for the surveillance techniques to also improve with the changing world. With the improvement in mass media and various forms of communication, it is now possible to monitor and control the environment to the advantage of the owners of the property. The latest technologies used in the fight against thefts and destruction are the video surveillance and monitoring. By using the technologies, it is possible to monitor and capture every inch and second of the area in interest. However, so far the technologies used are passive in nature, i.e., the monitoring systems only help in detecting the crime but do not actively participate in stopping or curbing the crime while it takes place. Therefore, we have developed a methodology to detect the motion in a video stream environment and this is an idea to ensure that the monitoring systems not only actively participate in stopping the crime, but do so while the crime is taking place. Hence, a system is used to detect any motion in a live streaming video and once motion has been detected in the live stream, the software will activate a warning system and capture the live streaming video.
Keywords: Motion, Detection, System, Video, Crime, Matlab, Surveillance.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 4290298 The Effection of Different Culturing Proportion of Deep Sea Water(DSW) to Surface Sea Water(SSW) in Reductive Ability and Phenolic Compositions of Sargassum Cristaefolium
Authors: H. L. Ku, K. C. Yang, S. Y. Jhou, S. C. Lee, C. S. Lin
Abstract:
Characterized as rich mineral substances, low temperature, few bacteria, and stability with numerous implementation aspects on aquaculture, food, drinking, and leisure, the deep sea water (DSW) development has become a new industry in the world. It has been report that marine algae contain various biologically active compounds. This research focued on the affections in cultivating Sagrassum cristaefolium with different concentration of deep sea water(DSW) and surface sea water(SSW). After two and four weeks, the total phenolic contents were compared in Sagrassum cristaefolium culturing with different ways, and the reductive activity of them was also be tried with potassium ferricyanide. Those fresh seaweeds were dried with oven and were ground to powder. Progressively, the marine algae we cultured was extracted by water under the condition with heating them at 90Ôäâ for 1hr.The total phenolic contents were be executed using Folin–Ciocalteu method. The results were explaining as follows: the highest total phenolic contents and the best reductive ability of all could be observed on the 1/4 proportion of DSW to SSW culturing in two weeks. Furthermore, the 1/2 proportion of DSW to SSW also showed good reductive ability and plentiful phenolic compositions. Finally, we confirmed that difference proportion of DSW and SSW is the major point relating to ether the total phenolic components or the reductive ability in the Sagrassum cristaefolium. In the future, we will use this way to mass production the marine algae or other micro algae on industry applications.Keywords: deep sea water(DSW), surface sea water(SSW), phenolic contents, reductive ability.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1619297 A Novel VLSI Architecture of Hybrid Image Compression Model based on Reversible Blockade Transform
Authors: C. Hemasundara Rao, M. Madhavi Latha
Abstract:
Image compression can improve the performance of the digital systems by reducing time and cost in image storage and transmission without significant reduction of the image quality. Furthermore, the discrete cosine transform has emerged as the new state-of-the art standard for image compression. In this paper, a hybrid image compression technique based on reversible blockade transform coding is proposed. The technique, implemented over regions of interest (ROIs), is based on selection of the coefficients that belong to different transforms, depending on the coefficients is proposed. This method allows: (1) codification of multiple kernals at various degrees of interest, (2) arbitrary shaped spectrum,and (3) flexible adjustment of the compression quality of the image and the background. No standard modification for JPEG2000 decoder was required. The method was applied over different types of images. Results show a better performance for the selected regions, when image coding methods were employed for the whole set of images. We believe that this method is an excellent tool for future image compression research, mainly on images where image coding can be of interest, such as the medical imaging modalities and several multimedia applications. Finally VLSI implementation of proposed method is shown. It is also shown that the kernal of Hartley and Cosine transform gives the better performance than any other model.Keywords: VLSI, Discrete Cosine Transform, JPEG, Hartleytransform, Radon Transform
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1836296 Experimental Study for the Development of a Wireless Communication System in a Solar Central Tower Facility
Authors: Victor H. Benitez, Ramon V. Armas-Flores, Jesus H. Pacheco-Ramirez
Abstract:
Systems transforming solar energy into electrical power have emerged as a viable source of clean, renewable energy. Solar power tower technology is a good example of this type of system, which consists of several mobile mirrors, called heliostats, which reflect the sun's radiation to the same point, located on top of a tower at the center of heliostat field, for collection or transformation into another type of energy. The so-called Hermosillo’s Solar Platform (Plataforma Solar de Hermosillo, PSH, in Spanish) is a facility constituted with several heliostats, its aim and scope is for research purposes. In this paper, the implementation of a wireless communication system based on intelligent nodes is proposed in order to allow the communication and control of the heliostats in PSH. Intelligent nodes transmit information from one point to another, and can perform other actions that allow them to adapt to the conditions and limitations of a field of heliostats, thus achieving effective communication system. After deployment of the nodes in the heliostats, tests were conducted to measure the effectiveness of the communication, and determine the feasibility of using the proposed technologies. The test results were always positive, exceeding expectations held for its operation in the field of heliostats. Therefore, it was possible to validate the efficiency of the wireless communication system to be implemented in PSH, allowing communication and control of the heliostats.Keywords: Solar energy, heliostat, wireless communication, intelligent node.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1667295 Prioritising the TQM Enablers and IT Resources in the ICT Industry: An AHP Approach
Authors: Suby Khanam, Jamshed Siddiqui, Faisal Talib
Abstract:
Total Quality Management (TQM) is a managerial approach that improves the competitiveness of the industry, meanwhile Information technology (IT) was introduced with TQM for handling the technical issues which is supported by quality experts for fulfilling the customers’ requirement. Present paper aims to utilise AHP (Analytic Hierarchy Process) methodology to priorities and rank the hierarchy levels of TQM enablers and IT resource together for its successful implementation in the Information and Communication Technology (ICT) industry. A total of 17 TQM enablers (nine) and IT resources (eight) were identified and partitioned into 3 categories and were prioritised by AHP approach. The finding indicates that the 17 sub-criteria can be grouped into three main categories namely organizing, tools and techniques, and culture and people. Further, out of 17 sub-criteria, three sub-criteria: top management commitment and support, total employee involvement, and continuous improvement got highest priority whereas three sub-criteria such as structural equation modelling, culture change, and customer satisfaction got lowest priority. The result suggests a hierarchy model for ICT industry to prioritise the enablers and resources as well as to improve the TQM and IT performance in the ICT industry. This paper has some managerial implication which suggests the managers of ICT industry to implement TQM and IT together in their organizations to get maximum benefits and how to utilize available resources. At the end, conclusions, limitation, future scope of the study are presented.Keywords: Analytic Hierarchy Process, Information Technology, Information and Communication Technology, Prioritization, Total Quality Management.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1956294 Teaching Math to Preschool Children with Autism
Authors: Hui Fang Huang Su, Jia Borror
Abstract:
This study compared two different interventions for math instruction among preschoolers with autism spectrum disorder (ASD). The first intervention, a combination of discrete trial teaching and Strategies for Teaching Based on Autism Research (STAR), was the regular math curriculum utilized at the preschool. The second activity-based, naturalistic intervention was Project Mind, also known as Math is Not Difficult. The curricular interventions were randomly assigned to four preschool classrooms with ASD students and implemented over three months for Project MIND. Measurements gained during the same three months for the STAR intervention were used. A quasi-experimental, pre-test/post-test design was selected to compare which intervention was the most effective in increasing mathematical knowledge and skills among preschoolers with ASD. Standardized pre and post-test instruments included the Bracken Basic Concept Scale-3 Receptive, the Applied Problems and Calculation subtests of the Woodcock-Johnson IV Tests of Achievement, and the TEMA 3: Test of Early Mathematics Ability – Third Edition. The STAR assessment is typically administered to all preschoolers at the study site three times per year, and those results were used in this study. We anticipated that the implementation of these two approaches would lead to improvement in the mathematical knowledge and skills of children with ASD. Still, it is essential to see whether a behavioral or naturalistic teaching approach leads to more significant results.
Keywords: Autism, mathematics, preschool, special education.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 886293 Evolutionary Techniques for Model Order Reduction of Large Scale Linear Systems
Authors: S. Panda, J. S. Yadav, N. P. Patidar, C. Ardil
Abstract:
Recently, genetic algorithms (GA) and particle swarm optimization (PSO) technique have attracted considerable attention among various modern heuristic optimization techniques. The GA has been popular in academia and the industry mainly because of its intuitiveness, ease of implementation, and the ability to effectively solve highly non-linear, mixed integer optimization problems that are typical of complex engineering systems. PSO technique is a relatively recent heuristic search method whose mechanics are inspired by the swarming or collaborative behavior of biological populations. In this paper both PSO and GA optimization are employed for finding stable reduced order models of single-input- single-output large-scale linear systems. Both the techniques guarantee stability of reduced order model if the original high order model is stable. PSO method is based on the minimization of the Integral Squared Error (ISE) between the transient responses of original higher order model and the reduced order model pertaining to a unit step input. Both the methods are illustrated through numerical example from literature and the results are compared with recently published conventional model reduction technique.
Keywords: Genetic Algorithm, Particle Swarm Optimization, Order Reduction, Stability, Transfer Function, Integral Squared Error.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2722292 Design and Implementation of a Counting and Differentiation System for Vehicles through Video Processing
Authors: Derlis Gregor, Kevin Cikel, Mario Arzamendia, Raúl Gregor
Abstract:
This paper presents a self-sustaining mobile system for counting and classification of vehicles through processing video. It proposes a counting and classification algorithm divided in four steps that can be executed multiple times in parallel in a SBC (Single Board Computer), like the Raspberry Pi 2, in such a way that it can be implemented in real time. The first step of the proposed algorithm limits the zone of the image that it will be processed. The second step performs the detection of the mobile objects using a BGS (Background Subtraction) algorithm based on the GMM (Gaussian Mixture Model), as well as a shadow removal algorithm using physical-based features, followed by morphological operations. In the first step the vehicle detection will be performed by using edge detection algorithms and the vehicle following through Kalman filters. The last step of the proposed algorithm registers the vehicle passing and performs their classification according to their areas. An auto-sustainable system is proposed, powered by batteries and photovoltaic solar panels, and the data transmission is done through GPRS (General Packet Radio Service)eliminating the need of using external cable, which will facilitate it deployment and translation to any location where it could operate. The self-sustaining trailer will allow the counting and classification of vehicles in specific zones with difficult access.Keywords: Intelligent transportation systems, object detection, video processing, road traffic, vehicle counting, vehicle classification.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1624291 Requirements Driven Multiple View Paradigm for Developing Security Architecture
Authors: K. Chandra Sekaran
Abstract:
This paper describes a paradigmatic approach to develop architecture of secure systems by describing the requirements from four different points of view: that of the owner, the administrator, the user, and the network. Deriving requirements and developing architecture implies the joint elicitation and describing the problem and the structure of the solution. The view points proposed in this paper are those we consider as requirements towards their contributions as major parties in the design, implementation, usage and maintenance of secure systems. The dramatic growth of the technology of Internet and the applications deployed in World Wide Web have lead to the situation where the security has become a very important concern in the development of secure systems. Many security approaches are currently being used in organizations. In spite of the widespread use of many different security solutions, the security remains a problem. It is argued that the approach that is described in this paper for the development of secure architecture is practical by all means. The models representing these multiple points of view are termed the requirements model (views of owner and administrator) and the operations model (views of user and network). In this paper, this multiple view paradigm is explained by first describing the specific requirements and or characteristics of secure systems (particularly in the domain of networks) and the secure architecture / system development methodology.
Keywords: Multiple view paradigms, requirements model, operations model, secure system, owner, administrator, user, network.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1370290 Development of Reliable Web-Based Laboratories for Developing Countries
Authors: Teyana S. Sapula, Damian D. Haule
Abstract:
In online context, the design and implementation of effective remote laboratories environment is highly challenging on account of hardware and software needs. This paper presents the remote laboratory software framework modified from ilab shared architecture (ISA). The ISA is a framework which enables students to remotely acccess and control experimental hardware using internet infrastructure. The need for remote laboratories came after experiencing problems imposed by traditional laboratories. Among them are: the high cost of laboratory equipment, scarcity of space, scarcity of technical personnel along with the restricted university budget creates a significant bottleneck on building required laboratory experiments. The solution to these problems is to build web-accessible laboratories. Remote laboratories allow students and educators to interact with real laboratory equipment located anywhere in the world at anytime. Recently, many universities and other educational institutions especially in third world countries rely on simulations because they do not afford the experimental equipment they require to their students. Remote laboratories enable users to get real data from real-time hand-on experiments. To implement many remote laboratories, the system architecture should be flexible, understandable and easy to implement, so that different laboratories with different hardware can be deployed easily. The modifications were made to enable developers to add more equipment in ISA framework and to attract the new developers to develop many online laboratories.Keywords: Batched, ISA, labserver, servicebroker.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1429289 Moving from Rule-based to Principle-based in Public Sector: Preparers' Perspective
Authors: Roshayani Arshad, Normah Omar, Siti Fatimah Awang
Abstract:
The move from cash accounting to accrual accounting, or rule-based to principle-based accounting, by many governments is part of an ongoing efforts in promoting a more business-like and performance-focused public sector. Using questionnaire responses from preparers of financial statements of public universities in Malaysia, this study examines the implementation challenges and benefits of principle-based accounting. Results from these responses suggest that most respondents perceived significant costs would be incurred in relation to staff training and recruitment of staffs with relevant technical knowledge. In addition, most respondents also perceived that there will be significant changes in the current accounting system and structure in order to comply with the principle-based accounting requirements. However, most respondents perceived that these changes might not result in significant benefits for management purposes, for example, financial management, budgeting and allocation of resources. Nevertheless, most respondents perceived that principle-based accounting information would facilitate the monitoring function of the board. The general perception is that adoption of principle-based accounting information is not significantly useful than rule-based accounting information is expected to change over time as preparers of the financial statements gradually understand and appreciate the benefits of principle-based accounting information. This infers that the perceived usefulness of different accounting system is a function of familiarity by the preparers.
Keywords: Accrual accounting, principle-based accounting, public sector, rule-based accounting.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2956288 A Generic Middleware to Instantly Sync Intensive Writes of Heterogeneous Massive Data via Internet
Authors: Haitao Yang, Zhenjiang Ruan, Fei Xu, Lanting Xia
Abstract:
Industry data centers often need to sync data changes reliably and instantly from a large-scale of heterogeneous autonomous relational databases accessed via the not-so-reliable Internet, for which a practical generic sync middleware of low maintenance and operation costs is most wanted. To this demand, this paper presented a generic sync middleware system (GSMS), which has been developed, applied and optimized since 2006, holding the principles or advantages that it must be SyncML-compliant and transparent to data application layer logic without referring to implementation details of databases synced, does not rely on host computer operating systems deployed, and its construction is light weighted and hence of low cost. Regarding these hard commitments of developing GSMS, in this paper we stressed the significant optimization breakthrough of GSMS sync delay being well below a fraction of millisecond per record sync. A series of ultimate tests with GSMS sync performance were conducted for a persuasive example, in which the source relational database underwent a broad range of write loads (from one thousand to one million intensive writes within a few minutes). All these tests showed that the performance of GSMS is competent and smooth even under ultimate write loads.
Keywords: Heterogeneous massive data, instantly sync intensive writes, Internet generic middleware design, optimization.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 454287 Method for Tuning Level Control Loops Based on Internal Model Control and Closed Loop Step Test Data
Authors: Arnaud Nougues
Abstract:
This paper describes a two-stage methodology derived from IMC (Internal Model Control) for tuning a PID (Proportional-Integral-Derivative) controller for levels or other integrating processes in an industrial environment. Focus is ease of use and implementation speed which are critical for an industrial application. Tuning can be done with minimum effort and without the need of time-consuming open-loop step tests on the plant. The first stage of the method applies to levels only: the vessel residence time is calculated from equipment dimensions and used to derive a set of preliminary PI (Proportional-Integral) settings with IMC. The second stage, re-tuning in closed-loop, applies to levels as well as other integrating processes: a tuning correction mechanism has been developed based on a series of closed-loop simulations with model errors. The tuning correction is done from a simple closed-loop step test and application of a generic correlation between observed overshoot and integral time correction. A spin-off of the method is that an estimate of the vessel residence time (levels) or open-loop process gain (other integrating process) is obtained from the closed-loop data.
Keywords: closed-loop model identification, IMC-PID tuning method, integrating process control, on-line PID tuning adaptation
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 577286 Children’s Literature in Primary School: An Opportunity to Develop Soft Skills
Abstract:
Emotions are manifestations of everything that happens around us, influencing, consequently, our actions. People experience emotions continuously when socialize with friends, when facing complex situations, and when at school, among many other situations. Although the influence of emotions in the teaching and learning process is nothing new, its study in the academic field has been more popular in recent years, distinguishing between positive (e.g., enjoyment and curiosity) and negative emotions (e.g., boredom and frustration). There is no doubt that emotions play an important role in the students’ learning process since the development of knowledge involves thoughts, actions, and emotions. Nowadays, one of the most significant changes in acquiring knowledge, accessing information, and communicating is the way we do it through technological and digital resources. Faced with an increasingly frequent use of technological or digital means with different purposes, whether in the acquisition of knowledge or in communicating with others, the emotions involved in these processes change naturally. The speed with which the Internet provides information reduces the excitement for searching for the answer, the gratification of discovering something through our own effort, the patience, the capacity for effort, and resilience. Thus, technological and digital devices are bringing changes to the emotional domain. For this reason and others, it is essential to educate children from an early age to understand that it is not possible to have everything with just one click and to deal with negative emotions. Currently, many curriculum guidelines highlight the importance of the development of so-called soft skills, in which the emotional domain is present, in academic contexts. Within the scope of the Portuguese reality, the “Students’ profile by the end of compulsory schooling” and the “Health education reference” also emphasize the importance of emotions in education. There are several resources to stimulate good emotions in articulation with cognitive development. One of the most predictable and not very used resources in the most diverse areas of knowledge after pre-school education is the literature. Due to its characteristics, in the narrative or in the illustrations, literature provides the reader with a journey full of emotions. On the other hand, literature makes it possible to establish bridges between narrative and different areas of knowledge, reconciling the cognitive and emotional domains. This study results from the presentation session of a children's book, entitled “From the Outside to Inside and from the Inside to Outside”, to children attending the 2nd, 3rd, and 4th years of basic education in the Portuguese education system. In this book, rationale and emotion are in constant dialogue, so in this session, based on excerpts from the book dramatized by the authors, some questions were asked to the children in a large group, with an aim to explore their perception regarding certain emotions or events that trigger them. According to the aim of this study, qualitative, descriptive, and interpretative research was carried out based on participant observation and audio records.
Keywords: Emotions, children’s literature, basic education, soft skills.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 350285 Applying Element Free Galerkin Method on Beam and Plate
Authors: Mahdad M’hamed, Belaidi Idir
Abstract:
This paper develops a meshless approach, called Element Free Galerkin (EFG) method, which is based on the weak form Moving Least Squares (MLS) of the partial differential governing equations and employs the interpolation to construct the meshless shape functions. The variation weak form is used in the EFG where the trial and test functions are approximated bye the MLS approximation. Since the shape functions constructed by this discretization have the weight function property based on the randomly distributed points, the essential boundary conditions can be implemented easily. The local weak form of the partial differential governing equations is obtained by the weighted residual method within the simple local quadrature domain. The spline function with high continuity is used as the weight function. The presently developed EFG method is a truly meshless method, as it does not require the mesh, either for the construction of the shape functions, or for the integration of the local weak form. Several numerical examples of two-dimensional static structural analysis are presented to illustrate the performance of the present EFG method. They show that the EFG method is highly efficient for the implementation and highly accurate for the computation. The present method is used to analyze the static deflection of beams and plate holeKeywords: Numerical computation, element-free Galerkin, moving least squares, meshless methods.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2436