Search results for: Quantum Computer Simulation
196 Availability, Accessibility and Utilization of Information and Communication Technology in Teaching and Learning Islamic Studies in Colleges of Education, North-Eastern, Nigeria
Authors: Bello Ali
Abstract:
The use of Information and Communication Technology (ICT) in tertiary institutions by lecturers and students has become a necessity for the enhancement of quality teaching and learning. This study examined availability, accessibility and utilization of ICT in Teaching-Learning Islamic Studies in Colleges of Education, North-East, Nigeria. The study adopted multi-stage sampling technique, in which, five out of the eleven Colleges of Education (both Federal and State owned) were purposively selected for the study. Primary data was drawn from the respondents by the use of questionnaire, interviews and observations. The results of the study, generally, indicate that the availability and accessibility to ICT facilities in Colleges of Education in North-East, Nigeria, especially in teaching/learning delivery of Islamic studies were relatively inadequate and rare to lecturers and students. The study further reveals that the respondents’ level of utilization of ICT is low and only few computer packages and internet services were involved in the ICT utilization, which is yet to reach the real expected situation of the globalization and advancement in the application of ICT if compared to other parts of the world, as far as the teaching and learning of Islamic studies is concerned. Observations and conclusion were drawn from the findings and finally, recommendations on how to improve on ICT availability, accessibility and utilization in teaching/ learning were suggested.
Keywords: Accessibility, availability, college of education, ICT, Islamic Studies, learning, North-Eastern, teaching, utilization.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1132195 Solar Energy Collection using a Double-layer Roof
Authors: S. Kong Wang
Abstract:
The purpose of this study is to investigate the efficiency of a double-layer roof in collecting solar energy as an application to the areas such as raising high-end temperature of organic Rankine cycle (ORC). The by-product of the solar roof is to reduce building air-conditioning loads. The experimental apparatus are arranged to evaluate the effects of the solar roof in absorbing solar energy. The flow channel is basically formed by an aluminum plate on top of a plywood plate. The geometric configurations in which the effects of absorbing energy is analyzed include: a bare uncovered aluminum plate, a glass-covered aluminum plate, a glass-covered/black-painted aluminum plate, a plate with variable lengths, a flow channel with stuffed material (in an attempt on enhancement of heat conduction), and a flow channel with variable slanted angles. The experimental results show that the efficiency of energy collection varies from 0.6 % to 11 % for the geometric configurations mentioned above. An additional study is carried out using CFD simulation to investigate the effects of fins on the aluminum plate. It shows that due to vastly enhanced heat conduction, the efficiency can reach ~23 % if 50 fins are installed on the aluminum plate. The study shows that a double-layer roof can efficiently absorb solar energy and substantially reduce building air-conditioning loads. On the high end of an organic Rankine cycle, a solar pond is used to replace the warm surface water of the sea as OTEC (ocean thermal energy conversion) is the driving energy for the ORC. The energy collected from the double-layered solar roof can be pumped into the pond and raise the pond temperature as the pond surface area is equivalently increased by nearly one-fourth of the total area of the double-layer solar roof. The effect of raising solar pond temperature is especially prominent if the double-layer solar roofs are installed in a community area.Keywords: solar energy collection, double-layer solar roof, energy conservation, ORC, OTEC
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2334194 Efficient Program Slicing Algorithms for Measuring Functional Cohesion and Parallelism
Authors: Jehad Al Dallal
Abstract:
Program slicing is the task of finding all statements in a program that directly or indirectly influence the value of a variable occurrence. The set of statements that can affect the value of a variable at some point in a program is called a program slice. In several software engineering applications, such as program debugging and measuring program cohesion and parallelism, several slices are computed at different program points. In this paper, algorithms are introduced to compute all backward and forward static slices of a computer program by traversing the program representation graph once. The program representation graph used in this paper is called Program Dependence Graph (PDG). We have conducted an experimental comparison study using 25 software modules to show the effectiveness of the introduced algorithm for computing all backward static slices over single-point slicing approaches in computing the parallelism and functional cohesion of program modules. The effectiveness of the algorithm is measured in terms of time execution and number of traversed PDG edges. The comparison study results indicate that using the introduced algorithm considerably saves the slicing time and effort required to measure module parallelism and functional cohesion.
Keywords: Backward slicing, cohesion measure, forward slicing, parallelism measure, program dependence graph, program slicing, static slicing.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1448193 Vision-Based Collision Avoidance for Unmanned Aerial Vehicles by Recurrent Neural Networks
Authors: Yao-Hong Tsai
Abstract:
Due to the sensor technology, video surveillance has become the main way for security control in every big city in the world. Surveillance is usually used by governments for intelligence gathering, the prevention of crime, the protection of a process, person, group or object, or the investigation of crime. Many surveillance systems based on computer vision technology have been developed in recent years. Moving target tracking is the most common task for Unmanned Aerial Vehicle (UAV) to find and track objects of interest in mobile aerial surveillance for civilian applications. The paper is focused on vision-based collision avoidance for UAVs by recurrent neural networks. First, images from cameras on UAV were fused based on deep convolutional neural network. Then, a recurrent neural network was constructed to obtain high-level image features for object tracking and extracting low-level image features for noise reducing. The system distributed the calculation of the whole system to local and cloud platform to efficiently perform object detection, tracking and collision avoidance based on multiple UAVs. The experiments on several challenging datasets showed that the proposed algorithm outperforms the state-of-the-art methods.Keywords: Unmanned aerial vehicle, object tracking, deep learning, collision avoidance.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 953192 Demonstration of Land Use Changes Simulation Using Urban Climate Model
Authors: Barbara Vojvodikova, Katerina Jupova, Iva Ticha
Abstract:
Cities in their historical evolution have always adapted their internal structure to the needs of society (for example protective city walls during classicism era lost their defense function, became unnecessary, were demolished and gave space for new features such as roads, museums or parks). Today it is necessary to modify the internal structure of the city in order to minimize the impact of climate changes on the environment of the population. This article discusses the results of the Urban Climate model owned by VITO, which was carried out as part of a project from the European Union's Horizon grant agreement No 730004 Pan-European Urban Climate Services Climate-Fit city. The use of the model was aimed at changes in land use and land cover in cities related to urban heat islands (UHI). The task of the application was to evaluate possible land use change scenarios in connection with city requirements and ideas. Two pilot areas in the Czech Republic were selected. One is Ostrava and the other Hodonín. The paper provides a demonstration of the application of the model for various possible future development scenarios. It contains an assessment of the suitability or inappropriateness of scenarios of future development depending on the temperature increase. Cities that are preparing to reconstruct the public space are interested in eliminating proposals that would lead to an increase in temperature stress as early as in the assignment phase. If they have evaluation on the unsuitability of some type of design, they can limit it into the proposal phases. Therefore, especially in the application of models on Local level - in 1 m spatial resolution, it was necessary to show which type of proposals would create a significant temperature island in its implementation. Such a type of proposal is considered unsuitable. The model shows that the building itself can create a shady place and thus contribute to the reduction of the UHI. If it sensitively approaches the protection of existing greenery, this new construction may not pose a significant problem. More massive interventions leading to the reduction of existing greenery create a new heat island space.
Keywords: Heat islands, land use, urban climate model.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 839191 Trapping Efficiency of Diesel Particles Through a Square Duct
Authors: Francis William S, Imtiaz Ahmed Choudhury, Ananda Kumar Eriki, A. John Rajan
Abstract:
Diesel Engines emit complex mixtures of inorganic and organic compounds in the form of both solid and vapour phase particles. Most of the particulates released are ultrafine nanoparticles which are detrimental to human health and can easily enter the body by respiration. The emissions standards on particulate matter release from diesel engines are constantly upgraded within the European Union and with future regulations based on the particles numbers released instead of merely mass, the need for effective aftertreatment devices will increase. Standard particulate filters in the form of wall flow filters can have problems with high soot accumulation, producing a large exhaust backpressure. A potential solution would be to combine the standard filter with a flow through filter to reduce the load on the wall flow filter. In this paper soot particle trapping has been simulated in different continuous flow filters of monolithic structure including the use of promoters, at laminar flow conditions. An Euler Lagrange model, the discrete phase model in Ansys used with user defined functions for forces acting on particles. A method to quickly screen trapping of 5 nm and 10 nm particles in different catalysts designs with tracers was also developed. Simulations of square duct monoliths with promoters show that the strength of the vortices produced are not enough to give a high amount of particle deposition on the catalyst walls. The smallest particles in the simulations, 5 and 10 nm particles were trapped to a higher extent, than larger particles up to 1000 nm, in all studied geometries with the predominant deposition mechanism being Brownian diffusion. The comparison of the different filters designed with a wall flow filter does show that the options for altering a design of a flow through filter, without imposing a too large pressure drop penalty are good.Keywords: Diesel Engine trap, thermophoresis, Exhaust pipe, PM-Simulation modeling.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2002190 Georgia Case: Tourism Expenses of International Visitors on the Basis of Growing Attractiveness
Authors: Nino Abesadze, Marine Mindorashvili, Nino Paresashvili
Abstract:
At present actual tourism indicators cannot be calculated in Georgia, making it impossible to perform their quantitative analysis. Therefore, the study conducted by us is highly important from a theoretical as well as practical standpoint. The main purpose of the article is to make complex statistical analysis of tourist expenses of foreign visitors and to calculate statistical attractiveness indices of the tourism potential of Georgia. During the research, the method involving random and proportional selection has been applied. Computer software SPSS was used to compute statistical data for corresponding analysis. Corresponding methodology of tourism statistics was implemented according to international standards. Important information was collected and grouped from major Georgian airports, and a representative population of foreign visitors and a rule of selection of respondents were determined. The results show a trend of growth in tourist numbers and the share of tourists from post-soviet countries are constantly increasing. The level of satisfaction with tourist facilities and quality of service has improved, but still we have a problem of disparity between the service quality and the prices. The design of tourist expenses of foreign visitors is diverse; competitiveness of tourist products of Georgian tourist companies is higher. Attractiveness of popular cities of Georgia has increased by 43%.
Keywords: Tourist, expenses, indexes, statistics, analysis.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 915189 Design and Implementation of a Counting and Differentiation System for Vehicles through Video Processing
Authors: Derlis Gregor, Kevin Cikel, Mario Arzamendia, Raúl Gregor
Abstract:
This paper presents a self-sustaining mobile system for counting and classification of vehicles through processing video. It proposes a counting and classification algorithm divided in four steps that can be executed multiple times in parallel in a SBC (Single Board Computer), like the Raspberry Pi 2, in such a way that it can be implemented in real time. The first step of the proposed algorithm limits the zone of the image that it will be processed. The second step performs the detection of the mobile objects using a BGS (Background Subtraction) algorithm based on the GMM (Gaussian Mixture Model), as well as a shadow removal algorithm using physical-based features, followed by morphological operations. In the first step the vehicle detection will be performed by using edge detection algorithms and the vehicle following through Kalman filters. The last step of the proposed algorithm registers the vehicle passing and performs their classification according to their areas. An auto-sustainable system is proposed, powered by batteries and photovoltaic solar panels, and the data transmission is done through GPRS (General Packet Radio Service)eliminating the need of using external cable, which will facilitate it deployment and translation to any location where it could operate. The self-sustaining trailer will allow the counting and classification of vehicles in specific zones with difficult access.Keywords: Intelligent transportation systems, object detection, video processing, road traffic, vehicle counting, vehicle classification.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1624188 Personas Help Understand Users’ Needs, Goals and Desires in an Online Institutional Repository
Authors: Maha ALjohani, James Blustein
Abstract:
Communicating users' needs, goals and problems help designers and developers overcome challenges faced by end users. Personas are used to represent end users’ needs. In our research, creating personas allowed the following questions to be answered: Who are the potential user groups? What do they want to achieve by using the service? What are the problems that users face? What should the service provide to them? To develop realistic personas, we conducted a focus group discussion with undergraduate and graduate students and also interviewed a university librarian. The personas were created to help evaluating the Institutional Repository that is based on the DSpace system. The profiles helped to communicate users' needs, abilities, tasks, and problems, and the task scenarios used in the heuristic evaluation were based on these personas. Four personas resulted of a focus group discussion with undergraduate and graduate students and from interviewing a university librarian. We then used these personas to create focused task-scenarios for a heuristic evaluation on the system interface to ensure that it met users' needs, goals, problems and desires. In this paper, we present the process that we used to create the personas that led to devise the task scenarios used in the heuristic evaluation as a follow up study of the DSpace university repository.Keywords: Heuristic Evaluation, Institutional Repositories, User Experience, Human Computer Interaction, User Profiles, Personas, Task Scenarios, Heuristics.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2167187 A Study of Two Disease Models: With and Without Incubation Period
Authors: H. C. Chinwenyi, H. D. Ibrahim, J. O. Adekunle
Abstract:
The incubation period is defined as the time from infection with a microorganism to development of symptoms. In this research, two disease models: one with incubation period and another without incubation period were studied. The study involves the use of a mathematical model with a single incubation period. The test for the existence and stability of the disease free and the endemic equilibrium states for both models were carried out. The fourth order Runge-Kutta method was used to solve both models numerically. Finally, a computer program in MATLAB was developed to run the numerical experiments. From the results, we are able to show that the endemic equilibrium state of the model with incubation period is locally asymptotically stable whereas the endemic equilibrium state of the model without incubation period is unstable under certain conditions on the given model parameters. It was also established that the disease free equilibrium states of the model with and without incubation period are locally asymptotically stable. Furthermore, results from numerical experiments using empirical data obtained from Nigeria Centre for Disease Control (NCDC) showed that the overall population of the infected people for the model with incubation period is higher than that without incubation period. We also established from the results obtained that as the transmission rate from susceptible to infected population increases, the peak values of the infected population for the model with incubation period decrease and are always less than those for the model without incubation period.
Keywords: Asymptotic stability, incubation period, Routh-Hurwitz criterion, Runge Kutta method.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 684186 On The Analysis of a Compound Neural Network for Detecting Atrio Ventricular Heart Block (AVB) in an ECG Signal
Authors: Salama Meghriche, Amer Draa, Mohammed Boulemden
Abstract:
Heart failure is the most common reason of death nowadays, but if the medical help is given directly, the patient-s life may be saved in many cases. Numerous heart diseases can be detected by means of analyzing electrocardiograms (ECG). Artificial Neural Networks (ANN) are computer-based expert systems that have proved to be useful in pattern recognition tasks. ANN can be used in different phases of the decision-making process, from classification to diagnostic procedures. This work concentrates on a review followed by a novel method. The purpose of the review is to assess the evidence of healthcare benefits involving the application of artificial neural networks to the clinical functions of diagnosis, prognosis and survival analysis, in ECG signals. The developed method is based on a compound neural network (CNN), to classify ECGs as normal or carrying an AtrioVentricular heart Block (AVB). This method uses three different feed forward multilayer neural networks. A single output unit encodes the probability of AVB occurrences. A value between 0 and 0.1 is the desired output for a normal ECG; a value between 0.1 and 1 would infer an occurrence of an AVB. The results show that this compound network has a good performance in detecting AVBs, with a sensitivity of 90.7% and a specificity of 86.05%. The accuracy value is 87.9%.Keywords: Artificial neural networks, Electrocardiogram(ECG), Feed forward multilayer neural network, Medical diagnosis, Pattern recognitionm, Signal processing.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2472185 Hierarchical Operation Strategies for Grid Connected Building Microgrid with Energy Storage and Photovoltatic Source
Authors: Seon-Ho Yoon, Jin-Young Choi, Dong-Jun Won
Abstract:
This paper presents hierarchical operation strategies which are minimizing operation error between day ahead operation plan and real time operation. Operating power systems between centralized and decentralized approaches can be represented as hierarchical control scheme, featured as primary control, secondary control and tertiary control. Primary control is known as local control, featuring fast response. Secondary control is referred to as microgrid Energy Management System (EMS). Tertiary control is responsible of coordinating the operations of multi-microgrids. In this paper, we formulated 3 stage microgrid operation strategies which are similar to hierarchical control scheme. First stage is to set a day ahead scheduled output power of Battery Energy Storage System (BESS) which is only controllable source in microgrid and it is optimized to minimize cost of exchanged power with main grid using Particle Swarm Optimization (PSO) method. Second stage is to control the active and reactive power of BESS to be operated in day ahead scheduled plan in case that State of Charge (SOC) error occurs between real time and scheduled plan. The third is rescheduling the system when the predicted error is over the limited value. The first stage can be compared with the secondary control in that it adjusts the active power. The second stage is comparable to the primary control in that it controls the error in local manner. The third stage is compared with the secondary control in that it manages power balancing. The proposed strategies will be applied to one of the buildings in Electronics and Telecommunication Research Institute (ETRI). The building microgrid is composed of Photovoltaic (PV) generation, BESS and load and it will be interconnected with the main grid. Main purpose of that is minimizing operation cost and to be operated in scheduled plan. Simulation results support validation of proposed strategies.
Keywords: Battery energy storage system, energy management system, microgrid, particle swarm optimization.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1093184 Computer Study of Cluster Mechanism of Anti-greenhouse Effect
Authors: A. Galashev
Abstract:
Absorption spectra of infra-red (IR) radiation of the disperse water medium absorbing the most important greenhouse gases: CO2 , N2O , CH4 , C2H2 , C2H6 have been calculated by the molecular dynamics method. Loss of the absorbing ability at the formation of clusters due to a reduction of the number of centers interacting with IR radiation, results in an anti-greenhouse effect. Absorption of O3 molecules by the (H2O)50 cluster is investigated at its interaction with Cl- ions. The splitting of ozone molecule on atoms near to cluster surface was observed. Interaction of water cluster with Cl- ions causes the increase of integrated intensity of emission spectra of IR radiation, and also essential reduction of the similar characteristic of Raman spectrum. Relative integrated intensity of absorption of IR radiation for small water clusters was designed. Dependences of the quantity of weight on altitude for vapor of monomers, clusters, droplets, crystals and mass of all moisture were determined. The anti-greenhouse effect of clusters was defined as the difference of increases of average global temperature of the Earth, caused by absorption of IR radiation by free water molecules forming clusters, and absorption of clusters themselves. The greenhouse effect caused by clusters makes 0.53 K, and the antigreenhouse one is equal to 1.14 K. The increase of concentration of CO2 in the atmosphere does not always correlate with the amplification of greenhouse effect.Keywords: Greenhouse gases, infrared absorption and Raman spectra, molecular dynamics method, water clusters.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1486183 A Generic Middleware to Instantly Sync Intensive Writes of Heterogeneous Massive Data via Internet
Authors: Haitao Yang, Zhenjiang Ruan, Fei Xu, Lanting Xia
Abstract:
Industry data centers often need to sync data changes reliably and instantly from a large-scale of heterogeneous autonomous relational databases accessed via the not-so-reliable Internet, for which a practical generic sync middleware of low maintenance and operation costs is most wanted. To this demand, this paper presented a generic sync middleware system (GSMS), which has been developed, applied and optimized since 2006, holding the principles or advantages that it must be SyncML-compliant and transparent to data application layer logic without referring to implementation details of databases synced, does not rely on host computer operating systems deployed, and its construction is light weighted and hence of low cost. Regarding these hard commitments of developing GSMS, in this paper we stressed the significant optimization breakthrough of GSMS sync delay being well below a fraction of millisecond per record sync. A series of ultimate tests with GSMS sync performance were conducted for a persuasive example, in which the source relational database underwent a broad range of write loads (from one thousand to one million intensive writes within a few minutes). All these tests showed that the performance of GSMS is competent and smooth even under ultimate write loads.
Keywords: Heterogeneous massive data, instantly sync intensive writes, Internet generic middleware design, optimization.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 454182 Seismic Behavior and Loss Assessment of High-Rise Buildings with Light Gauge Steel-Concrete Hybrid Structure
Authors: Bing Lu, Shuang Li, Hongyuan Zhou
Abstract:
The steel-concrete hybrid structure has been extensively employed in high-rise buildings and super high-rise buildings. The light gauge steel-concrete hybrid structure, including light gauge steel structure and concrete hybrid structure, is a type of steel-concrete hybrid structure, which possesses some advantages of light gauge steel structure and concrete hybrid structure. The seismic behavior and loss assessment of three high-rise buildings with three different concrete hybrid structures were investigated through finite element software. The three concrete hybrid structures are reinforced concrete column-steel beam (RC-S) hybrid structure, concrete-filled steel tube column-steel beam (CFST-S) hybrid structure, and tubed concrete column-steel beam (TC-S) hybrid structure. The nonlinear time-history analysis of three high-rise buildings under 80 earthquakes was carried out. After simulation, it indicated that the seismic performances of three high-rise buildings were superior. Under extremely rare earthquakes, the maximum inter-story drifts of three high-rise buildings are significantly lower than 1/50. The inter-story drift and floor acceleration of high-rise building with CFST-S hybrid structure were bigger than those of high-rise buildings with RC-S hybrid structure, and smaller than those of high-rise building with TC-S hybrid structure. Then, based on the time-history analysis results, the post-earthquake repair cost ratio and repair time of three high-rise buildings were predicted through an economic performance analysis method proposed in FEMA-P58 report. Under frequent earthquakes, basic earthquakes and rare earthquakes, the repair cost ratio and repair time of three high-rise buildings were less than 5% and 15 days, respectively. Under extremely rare earthquakes, the repair cost ratio and repair time of high-rise buildings with TC-S hybrid structure were the most among three high rise buildings. Due to the advantages of CFST-S hybrid structure, it could be extensively employed in high-rise buildings subjected to earthquake excitations.
Keywords: seismic behavior, loss assessment, light gauge steel, concrete hybrid structure, high-rise building, time-history analysis
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 491181 Texture Based Weed Detection Using Multi Resolution Combined Statistical and Spatial Frequency (MRCSF)
Authors: R.S.Sabeenian, V.Palanisamy
Abstract:
Texture classification is a trendy and a catchy technology in the field of texture analysis. Textures, the repeated patterns, have different frequency components along different orientations. Our work is based on Texture Classification and its applications. It finds its applications in various fields like Medical Image Classification, Computer Vision, Remote Sensing, Agricultural Field, and Textile Industry. Weed control has a major effect on agriculture. A large amount of herbicide has been used for controlling weeds in agriculture fields, lawns, golf courses, sport fields, etc. Random spraying of herbicides does not meet the exact requirement of the field. Certain areas in field have more weed patches than estimated. So, we need a visual system that can discriminate weeds from the field image which will reduce or even eliminate the amount of herbicide used. This would allow farmers to not use any herbicides or only apply them where they are needed. A machine vision precision automated weed control system could reduce the usage of chemicals in crop fields. In this paper, an intelligent system for automatic weeding strategy Multi Resolution Combined Statistical & spatial Frequency is used to discriminate the weeds from the crops and to classify them as narrow, little and broad weeds.Keywords: crop weed discrimination, MRCSF, MRFM, Weeddetection, Spatial Frequency.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1827180 Simulation and Design of an Aerospace Mission Powered by “Candy” Type Fuel Engines
Authors: N. Hernández Huertas, F. Rojas Mora
Abstract:
Sounding rockets are aerospace vehicles that were developed in the mid-20th century, and since then numerous investigations have been executed with the aim of innovate in this type of technology. However, the costs associated to the production of this type of technology are usually quite high, and therefore the challenge that exists today is to be able to reduce them. In this way, the main objective of this document is to present the design process of a Colombian aerospace mission capable to reach the thermosphere using low-cost “Candy” type solid fuel engines. This mission is the latest development of the Uniandes Aerospace Project (PUA for its Spanish acronym), which is an undergraduate and postgraduate research group at Universidad de los Andes (Bogotá, Colombia), dedicated to incurring in this type of technology. In this way, the investigations that have been carried out on Candy-type solid fuel, which is a compound of potassium nitrate and sorbitol, have allowed the production of engines powerful enough to reach space, and which represents a unique technological advance in Latin America and an important development in experimental rocketry. In this way, following the engineering iterative design methodology was possible to design a 2-stage sounding rocket with 1 solid fuel engine in each one, which was then simulated in RockSim V9.0 software and reached an apogee of approximately 150 km above sea level. Similarly, a speed equal to 5 Mach was obtained, which after performing a finite element analysis, it was shown that the rocket is strong enough to be able to withstand such speeds. Under these premises, it was demonstrated that it is possible to build a high-power aerospace mission at low cost, using Candy-type solid fuel engines. For this reason, the feasibility of carrying out similar missions clearly depends on the ability to replicate the engines in the best way, since as mentioned above, the design of the rocket is adequate to reach supersonic speeds and reach space. Consequently, with a team of at least 3 members, the mission can be obtained in less than 3 months. Therefore, when publishing this project, it is intended to be a reference for future research in this field and benefit the industry.
Keywords: Aerospace missions, candy type solid propellant engines, design of solid rockets, experimental rocketry, low costs missions.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 798179 Numerical Solution of Transient Natural Convection in Vertical Heated Rectangular Channel between Two Vertical Parallel MTR-Type Fuel Plates
Authors: Djalal Hamed
Abstract:
The aim of this paper is to perform, by mean of the finite volume method, a numerical solution of the transient natural convection in a narrow rectangular channel between two vertical parallel Material Testing Reactor (MTR)-type fuel plates, imposed under a heat flux with a cosine shape to determine the margin of the nuclear core power at which the natural convection cooling mode can ensure a safe core cooling, where the cladding temperature should not reach a specific safety limits (90 °C). For this purpose, a computer program is developed to determine the principal parameters related to the nuclear core safety, such as the temperature distribution in the fuel plate and in the coolant (light water) as a function of the reactor core power. Throughout the obtained results, we noticed that the core power should not reach 400 kW, to ensure a safe passive residual heat removing from the nuclear core by the upward natural convection cooling mode.
Keywords: Buoyancy force, friction force, friction factor, finite volume method, transient natural convection, thermal hydraulic analysis, vertical heated rectangular channel.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 754178 Application of Building Information Modeling in Energy Management of Individual Departments Occupying University Facilities
Authors: Kung-Jen Tu, Danny Vernatha
Abstract:
To assist individual departments within universities in their energy management tasks, this study explores the application of Building Information Modeling in establishing the ‘BIM based Energy Management Support System’ (BIM-EMSS). The BIM-EMSS consists of six components: (1) sensors installed for each occupant and each equipment, (2) electricity sub-meters (constantly logging lighting, HVAC, and socket electricity consumptions of each room), (3) BIM models of all rooms within individual departments’ facilities, (4) data warehouse (for storing occupancy status and logged electricity consumption data), (5) building energy management system that provides energy managers with various energy management functions, and (6) energy simulation tool (such as eQuest) that generates real time 'standard energy consumptions' data against which 'actual energy consumptions' data are compared and energy efficiency evaluated. Through the building energy management system, the energy manager is able to (a) have 3D visualization (BIM model) of each room, in which the occupancy and equipment status detected by the sensors and the electricity consumptions data logged are displayed constantly; (b) perform real time energy consumption analysis to compare the actual and standard energy consumption profiles of a space; (c) obtain energy consumption anomaly detection warnings on certain rooms so that energy management corrective actions can be further taken (data mining technique is employed to analyze the relation between space occupancy pattern with current space equipment setting to indicate an anomaly, such as when appliances turn on without occupancy); and (d) perform historical energy consumption analysis to review monthly and annually energy consumption profiles and compare them against historical energy profiles. The BIM-EMSS was further implemented in a research lab in the Department of Architecture of NTUST in Taiwan and implementation results presented to illustrate how it can be used to assist individual departments within universities in their energy management tasks.Keywords: Sensor, electricity sub-meters, database, energy anomaly detection.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2284177 MATLAB-based System for Centralized Monitoring and Self Restoration against Fiber Fault in FTTH
Authors: Mohammad Syuhaimi Ab-Rahman, Boonchuan Ng, Kasmiran Jumari
Abstract:
This paper presented a MATLAB-based system named Smart Access Network Testing, Analyzing and Database (SANTAD), purposely for in-service transmission surveillance and self restoration against fiber fault in fiber-to-the-home (FTTH) access network. The developed program will be installed with optical line terminal (OLT) at central office (CO) to monitor the status and detect any fiber fault that occurs in FTTH downwardly from CO towards residential customer locations. SANTAD is interfaced with optical time domain reflectometer (OTDR) to accumulate every network testing result to be displayed on a single computer screen for further analysis. This program will identify and present the parameters of each optical fiber line such as the line's status either in working or nonworking condition, magnitude of decreasing at each point, failure location, and other details as shown in the OTDR's screen. The failure status will be delivered to field engineers for promptly actions, meanwhile the failure line will be diverted to protection line to ensure the traffic flow continuously. This approach has a bright prospect to improve the survivability and reliability as well as increase the efficiency and monitoring capabilities in FTTH.
Keywords: MATLAB, SANTAD, in-service transmission surveillance, self restoration, fiber fault, FTTH
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2115176 Local Buckling of Web-Core and Foam-Core Sandwich Panels
Authors: Ali N. Suri, Ahmad A. Al-Makhlufi
Abstract:
Sandwich construction is widely accepted as a method of construction especially in the aircraft industry. It is a type of stressed skin construction formed by bonding two thin faces to a thick core, the faces resist all of the applied edge loads and provide all or nearly all of the required rigidities, the core spaces the faces to increase cross section moment of inertia about common neutral axis and transmit shear between them provides a perfect bond between core and faces is made.
Material for face sheets can be of metal or reinforced plastics laminates, core material can be metallic cores of thin sheets forming corrugation or honeycomb, or non metallic core of Balsa wood, plastic foams, or honeycomb made of reinforced plastics.
For in plane axial loading web core and web-foam core Sandwich panels can fail by local buckling of plates forming the cross section with buckling wave length of the order of length of spacing between webs.
In this study local buckling of web core and web-foam core Sandwich panels is carried out for given materials of facing and core, and given panel overall dimension for different combinations of cross section geometries.
The Finite Strip Method is used for the analysis, and Fortran based computer program is developed and used.
Keywords: Local Buckling, Finite Strip, Sandwich panels, Web and foam core.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2255175 3D Liver Segmentation from CT Images Using a Level Set Method Based on a Shape and Intensity Distribution Prior
Authors: Nuseiba M. Altarawneh, Suhuai Luo, Brian Regan, Guijin Tang
Abstract:
Liver segmentation from medical images poses more challenges than analogous segmentations of other organs. This contribution introduces a liver segmentation method from a series of computer tomography images. Overall, we present a novel method for segmenting liver by coupling density matching with shape priors. Density matching signifies a tracking method which operates via maximizing the Bhattacharyya similarity measure between the photometric distribution from an estimated image region and a model photometric distribution. Density matching controls the direction of the evolution process and slows down the evolving contour in regions with weak edges. The shape prior improves the robustness of density matching and discourages the evolving contour from exceeding liver’s boundaries at regions with weak boundaries. The model is implemented using a modified distance regularized level set (DRLS) model. The experimental results show that the method achieves a satisfactory result. By comparing with the original DRLS model, it is evident that the proposed model herein is more effective in addressing the over segmentation problem. Finally, we gauge our performance of our model against matrices comprising of accuracy, sensitivity, and specificity.
Keywords: Bhattacharyya distance, distance regularized level set (DRLS) model, liver segmentation, level set method.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2338174 Pre-Service Teachers’ Assessment of Information Technology Application to Instruction
Authors: Adesanya Anuoluwapo Olusola
Abstract:
Technology has moved into the classroom, and it becomes difficult talking of achievement in and attitude to learning without making mention of it. The use of technology makes learning easy, real and practical as it motivates learners, sustains their interest and improves their attitude to learning. This study, therefore examined the pre-service teachers’ assessment of information technology application to instruction. The use of technology emphasizes and encourages active learning in the classroom. The study involved 100 pre-service teachers in the selected two (2) Colleges of Education, Nigeria. Purposive random sampling was used in selecting the participants and ex-post facto design was adopted the in which there is no manipulation of variables. Two valid and reliable instruments were used for data collection: Access Point ICT facilities and Application of ICT. The study established that pre-service teachers have less access to ICT facilities and Application of ICT in the college, apart from those students having the access outside the college. Also fewer pre-service teachers used ICT facilities on weekly and monthly bases. It was concluded that the establishment of students’ resources centres and Campus wide wireless connectivity must be implemented so as to improve and enhance students’ achievement in and attitude to learning. The time and attention devoted to learning activities and strategic specialized ICT skills and requisite entrepreneur skills should be increased so as to have easy access to information sources and be able to apply it in teaching process.
Keywords: Computer, ICT Application, Learning Facilities, Pre-Service Teachers.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1932173 Topics of Blockchain Technology to Teach at Community College
Authors: Penn P. Wu, Jeannie Jo
Abstract:
Blockchain technology has rapidly gained popularity in industry. This paper attempts to assist academia to answer four questions. First, should community colleges begin offering education to nurture blockchain-literate students for the job market? Second, what are the appropriate topical areas to cover? Third, should it be an individual course? And forth, should it be a technical or management course? This paper starts with identifying the knowledge domains of blockchain technology and the topical areas each domain has, and continues with placing them in appropriate academic territories (Computer Sciences vs. Business) and subjects (programming, management, marketing, and laws), and then develops an evaluation model to determine the appropriate topical area for community colleges to teach. The evaluation is based on seven factors: maturity of technology, impacts on management, real-world applications, subject classification, knowledge prerequisites, textbook readiness, and recommended pedagogies. The evaluation results point to an interesting direction that offering an introductory course is an ideal option to guide students through the learning journey of what blockchain is and how it applies to business. Such an introductory course does not need to engage students in the discussions of mathematics and sciences that make blockchain technologies possible. While it is inevitable to brief technical topics to help students build a solid knowledge foundation of blockchain technologies, community colleges should avoid offering students a course centered on the discussion of developing blockchain applications.
Keywords: Blockchain, pedagogies, blockchain technologies, blockchain course, blockchain pedagogies.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 930172 Automatic Segmentation of Dermoscopy Images Using Histogram Thresholding on Optimal Color Channels
Authors: Rahil Garnavi, Mohammad Aldeen, M. Emre Celebi, Alauddin Bhuiyan, Constantinos Dolianitis, George Varigos
Abstract:
Automatic segmentation of skin lesions is the first step towards development of a computer-aided diagnosis of melanoma. Although numerous segmentation methods have been developed, few studies have focused on determining the most discriminative and effective color space for melanoma application. This paper proposes a novel automatic segmentation algorithm using color space analysis and clustering-based histogram thresholding, which is able to determine the optimal color channel for segmentation of skin lesions. To demonstrate the validity of the algorithm, it is tested on a set of 30 high resolution dermoscopy images and a comprehensive evaluation of the results is provided, where borders manually drawn by four dermatologists, are compared to automated borders detected by the proposed algorithm. The evaluation is carried out by applying three previously used metrics of accuracy, sensitivity, and specificity and a new metric of similarity. Through ROC analysis and ranking the metrics, it is shown that the best results are obtained with the X and XoYoR color channels which results in an accuracy of approximately 97%. The proposed method is also compared with two state-ofthe- art skin lesion segmentation methods, which demonstrates the effectiveness and superiority of the proposed segmentation method.Keywords: Border detection, Color space analysis, Dermoscopy, Histogram thresholding, Melanoma, Segmentation.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2085171 3D Rendering of American Sign Language Finger-Spelling: A Comparative Study of Two Animation Techniques
Authors: Nicoletta Adamo-Villani
Abstract:
In this paper we report a study aimed at determining the most effective animation technique for representing ASL (American Sign Language) finger-spelling. Specifically, in the study we compare two commonly used 3D computer animation methods (keyframe animation and motion capture) in order to ascertain which technique produces the most 'accurate', 'readable', and 'close to actual signing' (i.e. realistic) rendering of ASL finger-spelling. To accomplish this goal we have developed 20 animated clips of fingerspelled words and we have designed an experiment consisting of a web survey with rating questions. 71 subjects ages 19-45 participated in the study. Results showed that recognition of the words was correlated with the method used to animate the signs. In particular, keyframe technique produced the most accurate representation of the signs (i.e., participants were more likely to identify the words correctly in keyframed sequences rather than in motion captured ones). Further, findings showed that the animation method had an effect on the reported scores for readability and closeness to actual signing; the estimated marginal mean readability and closeness was greater for keyframed signs than for motion captured signs. To our knowledge, this is the first study aimed at measuring and comparing accuracy, readability and realism of ASL animations produced with different techniques.Keywords: 3D Animation, American Sign Language, DeafEducation, Motion Capture.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1998170 Facial Expression Phoenix (FePh): An Annotated Sequenced Dataset for Facial and Emotion-Specified Expressions in Sign Language
Authors: Marie Alaghband, Niloofar Yousefi, Ivan Garibay
Abstract:
Facial expressions are important parts of both gesture and sign language recognition systems. Despite the recent advances in both fields, annotated facial expression datasets in the context of sign language are still scarce resources. In this manuscript, we introduce an annotated sequenced facial expression dataset in the context of sign language, comprising over 3000 facial images extracted from the daily news and weather forecast of the public tv-station PHOENIX. Unlike the majority of currently existing facial expression datasets, FePh provides sequenced semi-blurry facial images with different head poses, orientations, and movements. In addition, in the majority of images, identities are mouthing the words, which makes the data more challenging. To annotate this dataset we consider primary, secondary, and tertiary dyads of seven basic emotions of "sad", "surprise", "fear", "angry", "neutral", "disgust", and "happy". We also considered the "None" class if the image’s facial expression could not be described by any of the aforementioned emotions. Although we provide FePh as a facial expression dataset of signers in sign language, it has a wider application in gesture recognition and Human Computer Interaction (HCI) systems.Keywords: Annotated Facial Expression Dataset, Sign Language Recognition, Gesture Recognition, Sequenced Facial Expression Dataset.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 719169 Enhancing Cache Performance Based on Improved Average Access Time
Authors: Jasim. A. Ghaeb
Abstract:
A high performance computer includes a fast processor and millions bytes of memory. During the data processing, huge amount of information are shuffled between the memory and processor. Because of its small size and its effectiveness speed, cache has become a common feature of high performance computers. Enhancing cache performance proved to be essential in the speed up of cache-based computers. Most enhancement approaches can be classified as either software based or hardware controlled. The performance of the cache is quantified in terms of hit ratio or miss ratio. In this paper, we are optimizing the cache performance based on enhancing the cache hit ratio. The optimum cache performance is obtained by focusing on the cache hardware modification in the way to make a quick rejection to the missed line's tags from the hit-or miss comparison stage, and thus a low hit time for the wanted line in the cache is achieved. In the proposed technique which we called Even- Odd Tabulation (EOT), the cache lines come from the main memory into cache are classified in two types; even line's tags and odd line's tags depending on their Least Significant Bit (LSB). This division is exploited by EOT technique to reject the miss match line's tags in very low time compared to the time spent by the main comparator in the cache, giving an optimum hitting time for the wanted cache line. The high performance of EOT technique against the familiar mapping technique FAM is shown in the simulated results.Keywords: Caches, Cache performance, Hit time, Cache hit ratio, Cache mapping, Cache memory.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1678168 Artificial Intelligent Approach for Machining Titanium Alloy in a Nonconventional Process
Authors: Md. Ashikur Rahman Khan, M. M. Rahman, K. Kadirgama
Abstract:
Artificial neural networks (ANN) are used in distinct researching fields and professions, and are prepared by cooperation of scientists in different fields such as computer engineering, electronic, structure, biology and so many different branches of science. Many models are built correlating the parameters and the outputs in electrical discharge machining (EDM) concern for different types of materials. Up till now model for Ti-5Al-2.5Sn alloy in the case of electrical discharge machining performance characteristics has not been developed. Therefore, in the present work, it is attempted to generate a model of material removal rate (MRR) for Ti-5Al-2.5Sn material by means of Artificial Neural Network. The experimentation is performed according to the design of experiment (DOE) of response surface methodology (RSM). To generate the DOE four parameters such as peak current, pulse on time, pulse off time and servo voltage and one output as MRR are considered. Ti-5Al-2.5Sn alloy is machined with positive polarity of copper electrode. Finally the developed model is tested with confirmation test. The confirmation test yields an error as within the agreeable limit. To investigate the effect of the parameters on performance sensitivity analysis is also carried out which reveals that the peak current having more effect on EDM performance.
Keywords: Ti-5Al-2.5Sn, material removal rate, copper tungsten, positive polarity, artificial neural network, multi-layer perceptron.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2399167 An Educational Data Mining System for Advising Higher Education Students
Authors: Heba Mohammed Nagy, Walid Mohamed Aly, Osama Fathy Hegazy
Abstract:
Educational data mining is a specific data mining field applied to data originating from educational environments, it relies on different approaches to discover hidden knowledge from the available data. Among these approaches are machine learning techniques which are used to build a system that acquires learning from previous data. Machine learning can be applied to solve different regression, classification, clustering and optimization problems.
In our research, we propose a “Student Advisory Framework” that utilizes classification and clustering to build an intelligent system. This system can be used to provide pieces of consultations to a first year university student to pursue a certain education track where he/she will likely succeed in, aiming to decrease the high rate of academic failure among these students. A real case study in Cairo Higher Institute for Engineering, Computer Science and Management is presented using real dataset collected from 2000−2012.The dataset has two main components: pre-higher education dataset and first year courses results dataset. Results have proved the efficiency of the suggested framework.
Keywords: Classification, Clustering, Educational Data Mining (EDM), Machine Learning.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 5213