Search results for: light processing technology
419 Optimizing the Performance of Thermoelectric for Cooling Computer Chips Using Different Types of Electrical Pulses
Authors: Saleh Alshehri
Abstract:
Thermoelectric technology is currently being used in many industrial applications for cooling, heating and generating electricity. This research mainly focuses on using thermoelectric to cool down high-speed computer chips at different operating conditions. A previously developed and validated three-dimensional model for optimizing and assessing the performance of cascaded thermoelectric and non-cascaded thermoelectric is used in this study to investigate the possibility of decreasing the hotspot temperature of computer chip. Additionally, a test assembly is built and tested at steady-state and transient conditions. The obtained optimum thermoelectric current at steady-state condition is used to conduct a number of pulsed tests (i.e. transient tests) with different shapes to cool the computer chips hotspots. The results of the steady-state tests showed that at hotspot heat rate of 15.58 W (5.97 W/cm2), using thermoelectric current of 4.5 A has resulted in decreasing the hotspot temperature at open circuit condition (89.3 °C) by 50.1 °C. Maximum and minimum hotspot temperatures have been affected by ON and OFF duration of the electrical current pulse. Maximum hotspot temperature was resulted by longer OFF pulse period. In addition, longer ON pulse period has generated the minimum hotspot temperature.
Keywords: Thermoelectric generator, thermoelectric cooler, chip hotspots, electronic cooling.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 619418 Exergetic and Life Cycle Assessment Analyses of Integrated Biowaste Gasification-Combustion System: A Study Case
Authors: Anabel Fernandez, Leandro Rodriguez-Ortiz, Rosa Rodríguez
Abstract:
Due to the negative impact of fossil fuels, renewable energies are promising sources to limit global temperature rise and damage to the environment. Also, the development of technology is focused on obtaining energetic products from renewable sources. In this study, a thermodynamic model including exergy balance and a subsequent Life Cycle Assessment (LCA) were carried out for four subsystems of the integrated gasification-combustion of pinewood. Results of exergy analysis and LCA showed the process feasibility in terms of exergy efficiency and global energy efficiency of the life cycle (GEELC). Moreover, the energy return on investment (EROI) index was calculated. The global exergy efficiency resulted in 67%. For pretreatment, reaction, cleaning, and electric generation subsystems, the results were 85%, 59%, 87%, and 29%, respectively. Results of LCA indicated that the emissions from the electric generation caused the most damage to the atmosphere, water, and soil. GEELC resulted in 31.09% for the global process. This result suggested the environmental feasibility of an integrated gasification-combustion system. EROI resulted in 3.15, which determines the sustainability of the process.
Keywords: Exergy analysis, Life Cycle Assessment, LCA, renewability, sustainability.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 462417 Exploring Social Impact of Emerging Technologies from Futuristic Data
Authors: Heeyeul Kwon, Yongtae Park
Abstract:
Despite the highly touted benefits, emerging technologies have unleashed pervasive concerns regarding unintended and unforeseen social impacts. Thus, those wishing to create safe and socially acceptable products need to identify such side effects and mitigate them prior to the market proliferation. Various methodologies in the field of technology assessment (TA), namely Delphi, impact assessment, and scenario planning, have been widely incorporated in such a circumstance. However, literatures face a major limitation in terms of sole reliance on participatory workshop activities. They unfortunately missed out the availability of a massive untapped data source of futuristic information flooding through the Internet. This research thus seeks to gain insights into utilization of futuristic data, future-oriented documents from the Internet, as a supplementary method to generate social impact scenarios whilst capturing perspectives of experts from a wide variety of disciplines. To this end, network analysis is conducted based on the social keywords extracted from the futuristic documents by text mining, which is then used as a guide to produce a comprehensive set of detailed scenarios. Our proposed approach facilitates harmonized depictions of possible hazardous consequences of emerging technologies and thereby makes decision makers more aware of, and responsive to, broad qualitative uncertainties.
Keywords: Emerging technologies, futuristic data, scenario, text mining.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2390416 An Evaluation of Kahoot Application and Its Environment as a Learning Tool
Authors: Muhammad Yasir Babar, Ebrahim Panah
Abstract:
Over the past 20 years, internet has seen continual advancement and with the advent of online technology, various types of web-based games have been developed. Games are frequently being used among different age groups from baby boomers to generation Z. Games are not only used for entertainment but also utilized as a learning approach transmitting education to a level that is more interesting and effective for students. One of the popular web-based education games is Kahoot with growing popularity and usage, which is being used in different fields of studies. However, little knowledge is available on university students’ perception of Kahoot environment and application for learning subjects. Hence, the objective of the current study is to investigate students’ perceptions of Kahoot application and environment as a learning tool. The study employed a survey approach by distributing Google Forms –created questionnaire, with high level of reliability index, to 62 students (11 males and 51 females). The findings show that students have positive attitudes towards Kahoot application and its environment for learning. Regarding Kahoot application, it was indicated that activities created using Kahoot are more interesting for students, Kahoot is useful for collaborative learning, and Kahoot enhances interest in learning lesson. In terms of Kahoot environment, it was found that using this application through mobile is easy for students, its design is simple and useful, Kahoot-created activities can easily be shared, and the application can easily be used on any platform. The findings of the study have implications for instructors, policymakers and curriculum developers.
Keywords: Application, environment, Kahoot, learning tool.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 791415 An Algorithm for Autonomous Aerial Navigation using MATLAB® Mapping Tool Box
Authors: Mansoor Ahsan, Suhail Akhtar, Adnan Ali, Farrukh Mazhar, Muddssar Khalid
Abstract:
In the present era of aviation technology, autonomous navigation and control have emerged as a prime area of active research. Owing to the tremendous developments in the field, autonomous controls have led today’s engineers to claim that future of aerospace vehicle is unmanned. Development of guidance and navigation algorithms for an unmanned aerial vehicle (UAV) is an extremely challenging task, which requires efforts to meet strict, and at times, conflicting goals of guidance and control. In this paper, aircraft altitude and heading controllers and an efficient algorithm for self-governing navigation using MATLAB® mapping toolbox is presented which also enables loitering of a fixed wing UAV over a specified area. For this purpose, a nonlinear mathematical model of a UAV is used. The nonlinear model is linearized around a stable trim point and decoupled for controller design. The linear controllers are tested on the nonlinear aircraft model and navigation algorithm is subsequently developed for for autonomous flight of the UAV. The results are presented for trajectory controllers and waypoint based navigation. Our investigation reveals that MATLAB® mapping toolbox can be exploited to successfully deliver an efficient algorithm for autonomous aerial navigation for a UAV.
Keywords: Navigation, trajectory-control, unmanned aerial vehicle, PID-control, MATLAB® mapping toolbox.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 4379414 Implementation of a Motion Detection System
Authors: Asif Ansari, T.C.Manjunath, C. Ardil
Abstract:
In today-s competitive environment, the security concerns have grown tremendously. In the modern world, possession is known to be 9/10-ths of the law. Hence, it is imperative for one to be able to safeguard one-s property from worldly harms such as thefts, destruction of property, people with malicious intent etc. Due to the advent of technology in the modern world, the methodologies used by thieves and robbers for stealing have been improving exponentially. Therefore, it is necessary for the surveillance techniques to also improve with the changing world. With the improvement in mass media and various forms of communication, it is now possible to monitor and control the environment to the advantage of the owners of the property. The latest technologies used in the fight against thefts and destruction are the video surveillance and monitoring. By using the technologies, it is possible to monitor and capture every inch and second of the area in interest. However, so far the technologies used are passive in nature, i.e., the monitoring systems only help in detecting the crime but do not actively participate in stopping or curbing the crime while it takes place. Therefore, we have developed a methodology to detect the motion in a video stream environment and this is an idea to ensure that the monitoring systems not only actively participate in stopping the crime, but do so while the crime is taking place. Hence, a system is used to detect any motion in a live streaming video and once motion has been detected in the live stream, the software will activate a warning system and capture the live streaming video.
Keywords: Motion, Detection, System, Video, Crime, Matlab, Surveillance.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 4290413 Implementation of an Improved Secure System Detection for E-passport by using EPC RFID Tags
Authors: A. Baith Mohamed, Ayman Abdel-Hamid, Kareem Youssri Mohamed
Abstract:
Current proposals for E-passport or ID-Card is similar to a regular passport with the addition of tiny contactless integrated circuit (computer chip) inserted in the back cover, which will act as a secure storage device of the same data visually displayed on the photo page of the passport. In addition, it will include a digital photograph that will enable biometric comparison, through the use of facial recognition technology at international borders. Moreover, the e-passport will have a new interface, incorporating additional antifraud and security features. However, its problems are reliability, security and privacy. Privacy is a serious issue since there is no encryption between the readers and the E-passport. However, security issues such as authentication, data protection and control techniques cannot be embedded in one process. In this paper, design and prototype implementation of an improved E-passport reader is presented. The passport holder is authenticated online by using GSM network. The GSM network is the main interface between identification center and the e-passport reader. The communication data is protected between server and e-passport reader by using AES to encrypt data for protection will transferring through GSM network. Performance measurements indicate a 19% improvement in encryption cycles versus previously reported results.
Keywords: RFID "Radio Frequency Identification", EPC"Electronic Product Code", ICAO "International Civil Aviation Organization", IFF "Identify Friend or Foe"
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2600412 Didactical and Semiotic Affordance of GeoGebra in a Productive Mathematical Discourse
Authors: I. Benning
Abstract:
Using technology to expand the learning space is critical for a productive mathematical discourse. This is a case study of two teachers who developed and enacted GeoGebra-based mathematics lessons following their engagement in a two-year professional development. The didactical and semiotic affordance of GeoGebra in widening the learning space for a productive mathematical discourse was explored. The approach of thematic analysis was used for lesson artefact, lesson observation, and interview data. The results indicated that constructing tools in GeoGebra provided a didactical milieu where students used them to explore mathematical concepts with little or no support from their teacher. The prompt feedback from the GeoGebra motivated students to practice mathematical concepts repeatedly in which they privately rethink their solutions before comparing their answers with that of their colleagues. The constructing tools enhanced self-discovery, team spirit, and dialogue among students. With regards to the semiotic construct, the tools widened the physical and psychological atmosphere of the classroom by providing animations that served as virtual concrete to enhance the recording, manipulation, testing of a mathematical idea, construction, and interpretation of geometric objects. These findings advance the discussion of widening the classroom for a productive mathematical discourse within the context of the mathematics curriculum of Ghana and similar sub-Saharan African countries.
Keywords: GeoGebra, theory of didactical situation, semiotic mediation, mathematics laboratory, mathematical discussion.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 393411 Flow-Through Supercritical Installation for Producing Biodiesel Fuel
Authors: Y. A. Shapovalov, F. M. Gumerov, M. K. Nauryzbaev, S. V. Mazanov, R. A. Usmanov, A. V. Klinov, L. K. Safiullina, S. A. Soshin
Abstract:
A flow-through installation was created and manufactured for the transesterification of triglycerides of fatty acids and production of biodiesel fuel under supercritical fluid conditions. Transesterification of rapeseed oil with ethanol was carried out according to two parameters: temperature and the ratio of alcohol/oil mixture at the constant pressure of 19 MPa. The kinetics of the yield of fatty acids ethyl esters (FAEE) was determined in the temperature range of 320-380 °C at the alcohol/oil molar ratio of 6:1-20:1. The content of the formed FAEE was determined by the method of correlation of the resulting biodiesel fuel by its kinematic viscosity. The maximum FAEE yield (about 90%) was obtained within 30 min at the ethanol/oil molar ratio of 12:1 and a temperature of 380 °C. When studying of transesterification of triglycerides, a kinetic model of an isothermal flow reactor was used. The reaction order implemented in the flow reactor has been determined. The first order of the reaction was confirmed by data on the conversion of FAEE during the reaction at different temperatures and the molar ratios of the initial reagents (ethanol/oil). Using the Arrhenius equation, the values of the effective constants of the transesterification reaction rate were calculated at different reaction temperatures. In addition, based on the experimental data, the activation energy and the pre-exponential factor of the transesterification reaction were determined.
Keywords: Biodiesel, fatty acid esters, supercritical fluid technology, transesterification.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 404410 The Client-Supplier Relationship in Managing Innovation: Delineating Defence Industry First Mover Challenges within the Government Contract Competition
Authors: Edward Pol
Abstract:
All companies are confronted with the need to innovate in order to meet market demands. In so doing they are challenged with the dilemma of whether to aim to be first into the market with a new innovative product, or to deliberately wait and learn from a pioneers’ mistakes; potentially avoiding higher risks. It is therefore important to critically understand from a first mover advantage and disadvantage perspective the decision-making implications of defence industry transformation onset by an innovative paradigm shift. This paper will argue that the type of industry characteristics matter, especially when considering what role the clients play in the innovation process and what their level of influence is. Through investigation of qualitative case study research, this inquiry will focus on first mover advantages and first mover disadvantages with a view to establish practical and value-added academic findings by focusing on specific industries where the clients play an active role in cooperation with the supplier innovation. The resulting findings will help managers to mitigate risk in innovative technology introduction. A selection from several defence industry innovations is specifically chosen because of the client–supplier relationship that typically differs from traditional first mover research. In this instance, case studies will be used referencing vertical-take-off-and-landing defence equipment innovations.
Keywords: innovation, pioneer, first mover advantage, first mover disadvantage, risk
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 465409 The Reproducibility and Repeatability of Modified Likelihood Ratio for Forensics Handwriting Examination
Authors: O. Abiodun Adeyinka, B. Adeyemo Adesesan
Abstract:
The forensic use of handwriting depends on the analysis, comparison, and evaluation decisions made by forensic document examiners. When using biometric technology in forensic applications, it is necessary to compute Likelihood Ratio (LR) for quantifying strength of evidence under two competing hypotheses, namely the prosecution and the defense hypotheses wherein a set of assumptions and methods for a given data set will be made. It is therefore important to know how repeatable and reproducible our estimated LR is. This paper evaluated the accuracy and reproducibility of examiners' decisions. Confidence interval for the estimated LR were presented so as not get an incorrect estimate that will be used to deliver wrong judgment in the court of Law. The estimate of LR is fundamentally a Bayesian concept and we used two LR estimators, namely Logistic Regression (LoR) and Kernel Density Estimator (KDE) for this paper. The repeatability evaluation was carried out by retesting the initial experiment after an interval of six months to observe whether examiners would repeat their decisions for the estimated LR. The experimental results, which are based on handwriting dataset, show that LR has different confidence intervals which therefore implies that LR cannot be estimated with the same certainty everywhere. Though the LoR performed better than the KDE when tested using the same dataset, the two LR estimators investigated showed a consistent region in which LR value can be estimated confidently. These two findings advance our understanding of LR when used in computing the strength of evidence in handwriting using forensics.Keywords: Logistic Regression LoR, Kernel Density Estimator KDE, Handwriting, Confidence Interval, Repeatability, Reproducibility.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 470408 Efficiency Enhancement of PWM Controlled Water Electrolysis Cells
Authors: S.K. Mazloomi, Nasri b. Sulaiman
Abstract:
By analyzing the sources of energy and power loss in PWM (Pulse Width Modulation) controlled drivers of water electrolysis cells, it is possible to reduce the power dissipation and enhance the efficiency of such hydrogen production units. A PWM controlled power driver is based on a semiconductor switching element where its power dissipation might be a remarkable fraction of the total power demand of an electrolysis system. Power dissipation in a semiconductor switching element is related to many different parameters which could be fitted into two main categories: switching losses and conduction losses. Conduction losses are directly related to the built, structure and capabilities of a switching device itself and indeed the conditions in which the element is handling the switching application such as voltage, current, temperature and of course the fabrication technology. On the other hand, switching losses have some other influencing variables other than the mentioned such as control system, switching method and power electronics circuitry of the PWM power driver. By analyzings the characteristics of recently developed power switching transistors from different families of Bipolar Junction Transistors (BJT), Metal Oxide Semiconductor Field Effect Transistors (MOSFET) and Insulated Gate Bipolar Transistors (IGBT), some recommendations are made in this paper which are able to lead to achieve higher hydrogen production efficiency by utilizing PWM controlled water electrolysis cells.Keywords: Power switch, PWM, Semiconductor switch, Waterelectrolysis
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3471407 Flexible Wormhole-Switched Network-on-chip with Two-Level Priority Data Delivery Service
Authors: Faizal A. Samman, Thomas Hollstein, Manfred Glesner
Abstract:
A synchronous network-on-chip using wormhole packet switching and supporting guaranteed-completion best-effort with low-priority (LP) and high-priority (HP) wormhole packet delivery service is presented in this paper. Both our proposed LP and HP message services deliver a good quality of service in term of lossless packet completion and in-order message data delivery. However, the LP message service does not guarantee minimal completion bound. The HP packets will absolutely use 100% bandwidth of their reserved links if the HP packets are injected from the source node with maximum injection. Hence, the service are suitable for small size messages (less than hundred bytes). Otherwise the other HP and LP messages, which require also the links, will experience relatively high latency depending on the size of the HP message. The LP packets are routed using a minimal adaptive routing, while the HP packets are routed using a non-minimal adaptive routing algorithm. Therefore, an additional 3-bit field, identifying the packet type, is introduced in their packet headers to classify and to determine the type of service committed to the packet. Our NoC prototypes have been also synthesized using a 180-nm CMOS standard-cell technology to evaluate the cost of implementing the combination of both services.Keywords: Network-on-Chip, Parallel Pipeline Router Architecture, Wormhole Switching, Two-Level Priority Service.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1765406 110 MW Geothermal Power Plant Multiple Simulator, Using Wireless Technology
Authors: Guillermo Romero-Jiménez, Luis A. Jiménez-Fraustro, Mayolo Salinas-Camacho, Heriberto Avalos-Valenzuela
Abstract:
A geothermal power plant multiple simulator for operators training is presented. The simulator is designed to be installed in a wireless local area network and has a capacity to train one to six operators simultaneously, each one with an independent simulation session. The sessions must be supervised only by one instructor. The main parts of this multiple simulator are: instructor and operator-s stations. On the instructor station, the instructor controls the simulation sessions, establishes training exercises and supervises each power plant operator in individual way. This station is hosted in a Main Personal Computer (NS) and its main functions are: to set initial conditions, snapshots, malfunctions or faults, monitoring trends, and process and soft-panel diagrams. On the other hand the operators carry out their actions over the power plant simulated on the operator-s stations; each one is also hosted in a PC. The main software of instructor and operator-s stations are executed on the same NS and displayed in PCs through graphical Interactive Process Diagrams (IDP). The geothermal multiple simulator has been installed in the Geothermal Simulation Training Center (GSTC) of the Comisi├│n Federal de Electricidad, (Federal Commission of Electricity, CFE), Mexico, and is being utilized as a part of the training courses for geothermal power plant operators.Keywords: Geothermal power plant, multiple simulator, training operator.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2047405 The Calculation of Electromagnetic Fields (EMF) in Substations of Shopping Centers
Authors: Adnan Muharemovic, Hidajet Salkic, Mario Klaric, Irfan Turkovic, Aida Muharemovic
Abstract:
In nature, electromagnetic fields always appear like atmosphere static electric field, the earth's static magnetic field and the wide-rang frequency electromagnetic field caused by lightening. However, besides natural electromagnetic fields (EMF), today human beings are mostly exposed to artificial electromagnetic fields due to technology progress and outspread use of electrical devices. To evaluate nuisance of EMF, it is necessary to know field intensity for every frequency which appears and compare it with allowed values. Low frequency EMF-s around transmission and distribution lines are time-varying quasi-static electromagnetic fields which have conservative component of low frequency electrical field caused by charges and eddy component of low frequency magnetic field caused by currents. Displacement current or field delay are negligible, so energy flow in quasi-static EMF involves diffusion, analog like heat transfer. Electrical and magnetic field can be analyzed separately. This paper analysis the numerical calculations in ELF-400 software of EMF in distribution substation in shopping center. Analyzing the results it is possible to specify locations exposed to the fields and give useful suggestion to eliminate electromagnetic effect or reduce it on acceptable level within the non-ionizing radiation norms and norms of protection from EMF.Keywords: Electromagnetic Field, Density of Electromagnetic Flow, Place of Proffesional Exposure, Place of Increased Sensitivity
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3857404 Experimental Study for the Development of a Wireless Communication System in a Solar Central Tower Facility
Authors: Victor H. Benitez, Ramon V. Armas-Flores, Jesus H. Pacheco-Ramirez
Abstract:
Systems transforming solar energy into electrical power have emerged as a viable source of clean, renewable energy. Solar power tower technology is a good example of this type of system, which consists of several mobile mirrors, called heliostats, which reflect the sun's radiation to the same point, located on top of a tower at the center of heliostat field, for collection or transformation into another type of energy. The so-called Hermosillo’s Solar Platform (Plataforma Solar de Hermosillo, PSH, in Spanish) is a facility constituted with several heliostats, its aim and scope is for research purposes. In this paper, the implementation of a wireless communication system based on intelligent nodes is proposed in order to allow the communication and control of the heliostats in PSH. Intelligent nodes transmit information from one point to another, and can perform other actions that allow them to adapt to the conditions and limitations of a field of heliostats, thus achieving effective communication system. After deployment of the nodes in the heliostats, tests were conducted to measure the effectiveness of the communication, and determine the feasibility of using the proposed technologies. The test results were always positive, exceeding expectations held for its operation in the field of heliostats. Therefore, it was possible to validate the efficiency of the wireless communication system to be implemented in PSH, allowing communication and control of the heliostats.Keywords: Solar energy, heliostat, wireless communication, intelligent node.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1666403 Carbothermic Reduction of Phosphoric Acid Extracted from Dephosphorization Slags to Produce Yellow Phosphorus
Authors: Ryoko Yoshida, Jyunpei Yoshida, Hua Fang Yu, Yasushi Sasaki, Tetsuya Nagasaka
Abstract:
Phosphorous is an important element for agriculture and industry and is a non-renewable resource. Especially, yellow phosphorus is an essential material in advanced industrial technology, but phosphorus resources were not produced in Japan at all, and all depend on imports. It has been suggested, however, that the remaining accessible reserves of phosphate ore will be depleted within 50 years. Therefore, alternative resources for phosphate ore must be found. In this research, we have developed a process that enables the production of high-purity yellow phosphorus from domestic unused phosphorus resources such as steelmaking slags. The process consists of two parts: (1) the production of crude phosphoric acid from wastes such as steelmaking slag; (2) producing high-purity yellow phosphorus by low-temperature carbothermic reduction of phosphoric acid (H3PO4). The details of the carbothermic reduction of phosphoric acid are presented in this paper. Yellow phosphorus is commercially produced by carbothermic reduction of phosphate ore in an electric arc furnace at more than 1673K. In the newly developed system, gaseous P4O10 evaporated from H3PO4 is successfully reduced to yellow phosphorus by using carbon packed bed at less than 1273K. To meet the depletion of phosphate ore, the proposed process in this study to produce yellow phosphorus by carbothermic reduction of H3PO4 that are extracted from dephosphorization slags will be one of the effective and economical solutions.
Keywords: Carbothermic reduction, dephosphorization slags, phosphoric acid, yellow phosphorus.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 991402 Statistical Measures and Optimization Algorithms for Gene Selection in Lung and Ovarian Tumor
Authors: C. Gunavathi, K. Premalatha
Abstract:
Microarray technology is universally used in the study of disease diagnosis using gene expression levels. The main shortcoming of gene expression data is that it includes thousands of genes and a small number of samples. Abundant methods and techniques have been proposed for tumor classification using microarray gene expression data. Feature or gene selection methods can be used to mine the genes that directly involve in the classification and to eliminate irrelevant genes. In this paper statistical measures like T-Statistics, Signal-to-Noise Ratio (SNR) and F-Statistics are used to rank the genes. The ranked genes are used for further classification. Particle Swarm Optimization (PSO) algorithm and Shuffled Frog Leaping (SFL) algorithm are used to find the significant genes from the top-m ranked genes. The Naïve Bayes Classifier (NBC) is used to classify the samples based on the significant genes. The proposed work is applied on Lung and Ovarian datasets. The experimental results show that the proposed method achieves 100% accuracy in all the three datasets and the results are compared with previous works.
Keywords: Microarray, T-Statistics, Signal-to-Noise Ratio, FStatistics, Particle Swarm Optimization, Shuffled Frog Leaping, Naïve Bayes Classifier.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1944401 Nano-Bioremediation of Contaminated Industrial Wastewater Using Biosynthesized AgNPs and Their Nano-Composite
Authors: Osama M. Darwesh, Sahar H. Hassan, Abd El-Raheem R. El-Shanshoury, Shawky Z. Sabae
Abstract:
Nanotechnology as multidisciplinary technology is growing rapidly with important applications in several sectors. Also, nanobiotechnology is known for the use of microorganisms for the synthesis of targeted nanoparticles. The present study deals with the green synthesis of silver nanoparticles using aquatic bacteria and the development of a biogenic nanocomposite for environmental applications. 20 morphologically different colonies were isolated from the collected water samples from eight different locations at the Rosetta branch of the Nile Delta, Egypt. The obtained results illustrated that the most effective bacterial isolate (produced the higher amount of AgNPs after 24 h of incubation time) is isolate R3. Bacillus tequilensis was the strongest extracellular bio-manufactory of AgNPs. Biosynthesized nanoparticles had a spherical shape with a mean diameter of 2.74 to 28.4 nm. The antimicrobial activity of silver nanoparticles against many pathogenic microbes indicated that the produced AgNPs had high activity against all tested multi-antibiotic resistant pathogens. Also, the stabilized prepared AgNPs-SA nanocomposite has greater catalytic activity for the decolourization of some dyes like Methylene blue (MB) and Crystal violet. Such results represent a promising stage for producing eco-friendly, cost-effective, and easy-to-handle devices for the bioremediation of contaminated industrial wastewater.
Keywords: Bioremediation, AgNPs, AgNPs-SA nanocomposite, Bacillus tequilensis, nanobiotechnology.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 360400 Analysis of Secondary School Students’ Perceptions about Information Technologies through a Word Association Test
Authors: Fetah Eren, Ismail Sahin, Ismail Celik, Ahmet Oguz Akturk
Abstract:
The aim of this study is to discover secondary school students’ perceptions related to information technologies and the connections between concepts in their cognitive structures. A word association test consisting of six concepts related to information technologies is used to collect data from 244 secondary school students. Concept maps that present students’ cognitive structures are drawn with the help of frequency data. Data are analyzed and interpreted according to the connections obtained as a result of the concept maps. It is determined students associate most with these concepts—computer, Internet, and communication of the given concepts, and associate least with these concepts—computer-assisted education and information technologies. These results show the concepts, Internet, communication, and computer, are an important part of students’ cognitive structures. In addition, students mostly answer computer, phone, game, Internet and Facebook as the key concepts. These answers show students regard information technologies as a means for entertainment and free time activity, not as a means for education.
Keywords: Word association test, cognitive structure, information technology, secondary school.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2078399 On Innovation and Knowledge Economy in Russia
Authors: Zhanna Mingaleva, Irina Mirskikh
Abstract:
Innovational development of regions in Russia is generally faced with the essential influence from federal and local authorities. The organization of effective mechanism of innovation development (and self-development) is impossible without establishment of defined institutional conditions in the analyzed field. Creative utilization of scientific concepts and information should merge, giving rise to continuing innovation and advanced production. The paper presents an analysis of institutional conditions in the field of creation and development of innovation activity infrastructure and transferring of knowledge and skills between different economic agents in Russia. Knowledge is mainly privately owned, developed through R&D investments and incorporated into technology or a product. Innovation infrastructure is a strong concentration mechanism of advanced facilities, which are mainly located inside large agglomerations or city-regions in order to benefit from scale effects in both input markets (human capital, private financial capital) and output markets (higher education services, research services). The empirical results of the paper show that in the presence of more efficient innovation and knowledge transfer and transcoding system and of a more open attitude of economic agents towards innovation, the innovation and knowledge capacity of regional economy is much higher.
Keywords: knowledge economy, innovational development, transfer of knowledge, institutional preconditions, innovation andknowledge capacity.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2278398 Artificial Neural Networks Technique for Seismic Hazard Prediction Using Seismic Bumps
Authors: Belkacem Selma, Boumediene Selma, Samira Chouraqui, Hanifi Missoum, Tourkia Guerzou
Abstract:
Natural disasters have occurred and will continue to cause human and material damage. Therefore, the idea of "preventing" natural disasters will never be possible. However, their prediction is possible with the advancement of technology. Even if natural disasters are effectively inevitable, their consequences may be partly controlled. The rapid growth and progress of artificial intelligence (AI) had a major impact on the prediction of natural disasters and risk assessment which are necessary for effective disaster reduction. Earthquake prediction to prevent the loss of human lives and even property damage is an important factor; that, is why it is crucial to develop techniques for predicting this natural disaster. This study aims to analyze the ability of artificial neural networks (ANNs) to predict earthquakes that occur in a given area. The used data describe the problem of high energy (higher than 104 J) seismic bumps forecasting in a coal mine using two long walls as an example. For this purpose, seismic bumps data obtained from mines have been analyzed. The results obtained show that the ANN is able to predict earthquake parameters with high accuracy; the classification accuracy through neural networks is more than 94%, and the models developed are efficient and robust and depend only weakly on the initial database.
Keywords: Earthquake prediction, artificial intelligence, AI, Artificial Neural Network, ANN, seismic bumps.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1184397 An Extended Domain-Specific Modeling Language for Marine Observatory Relying on Enterprise Architecture
Authors: Charbel Geryes Aoun, Loic Lagadec
Abstract:
A Sensor Network (SN) is considered as an operation of two phases: (1) the observation/measuring, which means the accumulation of the gathered data at each sensor node; (2) transferring the collected data to some processing center (e.g. Fusion Servers) within the SN. Therefore, an underwater sensor network can be defined as a sensor network deployed underwater that monitors underwater activity. The deployed sensors, such as hydrophones, are responsible for registering underwater activity and transferring it to more advanced components. The process of data exchange between the aforementioned components perfectly defines the Marine Observatory (MO) concept which provides information on ocean state, phenomena and processes. The first step towards the implementation of this concept is defining the environmental constraints and the required tools and components (Marine Cables, Smart Sensors, Data Fusion Server, etc). The logical and physical components that are used in these observatories perform some critical functions such as the localization of underwater moving objects. These functions can be orchestrated with other services (e.g. military or civilian reaction). In this paper, we present an extension to our MO meta-model that is used to generate a design tool (ArchiMO). We propose constraints to be taken into consideration at design time. We illustrate our proposal with an example from the MO domain. Additionally, we generate the corresponding simulation code using our self-developed domain-specific model compiler. On the one hand, this illustrates our approach in relying on Enterprise Architecture (EA) framework that respects: multiple-views, perspectives of stakeholders, and domain specificity. On the other hand, it helps reducing both complexity and time spent in design activity, while preventing from design modeling errors during porting this activity in the MO domain. As conclusion, this work aims to demonstrate that we can improve the design activity of complex system based on the use of MDE technologies and a domain-specific modeling language with the associated tooling. The major improvement is to provide an early validation step via models and simulation approach to consolidate the system design.
Keywords: Smart sensors, data fusion, distributed fusion architecture, sensor networks, domain specific modeling language, enterprise architecture, underwater moving object, localization, marine observatory, NS-3, IMS.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 255396 Development of Performance Indicators in Operational Level for Pre-hospital EMS in Thailand
Authors: Napisporn Memongkol, Runchana Sinthavalai, Nattapong Seneeratanaprayune Weerawat Ounsaneha, Chanisada Choosuk
Abstract:
The objective of this research is to develop the performance indicators (PIs) in operational level for the Pre-hospital Emergency Medical Service (EMS) system employing in Thailand. This research started with ascertaining the current pre-hospital care system. The team analyzed the strategies of Narerthorn, a government unit under the ministry of public health, and the existing PIs of the pre-hospital care. Afterwards, the current National Strategic Plan of EMS development (2008-2012) of the Emergency Medical Institute of Thailand (EMIT) was considered using strategic analysis to developed Strategy Map (SM) and identified the Success Factors (SFs). The analysis results from strategy map and SFs were used to develop the Performance Indicators (PIs). To verify the set of PIs, the team has interviewed with the relevant practitioners for the possibilities to implement the PIs. To this paper, it was to ascertain that all the developed PIs support the objectives of the strategic plan. Nevertheless, the results showed that the operational level PIs suited only with the first dimension of National Strategic Plan (infrastructure and information technology development). Besides, the SF was the infrastructure development (to contribute the EMS system to people throughout with standard and efficiency both in normally and disaster conditions). Finally, twenty-nine indicators were developed from the analysis results of SM and SFs.Keywords: Emergency Medical Service, Performance Indicator, Success Factor, Thailand
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2271395 Towards End-To-End Disease Prediction from Raw Metagenomic Data
Authors: Maxence Queyrel, Edi Prifti, Alexandre Templier, Jean-Daniel Zucker
Abstract:
Analysis of the human microbiome using metagenomic sequencing data has demonstrated high ability in discriminating various human diseases. Raw metagenomic sequencing data require multiple complex and computationally heavy bioinformatics steps prior to data analysis. Such data contain millions of short sequences read from the fragmented DNA sequences and stored as fastq files. Conventional processing pipelines consist in multiple steps including quality control, filtering, alignment of sequences against genomic catalogs (genes, species, taxonomic levels, functional pathways, etc.). These pipelines are complex to use, time consuming and rely on a large number of parameters that often provide variability and impact the estimation of the microbiome elements. Training Deep Neural Networks directly from raw sequencing data is a promising approach to bypass some of the challenges associated with mainstream bioinformatics pipelines. Most of these methods use the concept of word and sentence embeddings that create a meaningful and numerical representation of DNA sequences, while extracting features and reducing the dimensionality of the data. In this paper we present an end-to-end approach that classifies patients into disease groups directly from raw metagenomic reads: metagenome2vec. This approach is composed of four steps (i) generating a vocabulary of k-mers and learning their numerical embeddings; (ii) learning DNA sequence (read) embeddings; (iii) identifying the genome from which the sequence is most likely to come and (iv) training a multiple instance learning classifier which predicts the phenotype based on the vector representation of the raw data. An attention mechanism is applied in the network so that the model can be interpreted, assigning a weight to the influence of the prediction for each genome. Using two public real-life data-sets as well a simulated one, we demonstrated that this original approach reaches high performance, comparable with the state-of-the-art methods applied directly on processed data though mainstream bioinformatics workflows. These results are encouraging for this proof of concept work. We believe that with further dedication, the DNN models have the potential to surpass mainstream bioinformatics workflows in disease classification tasks.Keywords: Metagenomics, phenotype prediction, deep learning, embeddings, multiple instance learning.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 909394 Leveraging Hyperledger Iroha for the Issuance and Verification of Higher-Education Certificates
Authors: Vasiliki Vlachou, Christos Kontzinos, Ourania Markaki, Panagiotis Kokkinakos, Vagelis Karakolis, John Psarras
Abstract:
Higher Education is resisting the pull of technology, especially as this concerns the issuance and verification of degrees and certificates. It is widely known that education certificates are largely produced in paper form making them vulnerable to damage while holders of such certificates are dependent on the universities and other issuing organisations. QualiChain is an EU Horizon 2020 (H2020) research project aiming to transform and revolutionise the domain of public education and its ties with the job market by leveraging blockchain, analytics and decision support to develop a platform for the verification and sharing of education certificates. Blockchain plays an integral part in the QualiChain solution in providing a trustworthy environment to store, share and manage such accreditations. Under the context of this paper, three prominent blockchain platforms (Ethereum, Hyperledger Fabric, Hyperledger Iroha) were considered as a means of experimentation for creating a system with the basic functionalities that will be needed for trustworthy degree verification. The methodology and respective system developed and presented in this paper used Hyperledger Iroha and proved that this specific platform can be used to easily develop decentralize applications. Future papers will attempt to further experiment with other blockchain platforms and assess which has the best potential.
Keywords: Blockchain, degree verification, higher education certificates, Hyperledger Iroha.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 828393 Requirements Driven Multiple View Paradigm for Developing Security Architecture
Authors: K. Chandra Sekaran
Abstract:
This paper describes a paradigmatic approach to develop architecture of secure systems by describing the requirements from four different points of view: that of the owner, the administrator, the user, and the network. Deriving requirements and developing architecture implies the joint elicitation and describing the problem and the structure of the solution. The view points proposed in this paper are those we consider as requirements towards their contributions as major parties in the design, implementation, usage and maintenance of secure systems. The dramatic growth of the technology of Internet and the applications deployed in World Wide Web have lead to the situation where the security has become a very important concern in the development of secure systems. Many security approaches are currently being used in organizations. In spite of the widespread use of many different security solutions, the security remains a problem. It is argued that the approach that is described in this paper for the development of secure architecture is practical by all means. The models representing these multiple points of view are termed the requirements model (views of owner and administrator) and the operations model (views of user and network). In this paper, this multiple view paradigm is explained by first describing the specific requirements and or characteristics of secure systems (particularly in the domain of networks) and the secure architecture / system development methodology.
Keywords: Multiple view paradigms, requirements model, operations model, secure system, owner, administrator, user, network.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1368392 Applications of Building Information Modeling (BIM) in Knowledge Sharing and Management in Construction
Authors: Shu-Hui Jan, Shih-Ping Ho, Hui-Ping Tserng
Abstract:
Construction knowledge can be referred to and reused among involved project managers and jobsite engineers to alleviate problems on a construction jobsite and reduce the time and cost of solving problems related to constructability. This paper proposes a new methodology to provide sharing of construction knowledge by using the Building Information Modeling (BIM) approach. The main characteristics of BIM include illustrating 3D CAD-based presentations and keeping information in a digital format, and facilitation of easy updating and transfer of information in the 3D BIM environment. Using the BIM approach, project managers and engineers can gain knowledge related to 3D BIM and obtain feedback provided by jobsite engineers for future reference. This study addresses the application of knowledge sharing management in the construction phase of construction projects and proposes a BIM-based Knowledge Sharing Management (BIMKSM) system for project managers and engineers. The BIMKSM system is then applied in a selected case study of a construction project in Taiwan to verify the proposed methodology and demonstrate the effectiveness of sharing knowledge in the BIM environment. The combined results demonstrate that the BIMKSM system can be used as a visual BIM-based knowledge sharing management platform by utilizing the BIM approach and web technology.
Keywords: Construction knowledge management, building information modeling, project management, web-based information system.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 4346391 Design and Fabrication of Stent with Negative Poisson’s Ratio
Authors: S. K. Bhullar, J. Ko, F. Ahmed, M. B. G. Jun
Abstract:
The negative Poisson’s ratios can be described in terms of models based on the geometry of the system and the way this geometry changes due to applied loads. As the Poisson’s ratio does not depend on scale hence deformation can take place at the nano to macro level the only requirement is the right combination of the geometry. Our thrust in this paper is to combine our knowledge of tailored enhanced mechanical properties of the materials having negative Poisson’s ratio with the micromachining and electrospining technology to develop a novel stent carrying a drug delivery system. Therefore, the objective of this paper includes (i) fabrication of a micromachined metal sheet tailored with structure having negative Poisson’s ratio through rotating solid squares geometry using femtosecond laser ablation; (ii) rolling fabricated structure and welding to make a tubular structure (iii) wrapping it with nanofibers of biocompatible polymer PCL (polycaprolactone) for drug delivery (iv) analysis of the functional and mechanical performance of fabricated structure analytically and experimentally. Further, as the applications concerned, tubular structures have potential in biomedical for example hollow tubes called stents are placed inside to provide mechanical support to a damaged artery or diseased region and to open a blocked esophagus thus allowing feeding capacity and improving quality of life.
Keywords: Micromachining, electrospining, auxetic materials, enhanced mechanical properties.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3687390 Unbalanced Distribution Optimal Power Flow to Minimize Losses with Distributed Photovoltaic Plants
Authors: Malinwo Estone Ayikpa
Abstract:
Electric power systems are likely to operate with minimum losses and voltage meeting international standards. This is made possible generally by control actions provide by automatic voltage regulators, capacitors and transformers with on-load tap changer (OLTC). With the development of photovoltaic (PV) systems technology, their integration on distribution networks has increased over the last years to the extent of replacing the above mentioned techniques. The conventional analysis and simulation tools used for electrical networks are no longer able to take into account control actions necessary for studying distributed PV generation impact. This paper presents an unbalanced optimal power flow (OPF) model that minimizes losses with association of active power generation and reactive power control of single-phase and three-phase PV systems. Reactive power can be generated or absorbed using the available capacity and the adjustable power factor of the inverter. The unbalance OPF is formulated by current balance equations and solved by primal-dual interior point method. Several simulation cases have been carried out varying the size and location of PV systems and the results show a detailed view of the impact of PV distributed generation on distribution systems.
Keywords: Distribution system, losses, photovoltaic generation, primal-dual interior point method, reactive power control.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1079