Search results for: low-temperature district heating network
4778 Cascaded Neural Network for Internal Temperature Forecasting in Induction Motor
Authors: Hidir S. Nogay
Abstract:
In this study, two systems were created to predict interior temperature in induction motor. One of them consisted of a simple ANN model which has two layers, ten input parameters and one output parameter. The other one consisted of eight ANN models connected each other as cascaded. Cascaded ANN system has 17 inputs. Main reason of cascaded system being used in this study is to accomplish more accurate estimation by increasing inputs in the ANN system. Cascaded ANN system is compared with simple conventional ANN model to prove mentioned advantages. Dataset was obtained from experimental applications. Small part of the dataset was used to obtain more understandable graphs. Number of data is 329. 30% of the data was used for testing and validation. Test data and validation data were determined for each ANN model separately and reliability of each model was tested. As a result of this study, it has been understood that the cascaded ANN system produced more accurate estimates than conventional ANN model.Keywords: cascaded neural network, internal temperature, inverter, three-phase induction motor
Procedia PDF Downloads 3454777 Genome-Wide Expression Profiling of Cicer arietinum Heavy Metal Toxicity
Authors: B. S. Yadav, A. Mani, S. Srivastava
Abstract:
Chickpea (Cicer arietinum L.) is an annual, self-pollinating, diploid (2n = 2x = 16) pulse crop that ranks second in world legume production after common bean (Phaseolus vulgaris). ICC 4958 flowers approximately 39 days after sowing under peninsular Indian conditions and the crop matures in less than 90 days in rained environments. The estimated collective yield losses due to abiotic stresses (6.4 million t) have been significantly higher than for biotic stresses (4.8 million t). Most legumes are known to be salt sensitive, and therefore, it is becoming increasingly important to produce cultivars tolerant to high-salinity in addition to other abiotic and biotic stresses for sustainable chickpea production. Our aim was to identify the genes that are involved in the defence mechanism against heavy metal toxicity in chickpea and establish the biological network of heavy metal toxicity in chickpea. ICC4958 variety of chick pea was taken and grown in normal condition and 150µM concentration of different heavy metal salt like CdCl₂, K₂Cr2O₇, NaAsO₂. At 15th day leave samples were collected and stored in RNA Later solution microarray was performed for checking out differential gene expression pattern. Our studies revealed that 111 common genes that involved in defense mechanism were up regulated and 41 genes were commonly down regulated during treatment of 150µM concentration of CdCl₂, K₂Cr₂O₇, and NaAsO₂. Biological network study shows that the genes which are differentially expressed are highly connected and having high betweenness and centrality.Keywords: abiotic stress, biological network, chickpea, microarray
Procedia PDF Downloads 1974776 An Algorithm for Determining the Arrival Behavior of a Secondary User to a Base Station in Cognitive Radio Networks
Authors: Danilo López, Edwin Rivas, Leyla López
Abstract:
This paper presents the development of an algorithm that predicts the arrival of a secondary user (SU) to a base station (BS) in a cognitive network based on infrastructure, requesting a Best Effort (BE) or Real Time (RT) type of service with a determined bandwidth (BW) implementing neural networks. The algorithm dynamically uses a neural network construction technique using the geometric pyramid topology and trains a Multilayer Perceptron Neural Networks (MLPNN) based on the historical arrival of an SU to estimate future applications. This will allow efficiently managing the information in the BS, since it precedes the arrival of the SUs in the stage of selection of the best channel in CRN. As a result, the software application determines the probability of arrival at a future time point and calculates the performance metrics to measure the effectiveness of the predictions made.Keywords: cognitive radio, base station, best effort, MLPNN, prediction, real time
Procedia PDF Downloads 3314775 Attention-based Adaptive Convolution with Progressive Learning in Speech Enhancement
Authors: Tian Lan, Yixiang Wang, Wenxin Tai, Yilan Lyu, Zufeng Wu
Abstract:
The monaural speech enhancement task in the time-frequencydomain has a myriad of approaches, with the stacked con-volutional neural network (CNN) demonstrating superiorability in feature extraction and selection. However, usingstacked single convolutions method limits feature represen-tation capability and generalization ability. In order to solvethe aforementioned problem, we propose an attention-basedadaptive convolutional network that integrates the multi-scale convolutional operations into a operation-specific blockvia input dependent attention to adapt to complex auditoryscenes. In addition, we introduce a two-stage progressivelearning method to enlarge the receptive field without a dra-matic increase in computation burden. We conduct a series ofexperiments based on the TIMIT corpus, and the experimen-tal results prove that our proposed model is better than thestate-of-art models on all metrics.Keywords: speech enhancement, adaptive convolu-tion, progressive learning, time-frequency domain
Procedia PDF Downloads 1224774 Estimation of Endogenous Brain Noise from Brain Response to Flickering Visual Stimulation Magnetoencephalography Visual Perception Speed
Authors: Alexander N. Pisarchik, Parth Chholak
Abstract:
Intrinsic brain noise was estimated via magneto-encephalograms (MEG) recorded during perception of flickering visual stimuli with frequencies of 6.67 and 8.57 Hz. First, we measured the mean phase difference between the flicker signal and steady-state event-related field (SSERF) in the occipital area where the brain response at the flicker frequencies and their harmonics appeared in the power spectrum. Then, we calculated the probability distribution of the phase fluctuations in the regions of frequency locking and computed its kurtosis. Since kurtosis is a measure of the distribution’s sharpness, we suppose that inverse kurtosis is related to intrinsic brain noise. In our experiments, the kurtosis value varied among subjects from K = 3 to K = 5 for 6.67 Hz and from 2.6 to 4 for 8.57 Hz. The majority of subjects demonstrated leptokurtic kurtosis (K < 3), i.e., the distribution tails approached zero more slowly than Gaussian. In addition, we found a strong correlation between kurtosis and brain complexity measured as the correlation dimension, so that the MEGs of subjects with higher kurtosis exhibited lower complexity. The obtained results are discussed in the framework of nonlinear dynamics and complex network theories. Specifically, in a network of coupled oscillators, phase synchronization is mainly determined by two antagonistic factors, noise, and the coupling strength. While noise worsens phase synchronization, the coupling improves it. If we assume that each neuron and each synapse contribute to brain noise, the larger neuronal network should have stronger noise, and therefore phase synchronization should be worse, that results in smaller kurtosis. The described method for brain noise estimation can be useful for diagnostics of some brain pathologies associated with abnormal brain noise.Keywords: brain, flickering, magnetoencephalography, MEG, visual perception, perception time
Procedia PDF Downloads 1484773 The Friendship Network Stability of Preschool Children during One Pedagogical Season
Authors: Yili Wang, Jarmo Kinos, Tuire Palonen, Tarja-Riitta Hurme
Abstract:
This longitudinal study aims to examine how five- and six-year-old children’s peer relationships are formed and fostered during one preschool year in a southwestern Finnish preschool. All 16 kindergarteners participated in the study (at dyad level N=240; i.e., 16 x 15 relationships among the children). The children were divided into four daily groups, based on the table order during the daily routines, and four intervention groups, based on the teachers’ pedagogical plan. During the intervention, one iPad was given to each group in order to stimulate interaction among peers and, thus, enable the children to form new peer relationships. In the data gathering, sociometric nomination techniques were used to investigate the nature (i.e., stability and mutuality) of the peer relationships. The data was collected five times during the year to see what kind of peer relationship changes occurred at the dyad level and the group level, i.e., in establishing and losing friendship ties among the children. Social network analyses were used to analyze the data. The results indicate that the children’s preference for gender segregation was strong compared to age preference and intervention. In all, the number of reciprocal friendship ties and the mutual absence of friendship ties increased towards the end of the year, whereas the number of unilateral friendship ties decreased. This indicates that children’s nominations narrow down; thus, the group structure becomes more crystalized. Instead of extending their friendship networks, children seek stable and mutual relationships with their peers in their middle childhood years. The intervention only had a slightly negative influence on children’s peer relationships.Keywords: intervention study, peer relationship, preschool education, social network analysis, sociometric ratings
Procedia PDF Downloads 2734772 Generalized Rough Sets Applied to Graphs Related to Urban Problems
Authors: Mihai Rebenciuc, Simona Mihaela Bibic
Abstract:
Branch of modern mathematics, graphs represent instruments for optimization and solving practical applications in various fields such as economic networks, engineering, network optimization, the geometry of social action, generally, complex systems including contemporary urban problems (path or transport efficiencies, biourbanism, & c.). In this paper is studied the interconnection of some urban network, which can lead to a simulation problem of a digraph through another digraph. The simulation is made univoc or more general multivoc. The concepts of fragment and atom are very useful in the study of connectivity in the digraph that is simulation - including an alternative evaluation of k- connectivity. Rough set approach in (bi)digraph which is proposed in premier in this paper contribute to improved significantly the evaluation of k-connectivity. This rough set approach is based on generalized rough sets - basic facts are presented in this paper.Keywords: (bi)digraphs, rough set theory, systems of interacting agents, complex systems
Procedia PDF Downloads 2434771 Fake Accounts Detection in Twitter Based on Minimum Weighted Feature Set
Authors: Ahmed ElAzab, Amira M. Idrees, Mahmoud A. Mahmoud, Hesham Hefny
Abstract:
Social networking sites such as Twitter and Facebook attracts over 500 million users across the world, for those users, their social life, even their practical life, has become interrelated. Their interaction with social networking has affected their life forever. Accordingly, social networking sites have become among the main channels that are responsible for vast dissemination of different kinds of information during real time events. This popularity in Social networking has led to different problems including the possibility of exposing incorrect information to their users through fake accounts which results to the spread of malicious content during life events. This situation can result to a huge damage in the real world to the society in general including citizens, business entities, and others. In this paper, we present a classification method for detecting fake accounts on Twitter. The study determines the minimized set of the main factors that influence the detection of the fake accounts on Twitter, then the determined factors have been applied using different classification techniques, a comparison of the results for these techniques has been performed and the most accurate algorithm is selected according to the accuracy of the results. The study has been compared with different recent research in the same area, this comparison has proved the accuracy of the proposed study. We claim that this study can be continuously applied on Twitter social network to automatically detect the fake accounts, moreover, the study can be applied on different Social network sites such as Facebook with minor changes according to the nature of the social network which are discussed in this paper.Keywords: fake accounts detection, classification algorithms, twitter accounts analysis, features based techniques
Procedia PDF Downloads 4164770 Surface Modification of SUS-304 Using Nitriding Treatment for Application of Bipolar Plates of Proton Exchange Membrane Fuel Cells
Authors: Wei-Ru Chang, Jenn-Jiang Hwang, Zen-Ting Hsiao, Shu-Feng Lee
Abstract:
Proton exchange membrane (PEM) fuel cells are widely used in electrical systems as an economical, low-polluting energy source. This study investigates the effects of PEMFC gas nitriding treatment on metal bipolar plates. The test material was SUS304 stainless steel. The study explored five different pretreatment processes, varying the corrosion resistance and electrical conductivity conditions. The most effective process was industrial acid washing, followed by heating to 500 °C. Under the condition, the corrosion current density was 8.695 μA, significantly lower than that of the untreated pretreatment sample flakes, which was measured as 38.351 μA.Keywords: nitriding, bipolar, 304, corrosion, resistance, pretreatment
Procedia PDF Downloads 10874769 Scheduling in a Single-Stage, Multi-Item Compatible Process Using Multiple Arc Network Model
Authors: Bokkasam Sasidhar, Ibrahim Aljasser
Abstract:
The problem of finding optimal schedules for each equipment in a production process is considered, which consists of a single stage of manufacturing and which can handle different types of products, where changeover for handling one type of product to the other type incurs certain costs. The machine capacity is determined by the upper limit for the quantity that can be processed for each of the products in a set up. The changeover costs increase with the number of set ups and hence to minimize the costs associated with the product changeover, the planning should be such that similar types of products should be processed successively so that the total number of changeovers and in turn the associated set up costs are minimized. The problem of cost minimization is equivalent to the problem of minimizing the number of set ups or equivalently maximizing the capacity utilization in between every set up or maximizing the total capacity utilization. Further, the production is usually planned against customers’ orders, and generally different customers’ orders are assigned one of the two priorities – “normal” or “priority” order. The problem of production planning in such a situation can be formulated into a Multiple Arc Network (MAN) model and can be solved sequentially using the algorithm for maximizing flow along a MAN and the algorithm for maximizing flow along a MAN with priority arcs. The model aims to provide optimal production schedule with an objective of maximizing capacity utilization, so that the customer-wise delivery schedules are fulfilled, keeping in view the customer priorities. Algorithms have been presented for solving the MAN formulation of the production planning with customer priorities. The application of the model is demonstrated through numerical examples.Keywords: scheduling, maximal flow problem, multiple arc network model, optimization
Procedia PDF Downloads 4024768 Incorporating Lexical-Semantic Knowledge into Convolutional Neural Network Framework for Pediatric Disease Diagnosis
Authors: Xiaocong Liu, Huazhen Wang, Ting He, Xiaozheng Li, Weihan Zhang, Jian Chen
Abstract:
The utilization of electronic medical record (EMR) data to establish the disease diagnosis model has become an important research content of biomedical informatics. Deep learning can automatically extract features from the massive data, which brings about breakthroughs in the study of EMR data. The challenge is that deep learning lacks semantic knowledge, which leads to impracticability in medical science. This research proposes a method of incorporating lexical-semantic knowledge from abundant entities into a convolutional neural network (CNN) framework for pediatric disease diagnosis. Firstly, medical terms are vectorized into Lexical Semantic Vectors (LSV), which are concatenated with the embedded word vectors of word2vec to enrich the feature representation. Secondly, the semantic distribution of medical terms serves as Semantic Decision Guide (SDG) for the optimization of deep learning models. The study evaluate the performance of LSV-SDG-CNN model on four kinds of Chinese EMR datasets. Additionally, CNN, LSV-CNN, and SDG-CNN are designed as baseline models for comparison. The experimental results show that LSV-SDG-CNN model outperforms baseline models on four kinds of Chinese EMR datasets. The best configuration of the model yielded an F1 score of 86.20%. The results clearly demonstrate that CNN has been effectively guided and optimized by lexical-semantic knowledge, and LSV-SDG-CNN model improves the disease classification accuracy with a clear margin.Keywords: convolutional neural network, electronic medical record, feature representation, lexical semantics, semantic decision
Procedia PDF Downloads 1264767 Network Conditioning and Transfer Learning for Peripheral Nerve Segmentation in Ultrasound Images
Authors: Harold Mauricio Díaz-Vargas, Cristian Alfonso Jimenez-Castaño, David Augusto Cárdenas-Peña, Guillermo Alberto Ortiz-Gómez, Alvaro Angel Orozco-Gutierrez
Abstract:
Precise identification of the nerves is a crucial task performed by anesthesiologists for an effective Peripheral Nerve Blocking (PNB). Now, anesthesiologists use ultrasound imaging equipment to guide the PNB and detect nervous structures. However, visual identification of the nerves from ultrasound images is difficult, even for trained specialists, due to artifacts and low contrast. The recent advances in deep learning make neural networks a potential tool for accurate nerve segmentation systems, so addressing the above issues from raw data. The most widely spread U-Net network yields pixel-by-pixel segmentation by encoding the input image and decoding the attained feature vector into a semantic image. This work proposes a conditioning approach and encoder pre-training to enhance the nerve segmentation of traditional U-Nets. Conditioning is achieved by the one-hot encoding of the kind of target nerve a the network input, while the pre-training considers five well-known deep networks for image classification. The proposed approach is tested in a collection of 619 US images, where the best C-UNet architecture yields an 81% Dice coefficient, outperforming the 74% of the best traditional U-Net. Results prove that pre-trained models with the conditional approach outperform their equivalent baseline by supporting learning new features and enriching the discriminant capability of the tested networks.Keywords: nerve segmentation, U-Net, deep learning, ultrasound imaging, peripheral nerve blocking
Procedia PDF Downloads 1064766 Power Aware Modified I-LEACH Protocol Using Fuzzy IF Then Rules
Authors: Gagandeep Singh, Navdeep Singh
Abstract:
Due to limited battery of sensor nodes, so energy efficiency found to be main constraint in WSN. Therefore the main focus of the present work is to find the ways to minimize the energy consumption problem and will results; enhancement in the network stability period and life time. Many researchers have proposed different kind of the protocols to enhance the network lifetime further. This paper has evaluated the issues which have been neglected in the field of the WSNs. WSNs are composed of multiple unattended ultra-small, limited-power sensor nodes. Sensor nodes are deployed randomly in the area of interest. Sensor nodes have limited processing, wireless communication and power resource capabilities Sensor nodes send sensed data to sink or Base Station (BS). I-LEACH gives adaptive clustering mechanism which very efficiently deals with energy conservations. This paper ends up with the shortcomings of various adaptive clustering based WSNs protocols.Keywords: WSN, I-Leach, MATLAB, sensor
Procedia PDF Downloads 2754765 Microbiological Analysis, Cytotoxic and Genotoxic Effects from Material Captured in PM2.5 and PM10 Filters Used in the Aburrá Valley Air Quality Monitoring Network (Colombia)
Authors: Carmen E. Zapata, Juan Bautista, Olga Montoya, Claudia Moreno, Marisol Suarez, Alejandra Betancur, Duvan Nanclares, Natalia A. Cano
Abstract:
This study aims to evaluate the diversity of microorganisms in filters PM2.5 and PM10; and determine the genotoxic and cytotoxic activity of the complex mixture present in PM2.5 filters used in the Aburrá Valley Air Quality Monitoring Network (Colombia). The research results indicate that particulate matter PM2.5 of different monitoring stations are bacteria; however, this study of detection of bacteria and their phylogenetic relationship is not complete evidence to connect the microorganisms with pathogenic or degrading activities of compounds present in the air. Additionally, it was demonstrated the damage induced by the particulate material in the cell membrane, lysosomal and endosomal membrane and in the mitochondrial metabolism; this damage was independent of the PM2.5 concentrations in almost all the cases.Keywords: cytotoxic, genotoxic, microbiological analysis, PM10, PM2.5
Procedia PDF Downloads 3484764 Assessment of Factors Influencing Adoption of Agroforestry Technologies in Halaba Special Woreda, Southern Ethiopia
Authors: Mihretu Erjabo
Abstract:
Halaba special district is characterized by drought, soil erosion, high population pressure, poor livestock production, lack of feed for livestock, very deep water table, very low productivity of crops and food insufficiency. In order to address these problems, the woreda agricultural development office along with other management practices such as soil physical conservation measures agroforestry was introduced decades ago as a means to alleviate the problem. However, the level of agroforestry adoption remains low. Objective of this study was to identify the factors that influence adoption of agroforestry technologies by farmers in the district. Random sampling was employed to select two kebele administrations and respondents. Data collection was conducted by rural household questionnaire survey, participatory rural appraisal, questionnaires for local and woreda extension staff, secondary data resources and field observation. A sample of 12 key informants, 6 extension staffs, and 182 households, were used in the data collection. Chi square test used to determine significant relationships between adoption of agroforestry and 15 selected variables. Out of which eleven were found to be significant to affect farmers’ adoptiveness. These were frequency of visits of farmers (13.39%), participation in training (11.49%), farmers’ attitude towards agroforestry practices (10.61%), frequency of visits of extensionists (10.38%), participation in extension meeting (10.34%), participation in field day (10.28%), land holding size (9.29%), level of literacy (8.78%), awareness about the importance of agroforestry technology packages (7.06%), time taken from their residence to nearest extension (5.04%) and gender of respondents (3.34%). This study also identified various factors that result in low adoption rates of agroforestry including fear of competition, seedling, rainfall and labour shortage, free grazing, financial problem, expecting trees as soil degrader and long span of trees and lack of need ranking. To improve farmers’ adoption, the factors identified should be well addressed by launching a series and recurrent outreach extension program appropriate and suitable to farmers need.Keywords: farmers attitude, farmers participation, soil degradation, technology packages
Procedia PDF Downloads 1584763 Improving the Performance of Back-Propagation Training Algorithm by Using ANN
Authors: Vishnu Pratap Singh Kirar
Abstract:
Artificial Neural Network (ANN) can be trained using backpropagation (BP). It is the most widely used algorithm for supervised learning with multi-layered feed-forward networks. Efficient learning by the BP algorithm is required for many practical applications. The BP algorithm calculates the weight changes of artificial neural networks, and a common approach is to use a two-term algorithm consisting of a learning rate (LR) and a momentum factor (MF). The major drawbacks of the two-term BP learning algorithm are the problems of local minima and slow convergence speeds, which limit the scope for real-time applications. Recently the addition of an extra term, called a proportional factor (PF), to the two-term BP algorithm was proposed. The third increases the speed of the BP algorithm. However, the PF term also reduces the convergence of the BP algorithm, and criteria for evaluating convergence are required to facilitate the application of the three terms BP algorithm. Although these two seem to be closely related, as described later, we summarize various improvements to overcome the drawbacks. Here we compare the different methods of convergence of the new three-term BP algorithm.Keywords: neural network, backpropagation, local minima, fast convergence rate
Procedia PDF Downloads 4984762 Realization of a Temperature Based Automatic Controlled Domestic Electric Boiling System
Authors: Shengqi Yu, Jinwei Zhao
Abstract:
This paper presents a kind of analog circuit based temperature control system, which is mainly composed by threshold control signal circuit, synchronization signal circuit and trigger pulse circuit. Firstly, the temperature feedback signal function is realized by temperature sensor TS503F3950E. Secondly, the main control circuit forms the cycle controlled pulse signal to control the thyristor switching model. Finally two reverse paralleled thyristors regulate the output power by their switching state. In the consequence, this is a modernized and energy-saving domestic electric heating system.Keywords: time base circuit, automatic control, zero-crossing trigger, temperature control
Procedia PDF Downloads 4814761 Integration of Wireless Sensor Networks and Radio Frequency Identification (RFID): An Assesment
Authors: Arslan Murtaza
Abstract:
RFID (Radio Frequency Identification) and WSN (Wireless sensor network) are two significant wireless technologies that have extensive diversity of applications and provide limitless forthcoming potentials. RFID is used to identify existence and location of objects whereas WSN is used to intellect and monitor the environment. Incorporating RFID with WSN not only provides identity and location of an object but also provides information regarding the condition of the object carrying the sensors enabled RFID tag. It can be widely used in stock management, asset tracking, asset counting, security, military, environmental monitoring and forecasting, healthcare, intelligent home, intelligent transport vehicles, warehouse management, and precision agriculture. This assessment presents a brief introduction of RFID, WSN, and integration of WSN and RFID, and then applications related to both RFID and WSN. This assessment also deliberates status of the projects on RFID technology carried out in different computing group projects to be taken on WSN and RFID technology.Keywords: wireless sensor network, RFID, embedded sensor, Wi-Fi, Bluetooth, integration, time saving, cost efficient
Procedia PDF Downloads 3344760 Message Framework for Disaster Management: An Application Model for Mines
Authors: A. Baloglu, A. Çınar
Abstract:
Different tools and technologies were implemented for Crisis Response and Management (CRM) which is generally using available network infrastructure for information exchange. Depending on type of disaster or crisis, network infrastructure could be affected and it could not be able to provide reliable connectivity. Thus any tool or technology that depends on the connectivity could not be able to fulfill its functionalities. As a solution, a new message exchange framework has been developed. Framework provides offline/online information exchange platform for CRM Information Systems (CRMIS) and it uses XML compression and packet prioritization algorithms and is based on open source web technologies. By introducing offline capabilities to the web technologies, framework will be able to perform message exchange on unreliable networks. The experiments done on the simulation environment provide promising results on low bandwidth networks (56kbps and 28.8 kbps) with up to 50% packet loss and the solution is to successfully transfer all the information on these low quality networks where the traditional 2 and 3 tier applications failed.Keywords: crisis response and management, XML messaging, web services, XML compression, mining
Procedia PDF Downloads 3394759 Transformation of the Traditional Landscape of Kabul Old City: A Study for Its Conservation
Authors: Mohammad Umar Azizi, Tetsuya Ando
Abstract:
This study investigates the transformation of the traditional landscape of Kabul Old City through an examination of five case study areas. Based on physical observation, three types of houses are found: traditional, mixed and modern. Firstly, characteristics of the houses are described according to construction materials and the number of stories. Secondly, internal and external factors are considered in order to implement a conservation plan. Finally, an adaptive conservation plan is suggested to protect the traditional landscape of Kabul Old City.Keywords: conservation, district 1, Kabul Old City, landscape, transformation, traditional houses
Procedia PDF Downloads 2214758 Multilayer Neural Network and Fuzzy Logic Based Software Quality Prediction
Authors: Sadaf Sahar, Usman Qamar, Sadaf Ayaz
Abstract:
In the software development lifecycle, the quality prediction techniques hold a prime importance in order to minimize future design errors and expensive maintenance. There are many techniques proposed by various researchers, but with the increasing complexity of the software lifecycle model, it is crucial to develop a flexible system which can cater for the factors which in result have an impact on the quality of the end product. These factors include properties of the software development process and the product along with its operation conditions. In this paper, a neural network (perceptron) based software quality prediction technique is proposed. Using this technique, the stakeholders can predict the quality of the resulting software during the early phases of the lifecycle saving time and resources on future elimination of design errors and costly maintenance. This technique can be brought into practical use using successful training.Keywords: software quality, fuzzy logic, perception, prediction
Procedia PDF Downloads 3174757 Distribution Planning with Renewable Energy Units Based on Improved Honey Bee Mating Optimization
Authors: Noradin Ghadimi, Nima Amjady, Oveis Abedinia, Roza Poursoleiman
Abstract:
This paper proposed an Improved Honey Bee Mating Optimization (IHBMO) for a planning paradigm for network upgrade. The proposed technique is a new meta-heuristic algorithm which inspired by mating of the honey bee. The paradigm is able to select amongst several choices equi-cost one assuring the optimum in terms of voltage profile, considering various scenarios of DG penetration and load demand. The distributed generation (DG) has created a challenge and an opportunity for developing various novel technologies in power generation. DG prepares a multitude of services to utilities and consumers, containing standby generation, peaks chopping sufficiency, base load generation. The proposed algorithm is applied over the 30 lines, 28 buses power system. The achieved results demonstrate the good efficiency of the DG using the proposed technique in different scenarios.Keywords: distributed generation, IHBMO, renewable energy units, network upgrade
Procedia PDF Downloads 4874756 Design and Implementation of Security Middleware for Data Warehouse Signature, Framework
Authors: Mayada Al Meghari
Abstract:
Recently, grid middlewares have provided large integrated use of network resources as the shared data and the CPU to become a virtual supercomputer. In this work, we present the design and implementation of the middleware for Data Warehouse Signature, DWS Framework. The aim of using the middleware in our DWS framework is to achieve the high performance by the parallel computing. This middleware is developed on Alchemi.Net framework to increase the security among the network nodes through the authentication and group-key distribution model. This model achieves the key security and prevents any intermediate attacks in the middleware. This paper presents the flow process structures of the middleware design. In addition, the paper ensures the implementation of security for DWS middleware enhancement with the authentication and group-key distribution model. Finally, from the analysis of other middleware approaches, the developed middleware of DWS framework is the optimal solution of a complete covering of security issues.Keywords: middleware, parallel computing, data warehouse, security, group-key, high performance
Procedia PDF Downloads 1194755 Data Compression in Ultrasonic Network Communication via Sparse Signal Processing
Authors: Beata Zima, Octavio A. Márquez Reyes, Masoud Mohammadgholiha, Jochen Moll, Luca de Marchi
Abstract:
This document presents the approach of using compressed sensing in signal encoding and information transferring within a guided wave sensor network, comprised of specially designed frequency steerable acoustic transducers (FSATs). Wave propagation in a damaged plate was simulated using commercial FEM-based software COMSOL. Guided waves were excited by means of FSATs, characterized by the special shape of its electrodes, and modeled using PIC255 piezoelectric material. The special shape of the FSAT, allows for focusing wave energy in a certain direction, accordingly to the frequency components of its actuation signal, which makes available a larger monitored area. The process begins when a FSAT detects and records reflection from damage in the structure, this signal is then encoded and prepared for transmission, using a combined approach, based on Compressed Sensing Matching Pursuit and Quadrature Amplitude Modulation (QAM). After codification of the signal is in binary chars the information is transmitted between the nodes in the network. The message reaches the last node, where it is finally decoded and processed, to be used for damage detection and localization purposes. The main aim of the investigation is to determine the location of detected damage using reconstructed signals. The study demonstrates that the special steerable capabilities of FSATs, not only facilitate the detection of damage but also permit transmitting the damage information to a chosen area in a specific direction of the investigated structure.Keywords: data compression, ultrasonic communication, guided waves, FEM analysis
Procedia PDF Downloads 1244754 Self in Networks: Public Sphere in the Era of Globalisation
Authors: Sanghamitra Sadhu
Abstract:
A paradigm shift from capitalism to information technology is discerned in the era globalisation. The idea of public sphere, which was theorized in terms of its decline in the wake of the rise of commercial mass media has now emerged as a transnational or global sphere with the discourse being dominated by the ‘network society’. In other words, the dynamic of globalisation has brought about ‘a spatial turn’ in the social and political sciences which is also manifested in the public sphere, Especially the global public sphere. The paper revisits the Habermasian concept of the public sphere and focuses on the various social networking sites with their plausibility to create a virtual global public sphere. Situating Habermas’s notion of the bourgeois public sphere in the present context of global public sphere, it considers the changing dimensions of the public sphere across time and examines the concept of the ‘public’ with its shifting transformation from the concrete collective to the fluid ‘imagined’ category. The paper addresses the problematic of multimodal self-portraiture in the social networking sites as well as various online diaries/journals with an attempt to explore the nuances of the networked self.Keywords: globalisation, network society, public sphere, self-fashioning, identity, autonomy
Procedia PDF Downloads 4164753 Using Action Based Research to Examine the Effects of Co-Teaching on Middle School and High School Student Achievement in Math and Language Arts
Authors: Kathleen L. Seifert
Abstract:
Students with special needs are expected to achieve the same academic standards as their general education peers, yet many students with special needs are pulled-out of general content instruction. Because of this, many students with special needs are denied content knowledge from a content expert and instead receive content instruction in a more restrictive setting. Collaborative teaching, where a general education and special education teacher work alongside each other in the same classroom, has become increasingly popular as a means to meet the diverse needs of students in America’s public schools. The idea behind co-teaching is noble; to ensure students with special needs receive content area instruction from a content expert while also receiving the necessary supports to be successful. However, in spite of this noble effort, the effects of co-teaching are not always positive. The reasons why have produced several hypotheses, one of which has to do with lack of proper training and implementation of effective evidence-based co-teaching practices. In order to examine the effects of co-teacher training, eleven teaching pairs from a small mid-western school district in the United States participated in a study. The purpose of the study was to examine the effects of co-teacher training on middle and high school student achievement in Math and Language Arts. A local university instructor provided teachers with training in co-teaching via a three-day workshop. In addition, co-teaching pairs were given the opportunity for direct observation and feedback using the Co-teaching Core Competencies Observation Checklist throughout the academic year. Data are in the process of being collected on both the students enrolled in the co-taught classes as well as on the teachers themselves. Student data compared achievement on standardized assessments and classroom performance across three domains: 1. General education students compared to students with special needs in co-taught classrooms, 2. Students with special needs in classrooms with and without co-teaching, 3. Students in classrooms where teachers were given observation and feedback compared to teachers who refused the observation and feedback. Teacher data compared the perceptions of the co-teaching initiative between teacher pairs who received direct observation and feedback from those who did not. The findings from the study will be shared with the school district and used for program improvement.Keywords: collabortive teaching, collaboration, co-teaching, professional development
Procedia PDF Downloads 1194752 Deep Learning for Image Correction in Sparse-View Computed Tomography
Authors: Shubham Gogri, Lucia Florescu
Abstract:
Medical diagnosis and radiotherapy treatment planning using Computed Tomography (CT) rely on the quantitative accuracy and quality of the CT images. At the same time, requirements for CT imaging include reducing the radiation dose exposure to patients and minimizing scanning time. A solution to this is the sparse-view CT technique, based on a reduced number of projection views. This, however, introduces a new problem— the incomplete projection data results in lower quality of the reconstructed images. To tackle this issue, deep learning methods have been applied to enhance the quality of the sparse-view CT images. A first approach involved employing Mir-Net, a dedicated deep neural network designed for image enhancement. This showed promise, utilizing an intricate architecture comprising encoder and decoder networks, along with the incorporation of the Charbonnier Loss. However, this approach was computationally demanding. Subsequently, a specialized Generative Adversarial Network (GAN) architecture, rooted in the Pix2Pix framework, was implemented. This GAN framework involves a U-Net-based Generator and a Discriminator based on Convolutional Neural Networks. To bolster the GAN's performance, both Charbonnier and Wasserstein loss functions were introduced, collectively focusing on capturing minute details while ensuring training stability. The integration of the perceptual loss, calculated based on feature vectors extracted from the VGG16 network pretrained on the ImageNet dataset, further enhanced the network's ability to synthesize relevant images. A series of comprehensive experiments with clinical CT data were conducted, exploring various GAN loss functions, including Wasserstein, Charbonnier, and perceptual loss. The outcomes demonstrated significant image quality improvements, confirmed through pertinent metrics such as Peak Signal-to-Noise Ratio (PSNR) and Structural Similarity Index (SSIM) between the corrected images and the ground truth. Furthermore, learning curves and qualitative comparisons added evidence of the enhanced image quality and the network's increased stability, while preserving pixel value intensity. The experiments underscored the potential of deep learning frameworks in enhancing the visual interpretation of CT scans, achieving outcomes with SSIM values close to one and PSNR values reaching up to 76.Keywords: generative adversarial networks, sparse view computed tomography, CT image correction, Mir-Net
Procedia PDF Downloads 1624751 Blockchain’s Feasibility in Military Data Networks
Authors: Brenden M. Shutt, Lubjana Beshaj, Paul L. Goethals, Ambrose Kam
Abstract:
Communication security is of particular interest to military data networks. A relatively novel approach to network security is blockchain, a cryptographically secured distribution ledger with a decentralized consensus mechanism for data transaction processing. Recent advances in blockchain technology have proposed new techniques for both data validation and trust management, as well as different frameworks for managing dataflow. The purpose of this work is to test the feasibility of different blockchain architectures as applied to military command and control networks. Various architectures are tested through discrete-event simulation and the feasibility is determined based upon a blockchain design’s ability to maintain long-term stable performance at industry standards of throughput, network latency, and security. This work proposes a consortium blockchain architecture with a computationally inexpensive consensus mechanism, one that leverages a Proof-of-Identity (PoI) concept and a reputation management mechanism.Keywords: blockchain, consensus mechanism, discrete-event simulation, fog computing
Procedia PDF Downloads 1384750 Strategies to Achieve Deep Decarbonisation in Power Generation: A Review
Authors: Abdullah Alotaiq
Abstract:
The transition to low-carbon power generation is essential for mitigating climate change and achieving sustainability. This process, however, entails considerable costs, and understanding the factors influencing these costs is critical. This is necessary to cater to the increasing demand for low-carbon electricity across the heating, industry, and transportation sectors. A crucial aspect of this transition is identifying cost-effective and feasible paths for decarbonization, which is integral to global climate mitigation efforts. It is concluded that hybrid solutions, combining different low-carbon technologies, are optimal for minimizing costs and enhancing flexibility. These solutions also address the challenges associated with phasing out existing fossil fuel-based power plants and broadening the spectrum of low-carbon power generation options.Keywords: review, power generation, energy transition, decarbonisation
Procedia PDF Downloads 544749 Simulation of Solar Assisted Absorption Cooling and Electricity Generation along with Thermal Storage
Authors: Faezeh Mosallat, Eric L. Bibeau, Tarek El Mekkawy
Abstract:
Availability of a wide variety of renewable resources, such as large reserves of hydro, biomass, solar and wind in Canada provides significant potential to improve the sustainability of energy uses. As buildings represent a considerable portion of energy use in Canada, application of distributed solar energy systems for heating and cooling may increase the amount of renewable energy use. Parabolic solar trough systems have seen limited deployments in cold northern climates as they are more suitable for electricity production in southern latitudes. Heat production by concentrating solar rays using parabolic troughs can overcome the poor efficiencies of flat panels and evacuated tubes in cold climates. A numerical dynamic model is developed to simulate an installed parabolic solar trough facility in Winnipeg. The results of the numerical model are validated using the experimental data obtained from this system. The model is developed in Simulink and will be utilized to simulate a tri-generation system for heating, cooling and electricity generation in remote northern communities. The main objective of this simulation is to obtain operational data of solar troughs in cold climates as this is lacking in the literature. In this paper, the validated Simulink model is applied to simulate a solar assisted absorption cooling system along with electricity generation using organic Rankine cycle (ORC) and thermal storage. A control strategy is employed to distribute the heated oil from solar collectors among the above three systems considering the temperature requirements. This modeling provides dynamic performance results using real time minutely meteorological data which are collected at the same location the solar system is installed. This is a big step ahead of the current models by accurately calculating the available solar energy at each time step considering the solar radiation fluctuations due to passing clouds. The solar absorption cooling is modeled to use the generated heat from the solar trough system and provide cooling in summer for a greenhouse which is located next to the solar field. A natural gas water heater provides the required excess heat for the absorption cooling at low or no solar radiation periods. The results of the simulation are presented for a summer month in Winnipeg which includes the amount of generated electric power from ORC and contribution of solar energy in the cooling load provisionKeywords: absorption cooling, parabolic solar trough, remote community, validated model
Procedia PDF Downloads 216