Search results for: ultra-dense networks
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 2801

Search results for: ultra-dense networks

611 A Multilayer Perceptron Neural Network Model Optimized by Genetic Algorithm for Significant Wave Height Prediction

Authors: Luis C. Parra

Abstract:

The significant wave height prediction is an issue of great interest in the field of coastal activities because of the non-linear behavior of the wave height and its complexity of prediction. This study aims to present a machine learning model to forecast the significant wave height of the oceanographic wave measuring buoys anchored at Mooloolaba of the Queensland Government Data. Modeling was performed by a multilayer perceptron neural network-genetic algorithm (GA-MLP), considering Relu(x) as the activation function of the MLPNN. The GA is in charge of optimized the MLPNN hyperparameters (learning rate, hidden layers, neurons, and activation functions) and wrapper feature selection for the window width size. Results are assessed using Mean Square Error (MSE), Root Mean Square Error (RMSE), and Mean Absolute Error (MAE). The GAMLPNN algorithm was performed with a population size of thirty individuals for eight generations for the prediction optimization of 5 steps forward, obtaining a performance evaluation of 0.00104 MSE, 0.03222 RMSE, 0.02338 MAE, and 0.71163% of MAPE. The results of the analysis suggest that the MLPNNGA model is effective in predicting significant wave height in a one-step forecast with distant time windows, presenting 0.00014 MSE, 0.01180 RMSE, 0.00912 MAE, and 0.52500% of MAPE with 0.99940 of correlation factor. The GA-MLP algorithm was compared with the ARIMA forecasting model, presenting better performance criteria in all performance criteria, validating the potential of this algorithm.

Keywords: significant wave height, machine learning optimization, multilayer perceptron neural networks, evolutionary algorithms

Procedia PDF Downloads 107
610 Self-Supervised Attributed Graph Clustering with Dual Contrastive Loss Constraints

Authors: Lijuan Zhou, Mengqi Wu, Changyong Niu

Abstract:

Attributed graph clustering can utilize the graph topology and node attributes to uncover hidden community structures and patterns in complex networks, aiding in the understanding and analysis of complex systems. Utilizing contrastive learning for attributed graph clustering can effectively exploit meaningful implicit relationships between data. However, existing attributed graph clustering methods based on contrastive learning suffer from the following drawbacks: 1) Complex data augmentation increases computational cost, and inappropriate data augmentation may lead to semantic drift. 2) The selection of positive and negative samples neglects the intrinsic cluster structure learned from graph topology and node attributes. Therefore, this paper proposes a method called self-supervised Attributed Graph Clustering with Dual Contrastive Loss constraints (AGC-DCL). Firstly, Siamese Multilayer Perceptron (MLP) encoders are employed to generate two views separately to avoid complex data augmentation. Secondly, the neighborhood contrastive loss is introduced to constrain node representation using local topological structure while effectively embedding attribute information through attribute reconstruction. Additionally, clustering-oriented contrastive loss is applied to fully utilize clustering information in global semantics for discriminative node representations, regarding the cluster centers from two views as negative samples to fully leverage effective clustering information from different views. Comparative clustering results with existing attributed graph clustering algorithms on six datasets demonstrate the superiority of the proposed method.

Keywords: attributed graph clustering, contrastive learning, clustering-oriented, self-supervised learning

Procedia PDF Downloads 53
609 A Visual Analytics Tool for the Structural Health Monitoring of an Aircraft Panel

Authors: F. M. Pisano, M. Ciminello

Abstract:

Aerospace, mechanical, and civil engineering infrastructures can take advantages from damage detection and identification strategies in terms of maintenance cost reduction and operational life improvements, as well for safety scopes. The challenge is to detect so called “barely visible impact damage” (BVID), due to low/medium energy impacts, that can progressively compromise the structure integrity. The occurrence of any local change in material properties, that can degrade the structure performance, is to be monitored using so called Structural Health Monitoring (SHM) systems, in charge of comparing the structure states before and after damage occurs. SHM seeks for any "anomalous" response collected by means of sensor networks and then analyzed using appropriate algorithms. Independently of the specific analysis approach adopted for structural damage detection and localization, textual reports, tables and graphs describing possible outlier coordinates and damage severity are usually provided as artifacts to be elaborated for information extraction about the current health conditions of the structure under investigation. Visual Analytics can support the processing of monitored measurements offering data navigation and exploration tools leveraging the native human capabilities of understanding images faster than texts and tables. Herein, a SHM system enrichment by integration of a Visual Analytics component is investigated. Analytical dashboards have been created by combining worksheets, so that a useful Visual Analytics tool is provided to structural analysts for exploring the structure health conditions examined by a Principal Component Analysis based algorithm.

Keywords: interactive dashboards, optical fibers, structural health monitoring, visual analytics

Procedia PDF Downloads 124
608 The Security Trade-Offs in Resource Constrained Nodes for IoT Application

Authors: Sultan Alharby, Nick Harris, Alex Weddell, Jeff Reeve

Abstract:

The concept of the Internet of Things (IoT) has received much attention over the last five years. It is predicted that the IoT will influence every aspect of our lifestyles in the near future. Wireless Sensor Networks are one of the key enablers of the operation of IoTs, allowing data to be collected from the surrounding environment. However, due to limited resources, nature of deployment and unattended operation, a WSN is vulnerable to various types of attack. Security is paramount for reliable and safe communication between IoT embedded devices, but it does, however, come at a cost to resources. Nodes are usually equipped with small batteries, which makes energy conservation crucial to IoT devices. Nevertheless, security cost in terms of energy consumption has not been studied sufficiently. Previous research has used a security specification of 802.15.4 for IoT applications, but the energy cost of each security level and the impact on quality of services (QoS) parameters remain unknown. This research focuses on the cost of security at the IoT media access control (MAC) layer. It begins by studying the energy consumption of IEEE 802.15.4 security levels, which is followed by an evaluation for the impact of security on data latency and throughput, and then presents the impact of transmission power on security overhead, and finally shows the effects of security on memory footprint. The results show that security overhead in terms of energy consumption with a payload of 24 bytes fluctuates between 31.5% at minimum level over non-secure packets and 60.4% at the top security level of 802.15.4 security specification. Also, it shows that security cost has less impact at longer packet lengths, and more with smaller packet size. In addition, the results depicts a significant impact on data latency and throughput. Overall, maximum authentication length decreases throughput by almost 53%, and encryption and authentication together by almost 62%.

Keywords: energy consumption, IEEE 802.15.4, IoT security, security cost evaluation

Procedia PDF Downloads 168
607 Retail of Organic Food in Poland

Authors: Joanna Smoluk-Sikorska, Władysława Łuczka

Abstract:

Organic farming is an important element of sustainable agriculture. It has been developing very dynamically in Poland, especially since Poland’s accession to the EU. Nevertheless, properly functioning organic market is a necessary condition justifying development of organic agriculture. Despite significant improvement, this market in Poland is still in the initial stage of growth. An important element of the market is distribution, especially retail, which offers specified product range to consumers. Therefore, there is a need to investigate retail outlets offering organic food in order to improve functioning of this part of the market. The inquiry research conducted in three types of outlets offering organic food, between 2011 and 2012 in the 8 largest Polish cities, shows that the majority of outlets offer cereals, processed fruit and vegetables as well as spices and the least shops – meat and sausages. The distributors mostly indicate unsatisfactory product range of suppliers as the reason for this situation. The main providers of the outlets are wholesalers, particularly in case of processed products, and in fresh products – organic farms. A very important distribution obstacle is dispersion of producers, which generates high transportation costs and what follows that, high price of organics. In the investigated shops, the most often used price calculation method is a cost method. The majority of the groceries and specialist shops apply margins between 21 and 40%. The margin in specialist outlets is the highest, in regard to the qualified service and advice. In turn, most retail networks declare the margin between 0 and 20%, which is consistent with low-price strategy applied in these shops. Some lacks in the product range of organics and in particular high prices cause that the demand volume is rather low. Therefore there is a need to support certain market actions, e.g. on-farm processing or promotion.

Keywords: organic food, retail, product range, supply sources

Procedia PDF Downloads 297
606 Cyber Security and Risk Assessment of the e-Banking Services

Authors: Aisha F. Bushager

Abstract:

Today we are more exposed than ever to cyber threats and attacks at personal, community, organizational, national, and international levels. More aspects of our lives are operating on computer networks simply because we are living in the fifth domain, which is called the Cyberspace. One of the most sensitive areas that are vulnerable to cyber threats and attacks is the Electronic Banking (e-Banking) area, where the banking sector is providing online banking services to its clients. To be able to obtain the clients trust and encourage them to practice e-Banking, also, to maintain the services provided by the banks and ensure safety, cyber security and risks control should be given a high priority in the e-banking area. The aim of the study is to carry out risk assessment on the e-banking services and determine the cyber threats, cyber attacks, and vulnerabilities that are facing the e-banking area specifically in the Kingdom of Bahrain. To collect relevant data, structured interviews were taken place with e-banking experts in different banks. Then, collected data where used as in input to the risk management framework provided by the National Institute of Standards and Technology (NIST), which was the model used in the study to assess the risks associated with e-banking services. The findings of the study showed that the cyber threats are commonly human errors, technical software or hardware failure, and hackers, on the other hand, the most common attacks facing the e-banking sector were phishing, malware attacks, and denial-of-service. The risks associated with the e-banking services were around the moderate level, however, more controls and countermeasures must be applied to maintain the moderate level of risks. The results of the study will help banks discover their vulnerabilities and maintain their online services, in addition, it will enhance the cyber security and contribute to the management and control of risks that are facing the e-banking sector.

Keywords: cyber security, e-banking, risk assessment, threats identification

Procedia PDF Downloads 350
605 Preserving Urban Cultural Heritage with Deep Learning: Color Planning for Japanese Merchant Towns

Authors: Dongqi Li, Yunjia Huang, Tomo Inoue, Kohei Inoue

Abstract:

With urbanization, urban cultural heritage is facing the impact and destruction of modernization and urbanization. Many historical areas are losing their historical information and regional cultural characteristics, so it is necessary to carry out systematic color planning for historical areas in conservation. As an early focus on urban color planning, Japan has a systematic approach to urban color planning. Hence, this paper selects five merchant towns from the category of important traditional building preservation areas in Japan as the subject of this study to explore the color structure and emotion of this type of historic area. First, the image semantic segmentation method identifies the buildings, roads, and landscape environments. Their color data were extracted for color composition and emotion analysis to summarize their common features. Second, the obtained Internet evaluations were extracted by natural language processing for keyword extraction. The correlation analysis of the color structure and keywords provides a valuable reference for conservation decisions for this historic area in the town. This paper also combines the color structure and Internet evaluation results with generative adversarial networks to generate predicted images of color structure improvements and color improvement schemes. The methods and conclusions of this paper can provide new ideas for the digital management of environmental colors in historic districts and provide a valuable reference for the inheritance of local traditional culture.

Keywords: historic districts, color planning, semantic segmentation, natural language processing

Procedia PDF Downloads 88
604 Construction and Optimization of Green Infrastructure Network in Mountainous Counties Based on Morphological Spatial Pattern Analysis and Minimum Cumulative Resistance Models: A Case Study of Shapingba District, Chongqing

Authors: Yuning Guan

Abstract:

Under the background of rapid urbanization, mountainous counties need to break through mountain barriers for urban expansion due to undulating topography, resulting in ecological problems such as landscape fragmentation and reduced biodiversity. Green infrastructure networks are constructed to alleviate the contradiction between urban expansion and ecological protection, promoting the healthy and sustainable development of urban ecosystems. This study applies the MSPA model, the MCR model and Linkage Mapper Tools to identify eco-sources and eco-corridors in the Shapingba District of Chongqing and combined with landscape connectivity assessment and circuit theory to delineate the importance levels to extract ecological pinch point areas on the corridors. The results show that: (1) 20 ecological sources are identified, with a total area of 126.47 km², accounting for 31.88% of the study area, and showing a pattern of ‘one core, three corridors, multi-point distribution’. (2) 37 ecological corridors are formed in the area, with a total length of 62.52km, with a ‘more in the west, less in the east’ pattern. (3) 42 ecological pinch points are extracted, accounting for 25.85% of the length of the corridors, which are mainly distributed in the eastern new area. Accordingly, this study proposes optimization strategies for sub-area protection of ecological sources, grade-level construction of ecological corridors, and precise restoration of ecological pinch points.

Keywords: green infrastructure network, morphological spatial pattern, minimal cumulative resistance, mountainous counties, circuit theory, shapingba district

Procedia PDF Downloads 43
603 Security Issues on Smart Grid and Blockchain-Based Secure Smart Energy Management Systems

Authors: Surah Aldakhl, Dafer Alali, Mohamed Zohdy

Abstract:

The next generation of electricity grid infrastructure, known as the "smart grid," integrates smart ICT (information and communication technology) into existing grids in order to alleviate the drawbacks of existing one-way grid systems. Future power systems' efficiency and dependability are anticipated to significantly increase thanks to the Smart Grid, especially given the desire for renewable energy sources. The security of the Smart Grid's cyber infrastructure is a growing concern, though, as a result of the interconnection of significant power plants through communication networks. Since cyber-attacks can destroy energy data, beginning with personal information leaking from grid members, they can result in serious incidents like huge outages and the destruction of power network infrastructure. We shall thus propose a secure smart energy management system based on the Blockchain as a remedy for this problem. The power transmission and distribution system may undergo a transformation as a result of the inclusion of optical fiber sensors and blockchain technology in smart grids. While optical fiber sensors allow real-time monitoring and management of electrical energy flow, Blockchain offers a secure platform to safeguard the smart grid against cyberattacks and unauthorized access. Additionally, this integration makes it possible to see how energy is produced, distributed, and used in real time, increasing transparency. This strategy has advantages in terms of improved security, efficiency, dependability, and flexibility in energy management. An in-depth analysis of the advantages and drawbacks of combining blockchain technology with optical fiber is provided in this paper.

Keywords: smart grids, blockchain, fiber optic sensor, security

Procedia PDF Downloads 120
602 Ageing Patterns and Concerns in the Arabian Gulf: A Systematic Review

Authors: Asharaf Abdul Salam

Abstract:

Arabian Gulf countries have norms and rules different from others: having an exodus of male immigrant labor contract holders of age 20-60 years as a floating population. Such a demographic scenario camouflages population ageing. However, it is observed on examining vigilantly, not only in the native population but also in the general population. This research on population ageing in the Arabian Gulf examines ageing scenario and concerns through analyses of international databases (US Census Bureau and United Nations) and national level databases (Censuses and Surveys) apart from a review of published research. Transitions in demography and epidemiology lead to gains in life expectancy and thereby reductions in fertility, leading to ageing of the population in the region. Even after bringing adult immigrants, indices and age pyramids show an increasing ageing trend in the total population, demonstrating an ageing workforce. Besides, the exclusive native population analysis reveals a trend of expansive pyramids (pre-transitional stage) turning to constrictive (transition stage) and cylindrical (post-transition stage) shapes. Age-based indices such as the index of ageing, age dependency ratio, and median age confirm this trend. While the feminine nature of ageing is vivid, gains in life expectancy and causes of death in old age, indicating co-morbidity compression, are concerns to ageing. Preparations are in demand to cope with ageing from different dimensions, as explained in the United Nations Plans of Action. A strategy of strengthening informal care with supportive semi-formal and supplementary formal care networks would alleviate this crisis associated with population ageing.

Keywords: total versus native population, indices of ageing, age pyramids, feminine nature, comorbidity compression, strategic interventions

Procedia PDF Downloads 93
601 Isolation and Identification of Microorganisms from Marine-Associated Samples under Laboratory Conditions

Authors: Sameen Tariq, Saira Bano, Sayyada Ghufrana Nadeem

Abstract:

The Ocean, which covers over 70% of the world's surface, is wealthy in biodiversity as well as a rich wellspring of microorganisms with huge potential. The oceanic climate is home to an expansive scope of plants, creatures, and microorganisms. Marine microbial networks, which incorporate microscopic organisms, infections, and different microorganisms, enjoy different benefits in biotechnological processes. Samples were collected from marine environments, including soil and water samples, to cultivate the uncultured marine organisms by using Zobell’s medium, Sabouraud’s dextrose agar, and casein media for this purpose. Following isolation, we conduct microscopy and biochemical tests, including gelatin, starch, glucose, casein, catalase, and carbohydrate hydrolysis for further identification. The results show that more gram-positive and gram-negative bacteria. The isolation process of marine organisms is essential for understanding their ecological roles, unraveling their biological secrets, and harnessing their potential for various applications. Marine organisms exhibit remarkable adaptations to thrive in the diverse and challenging marine environment, offering vast potential for scientific, medical, and industrial applications. The isolation process plays a crucial role in unlocking the secrets of marine organisms, understanding their biological functions, and harnessing their valuable properties. They offer a rich source of bioactive compounds with pharmaceutical potential, including antibiotics, anticancer agents, and novel therapeutics. This study is an attempt to explore the diversity and dynamics related to marine microflora and their role in biofilm formation.

Keywords: marine microorganisms, ecosystem, fungi, biofilm, gram-positive, gram-negative

Procedia PDF Downloads 45
600 Understanding Algerian International Student Mental Health Experiences in UK (United Kingdom) Universities: Difficulties of Disclosure, Help-Seeking and Coping Strategies

Authors: Nesrine Boussaoui

Abstract:

Background: International students often encounter challenges while studying in the UK, including communication and language barriers, lack of social networks, and socio-cultural differences that adversely impact on their mental health. For Algerian international students (AISs), these challenges may be heightened as English is not their first language and the culture of their homeland is substantially different from British culture, yet research has to incorporate their experiences and perspectives. Aim: The current study aimed to explore AISs’ 1) understandings of mental health; 2) issues of disclosure for mental health difficulties; and 3) mental health help-seeking and coping strategies. Method: In-depth, audio recorded semi-structured interviews (n = 20) with AISs in UK universities were conducted. An inductive, reflective thematic approach analysis was used. Finding: The following themes and associated sub-themes were developed: (1) Algerian cultural influences on mental health understanding(socio-cultural comparisons); (2) the paradox of the family (pressure vs. support); (3) stigma and fear of disclosure; (4) Barriers to formal help-seeking (informal disclosure as first step to seeking help); (5) Communication barriers (resort to mother tongue to disclose); (6) Self-reliance and religious coping. Conclusion: Recognising and understanding the challenges faced by AISs in terms of disclosure and mental health help-seeking is essential to reduce barriers to formal help-seeking. Informal disclosure among peers is often the first step to seeking help. Enhancing practitioners’ cultural competences and awareness of diverse understandings of mental health and the role of religious coping among AISs’ may have transferable benefits to a wider international student population.

Keywords: mental health, stegma, coping, disclosure

Procedia PDF Downloads 142
599 Management of Interdependence in Manufacturing Networks

Authors: Atour Taghipour

Abstract:

In the real world each manufacturing company is an independent business unit. These business units are linked to each other through upstream and downstream linkages. The management of these linkages is called coordination which, could be considered as a difficult engineering task. The degree of difficulty of coordination depends on the type and the nature of information exchanged between partners as well as the structure of relationship from mutual to the network structure. The literature of manufacturing systems comprises a wide range of varieties of methods and approaches of coordination. In fact, two main streams of research can be distinguished: central coordination versus decentralized coordination. In the centralized systems a high degree of information exchanges is required. The high degree of information exchanges sometimes leads to difficulties when independent members do not want to share information. In order to address these difficulties, decentralized approaches of coordination of operations planning decisions based on some minimal information sharing have been proposed in many academic disciplines. This paper first proposes a framework of analysis in order to analyze the proposed approaches in the literature, based on this framework which includes the similarities between approaches we categorize the existing approaches. This classification can be used as a research map for future researches. The result of our paper highlights several opportunities for future research. First, it is proposed to develop more dynamic and stochastic mechanisms of planning coordination of manufacturing units. Second, in order to exploit the complementarities of approaches proposed by diverse science discipline, we propose to integrate the techniques of coordination. Finally, based on our approach we proposed to develop coordination standards to guaranty both the complementarity of these approaches as well as the freedom of companies to adopt any planning tools.

Keywords: network coordination, manufacturing, operations planning, supply chain

Procedia PDF Downloads 281
598 Assessing Renewal Needs of Urban Water Infrastructure Systems: Case Study of Linköping in Sweden

Authors: Eman Hegazy, Stefan Anderberg, Joakim Krook

Abstract:

Urban water infrastructure systems are central to functioning cities. For securing a continuous and efficient supply of the systems services, continuous investment, maintenance, and renewal are needed. Neglecting maintenance and renewal can lead to recurrent breakdown problems as systems age, which makes it more and more difficult to secure efficient long-term supply. Globally, many cities struggle with aging water infrastructure, often due to competing funding priorities. Investment in maintenance and renewal is not prioritized. The problem primarily stems from the challenge of reaping the benefits of investments promptly. The long-term benefits gained from investing in the renewal of water infrastructure may be achievable in the long run, resulting in the oversight of such investments. This leads to a build-up of "renewal debt" for future generations to inherit. Addressing this issue is difficult due to various contributing factors and the complex nature of the systems. The study aims to contribute to an increased understanding of the long-term management challenges of urban water infrastructure, the development of improved maintenance and renewal strategies through the examination of water infrastructure management, and the assessment of the adequacy of the maintenance and renewal in a case study, the city of Linköping, Sweden. Employing a multi-methods approach, this study utilized both qualitative and quantitative methods, including interviews, workshops, and data analysis. The findings of the study provided insights into the current status of the water and sewerage networks in Linkoping, highlighting the risks to ensuring reliable and sustainable water supply and discussing strategies for improving maintenance and renewal.

Keywords: case study, infrastructure management, renewal needs, Sweden, urban water infrastructure

Procedia PDF Downloads 68
597 Self-Regulated Learning: A Required Skill for Web 2.0 Internet-Based Learning

Authors: Pieter Conradie, M. Marina Moller

Abstract:

Web 2.0 Internet-based technologies have intruded all aspects of human life. Presently, this phenomenon is especially evident in the educational context, with increased disruptive Web 2.0 technology infusions dramatically changing educational practice. The most prominent of these Web 2.0 intrusions can be identified as Massive Open Online Courses (Coursera, EdX), video and photo sharing sites (Youtube, Flickr, Instagram), and Web 2.0 online tools utilize to create Personal Learning Environments (PLEs) (Symbaloo (aggregator), Delicious (social bookmarking), PBWorks (collaboration), Google+ (social networks), Wordspress (blogs), Wikispaces (wiki)). These Web 2.0 technologies have supported the realignment from a teacher-based pedagogy (didactic presentation) to a learner-based pedagogy (problem-based learning, project-based learning, blended learning), allowing greater learner autonomy. No longer is the educator the source of knowledge. Instead the educator has become the facilitator and mediator of the learner, involved in developing learner competencies to support life-long learning (continuous learning) in the 21st century. In this study, the self-regulated learning skills of thirty first-year university learners were explored by utilizing the Online Self-regulated Learning Questionnaire. Implementing an action research method, an intervention was affected towards improving the self-regulation skill set of the participants. Statistical significant results were obtained with increased self-regulated learning proficiency, positively impacting learner performance. Goal setting, time management, environment structuring, help seeking, task (learning) strategies and self-evaluation skills were confirmed as determinants of improved learner success.

Keywords: andragogy, online self-regulated learning questionnaire, self-regulated learning, web 2.0

Procedia PDF Downloads 417
596 Performance Analysis of Pumps-as-Turbine Under Cavitating Conditions

Authors: Calvin Stephen, Biswajit Basu, Aonghus McNabola

Abstract:

Market liberalization in the power sector has led to the emergence of micro-hydropower schemes that are dependent on the use of pumps-as-turbines in applications that were not suitable as potential hydropower sites in earlier years. These applications include energy recovery in water supply networks, sewage systems, irrigation systems, alcohol breweries, underground mining and desalination plants. As a result, there has been an accelerated adoption of pumpsas-turbine technology due to the economic advantages it presents in comparison to the conventional turbines in the micro-hydropower space. The performance of this machines under cavitation conditions, however, is not well understood as there is a deficiency of knowledge in literature focused on their turbine mode of operation. In hydraulic machines, cavitation is a common occurrence which needs to be understood to safeguard them and prolong their operation life. The overall purpose of this study is to investigate the effects of cavitation on the performance of a pumps-as-turbine system over its entire operating range. At various operating speeds, the cavitating region is identified experimentally while monitoring the effects this has on the power produced by the machine. Initial results indicate occurrence of cavitation at higher flow rates for lower operating speeds and at lower flow rates at higher operating speeds. This implies that for cavitation free operation, low speed pumps-as-turbine must be used for low flow rate conditions whereas for sites with higher flow rate conditions high speed turbines should be adopted. Such a complete understanding of pumps-as-turbine suction performance can aid avoid cavitation induced failures hence improved reliability of the micro-hydropower plant.

Keywords: cavitation, micro-hydropower, pumps-as-turbine, system design

Procedia PDF Downloads 119
595 Investigation on Reducing the Bandgap in Nanocomposite Polymers by Doping

Authors: Sharvare Palwai, Padmaja Guggilla

Abstract:

Smart materials, also called as responsive materials, undergo reversible physical or chemical changes in their properties as a consequence of small environmental variations. They can respond to a single or multiple stimuli such as stress, temperature, moist, electric or magnetic fields, light, or chemical compounds. Hence smart materials are the basis of many applications, including biosensors and transducers, particularly electroactive polymers. As the polymers exhibit good flexibility, high transparency, easy processing, and low cost, they would be promising for the sensor material. Polyvinylidene Fluoride (PVDF), being a ferroelectric polymer, exhibits piezoelectric and pyro electric properties. Pyroelectric materials convert heat directly into electricity, while piezoelectric materials convert mechanical energy into electricity. These characteristics of PVDF make it useful in biosensor devices and batteries. However, the influence of nanoparticle fillers such as Lithium Tantalate (LiTaO₃/LT), Potassium Niobate (KNbO₃/PN), and Zinc Titanate (ZnTiO₃/ZT) in polymer films will be studied comprehensively. Developing advanced and cost-effective biosensors is pivotal to foresee the fullest potential of polymer based wireless sensor networks, which will further enable new types of self-powered applications. Finally, nanocomposites films with best set of properties; the sensory elements will be designed and tested for their performance as electric generators under laboratory conditions. By characterizing the materials for their optical properties and investigate the effects of doping on the bandgap energies, the science in the next-generation biosensor technologies can be advanced.

Keywords: polyvinylidene fluoride, PVDF, lithium tantalate, potassium niobate, zinc titanate

Procedia PDF Downloads 134
594 Bayesian System and Copula for Event Detection and Summarization of Soccer Videos

Authors: Dhanuja S. Patil, Sanjay B. Waykar

Abstract:

Event detection is a standout amongst the most key parts for distinctive sorts of area applications of video data framework. Recently, it has picked up an extensive interest of experts and in scholastics from different zones. While detecting video event has been the subject of broad study efforts recently, impressively less existing methodology has considered multi-model data and issues related efficiency. Start of soccer matches different doubtful circumstances rise that can't be effectively judged by the referee committee. A framework that checks objectively image arrangements would prevent not right interpretations because of some errors, or high velocity of the events. Bayesian networks give a structure for dealing with this vulnerability using an essential graphical structure likewise the probability analytics. We propose an efficient structure for analysing and summarization of soccer videos utilizing object-based features. The proposed work utilizes the t-cherry junction tree, an exceptionally recent advancement in probabilistic graphical models, to create a compact representation and great approximation intractable model for client’s relationships in an interpersonal organization. There are various advantages in this approach firstly; the t-cherry gives best approximation by means of junction trees class. Secondly, to construct a t-cherry junction tree can be to a great extent parallelized; and at last inference can be performed utilizing distributed computation. Examination results demonstrates the effectiveness, adequacy, and the strength of the proposed work which is shown over a far reaching information set, comprising more soccer feature, caught at better places.

Keywords: summarization, detection, Bayesian network, t-cherry tree

Procedia PDF Downloads 324
593 Neural Network based Risk Detection for Dyslexia and Dysgraphia in Sinhala Language Speaking Children

Authors: Budhvin T. Withana, Sulochana Rupasinghe

Abstract:

The educational system faces a significant concern with regards to Dyslexia and Dysgraphia, which are learning disabilities impacting reading and writing abilities. This is particularly challenging for children who speak the Sinhala language due to its complexity and uniqueness. Commonly used methods to detect the risk of Dyslexia and Dysgraphia rely on subjective assessments, leading to limited coverage and time-consuming processes. Consequently, delays in diagnoses and missed opportunities for early intervention can occur. To address this issue, the project developed a hybrid model that incorporates various deep learning techniques to detect the risk of Dyslexia and Dysgraphia. Specifically, Resnet50, VGG16, and YOLOv8 models were integrated to identify handwriting issues. The outputs of these models were then combined with other input data and fed into an MLP model. Hyperparameters of the MLP model were fine-tuned using Grid Search CV, enabling the identification of optimal values for the model. This approach proved to be highly effective in accurately predicting the risk of Dyslexia and Dysgraphia, providing a valuable tool for early detection and intervention. The Resnet50 model exhibited a training accuracy of 0.9804 and a validation accuracy of 0.9653. The VGG16 model achieved a training accuracy of 0.9991 and a validation accuracy of 0.9891. The MLP model demonstrated impressive results with a training accuracy of 0.99918, a testing accuracy of 0.99223, and a loss of 0.01371. These outcomes showcase the high accuracy achieved by the proposed hybrid model in predicting the risk of Dyslexia and Dysgraphia.

Keywords: neural networks, risk detection system, dyslexia, dysgraphia, deep learning, learning disabilities, data science

Procedia PDF Downloads 64
592 A Mega-Analysis of the Predictive Power of Initial Contact within Minimal Social Network

Authors: Cathal Ffrench, Ryan Barrett, Mike Quayle

Abstract:

It is accepted in social psychology that categorization leads to ingroup favoritism, without further thought given to the processes that may co-occur or even precede categorization. These categorizations move away from the conceptualization of the self as a unique social being toward an increasingly collective identity. Subsequently, many individuals derive much of their self-evaluations from these collective identities. The seminal literature on this topic argues that it is primarily categorization that evokes instances of ingroup favoritism. Apropos to these theories, we argue that categorization acts to enhance and further intergroup processes rather than defining them. More accurately, we propose categorization aids initial ingroup contact and this first contact is predictive of subsequent favoritism on individual and collective levels. This analysis focuses on Virtual Interaction APPLication (VIAPPL) based studies, a software interface that builds on the flaws of the original minimal group studies. The VIAPPL allows the exchange of tokens in an intra and inter-group manner. This token exchange is how we classified the first contact. The study involves binary longitudinal analysis to better understand the subsequent exchanges of individuals based on who they first interacted with. Studies were selected on the criteria of evidence of explicit first interactions and two-group designs. Our findings paint a compelling picture in support of a motivated contact hypothesis, which suggests that an individual’s first motivated contact toward another has strong predictive capabilities for future behavior. This contact can lead to habit formation and specific favoritism towards individuals where contact has been established. This has important implications for understanding how group conflict occurs, and how intra-group individual bias can develop.

Keywords: categorization, group dynamics, initial contact, minimal social networks, momentary contact

Procedia PDF Downloads 148
591 Disarmament and Rehabilitation of Women Maoists: A Case Study of Chhattisgarh, India

Authors: Pinal Patel

Abstract:

The study defines the problems and issues of women in Maoist groups, also referred as ‘Naxalites’, in Chhattisgarh, India. It analyses the causes and consequences of increasing number of women joining Maoists groups and measures taken by the central and state government to retreat them. The main aspect of the study is, how to counter the challenges to resolve the issues and restore normalcy in the life of women Maoists to resettle them in mainstream once they become physically inactive and wish to become part of the society. The rationale behind this study is that women Maoists once inactive, has no place either with Maoist camps/rebel groups or particularly in society. The problems faced by the women Maoists, in society as well as in Maoists camps, can be studied through social, economic, cultural, political and humanitarian aspects. The methodology of the study is dependent on primary sources of information which includes a research survey in majorly affected areas, statistical analysis. Secondary sources of information are helpful for understanding the background of the problem. Government’s strategy of rewarding with cash and providing resettlement and rehabilitation benefits including houses and jobs to ex-women Maoists and their families is a well formulated and feasible policy and effectively implemented by the concerned authorities. But, the survey results show that the policy has not been able to have impacts as it was intended. Because inactive and physically disabled women are still left deserted in deep forests to die and police or authorities are not able to reach them and bring them back. The difficult terrain and dense forest areas are major hurdles to reach to Maoists camps. Moreover, to make people aware of government’s surrendering and rehabilitation schemes and policies as communication networks are very poor due to the lack of development in the state.

Keywords: maoists, women, government, policy

Procedia PDF Downloads 121
590 Assessing Significance of Correlation with Binomial Distribution

Authors: Vijay Kumar Singh, Pooja Kushwaha, Prabhat Ranjan, Krishna Kumar Ojha, Jitendra Kumar

Abstract:

Present day high-throughput genomic technologies, NGS/microarrays, are producing large volume of data that require improved analysis methods to make sense of the data. The correlation between genes and samples has been regularly used to gain insight into many biological phenomena including, but not limited to, co-expression/co-regulation, gene regulatory networks, clustering and pattern identification. However, presence of outliers and violation of assumptions underlying Pearson correlation is frequent and may distort the actual correlation between the genes and lead to spurious conclusions. Here, we report a method to measure the strength of association between genes. The method assumes that the expression values of a gene are Bernoulli random variables whose outcome depends on the sample being probed. The method considers the two genes as uncorrelated if the number of sample with same outcome for both the genes (Ns) is equal to certainly expected number (Es). The extent of correlation depends on how far Ns can deviate from the Es. The method does not assume normality for the parent population, fairly unaffected by the presence of outliers, can be applied to qualitative data and it uses the binomial distribution to assess the significance of association. At this stage, we would not claim about the superiority of the method over other existing correlation methods, but our method could be another way of calculating correlation in addition to existing methods. The method uses binomial distribution, which has not been used until yet, to assess the significance of association between two variables. We are evaluating the performance of our method on NGS/microarray data, which is noisy and pierce by the outliers, to see if our method can differentiate between spurious and actual correlation. While working with the method, it has not escaped our notice that the method could also be generalized to measure the association of more than two variables which has been proven difficult with the existing methods.

Keywords: binomial distribution, correlation, microarray, outliers, transcriptome

Procedia PDF Downloads 415
589 Artificial Neural Network-Based Prediction of Effluent Quality of Wastewater Treatment Plant Employing Data Preprocessing Approaches

Authors: Vahid Nourani, Atefeh Ashrafi

Abstract:

Prediction of treated wastewater quality is a matter of growing importance in water treatment procedure. In this way artificial neural network (ANN), as a robust data-driven approach, has been widely used for forecasting the effluent quality of wastewater treatment. However, developing ANN model based on appropriate input variables is a major concern due to the numerous parameters which are collected from treatment process and the number of them are increasing in the light of electronic sensors development. Various studies have been conducted, using different clustering methods, in order to classify most related and effective input variables. This issue has been overlooked in the selecting dominant input variables among wastewater treatment parameters which could effectively lead to more accurate prediction of water quality. In the presented study two ANN models were developed with the aim of forecasting effluent quality of Tabriz city’s wastewater treatment plant. Biochemical oxygen demand (BOD) was utilized to determine water quality as a target parameter. Model A used Principal Component Analysis (PCA) for input selection as a linear variance-based clustering method. Model B used those variables identified by the mutual information (MI) measure. Therefore, the optimal ANN structure when the result of model B compared with model A showed up to 15% percent increment in Determination Coefficient (DC). Thus, this study highlights the advantage of PCA method in selecting dominant input variables for ANN modeling of wastewater plant efficiency performance.

Keywords: Artificial Neural Networks, biochemical oxygen demand, principal component analysis, mutual information, Tabriz wastewater treatment plant, wastewater treatment plant

Procedia PDF Downloads 128
588 A Comprehensive Study and Evaluation on Image Fashion Features Extraction

Authors: Yuanchao Sang, Zhihao Gong, Longsheng Chen, Long Chen

Abstract:

Clothing fashion represents a human’s aesthetic appreciation towards everyday outfits and appetite for fashion, and it reflects the development of status in society, humanity, and economics. However, modelling fashion by machine is extremely challenging because fashion is too abstract to be efficiently described by machines. Even human beings can hardly reach a consensus about fashion. In this paper, we are dedicated to answering a fundamental fashion-related problem: what image feature best describes clothing fashion? To address this issue, we have designed and evaluated various image features, ranging from traditional low-level hand-crafted features to mid-level style awareness features to various current popular deep neural network-based features, which have shown state-of-the-art performance in various vision tasks. In summary, we tested the following 9 feature representations: color, texture, shape, style, convolutional neural networks (CNNs), CNNs with distance metric learning (CNNs&DML), AutoEncoder, CNNs with multiple layer combination (CNNs&MLC) and CNNs with dynamic feature clustering (CNNs&DFC). Finally, we validated the performance of these features on two publicly available datasets. Quantitative and qualitative experimental results on both intra-domain and inter-domain fashion clothing image retrieval showed that deep learning based feature representations far outweigh traditional hand-crafted feature representation. Additionally, among all deep learning based methods, CNNs with explicit feature clustering performs best, which shows feature clustering is essential for discriminative fashion feature representation.

Keywords: convolutional neural network, feature representation, image processing, machine modelling

Procedia PDF Downloads 139
587 Redefining “Minor”: An Empirical Research on Two Biennials in Contemporary China

Authors: Mengwei Li

Abstract:

Since the 1990s, biennials, and large-scale transnational art exhibitions, have proliferated exponentially across the globe, particularly in Asia, Africa, and Latin America. It has spurred debates regarding the inclusion of "new art cultures" and the deconstruction of the mechanism of exclusion embedded in the Western monopoly on art. Hans Belting introduced the concept of "global art" in 2013 to denounce the West's privileged canons in art by emphasising the inclusion of art practices from alleged non-Western regions. Arguably, the rise of new biennial networks developed by these locations has contributed to the asserted "inclusion of new art worlds." However, phrases such as "non-Western" and "beyond Euro-American" attached to these discussions raise the question of non- or beyond- in relation to whom. In this narrative, to become "integrated" and "equal" implies entry into the "core," a universal system in which preexisting authoritative voices define "newcomers" by what they are not. Possibly, if there is a global biennial system that symbolises a "universal language" of the contemporary art world, it is centered on the inherently dynamic yet asymmetrical interaction and negotiation between the "core" and the rest of the world's "periphery." Engaging with theories of "minor literature" developed by Deleuze and Guattari, this research proposes an epistemological framework to comprehend the global biennial discourse since the 1990s. Using this framework, this research looks at two biennial models in China: the 13th Shanghai Biennale, which was organised in the country's metropolitan art centre, and the 2nd Yinchuan Biennale, which was inaugurated in a geographically and economically marginalised city compared to domestic centres. By analysing how these two biennials from different locations in China positioned themselves and conveyed their local profiles through the universal language of the biennial, this research identifies a potential "minor" positionality within the global biennial discourse from China's perspective.

Keywords: biennials, China, contemporary, global art, minor literature

Procedia PDF Downloads 87
586 Questioning the Relationship Between Young People and Fake News Through Their Use of Social Media

Authors: Marion Billard

Abstract:

This paper will focus on the question of the real relationship between young people and fake news. Fake news is one of today’s main issues in the world of information and communication. Social media and its democratization helped to spread false information. According to traditional beliefs, young people are more inclined to believe what they read through social media. But, the individuals concerned, think that they are more inclined to make a distinction between real and fake news. This phenomenon is due to their use of the internet and social media from an early age. During the 2016 and 2017 French and American presidential campaigns, the term fake news was in the mouth of the entire world and became a real issue in the field of information. While young people were informing themselves with newspapers or television until the beginning of the ’90s, Gen Z (meaning people born between 1997 and 2010), has always been immersed in this world of fast communication. They know how to use social media from a young age and the internet has no secret for them. Today, despite the sporadic use of traditional media, young people tend to turn to their smartphones and social networks such as Instagram or Twitter to stay abreast of the latest news. The growth of social media information led to an “ambient journalism”, giving access to an endless quantity of information. Waking up in the morning, young people will see little posts with short texts supplying the essential of the news, without, for the most, many details. As a result, impressionable people are not able to do a distinction between real media, and “junk news” or Fake News. This massive use of social media is probably explained by the inability of the youngsters to find connections between the communication of the traditional media and what they are living. The question arises if this over-confidence of the young people in their ability to distinguish between accurate and fake news would not make it more difficult for them to examine critically the information. Their relationship with media and fake news is more complex than popular opinion. Today’s young people are not the master in the quest for information, nor inherently the most impressionable public on social media.

Keywords: fake news, youngsters, social media, information, generation

Procedia PDF Downloads 161
585 Neural Network-based Risk Detection for Dyslexia and Dysgraphia in Sinhala Language Speaking Children

Authors: Budhvin T. Withana, Sulochana Rupasinghe

Abstract:

The problem of Dyslexia and Dysgraphia, two learning disabilities that affect reading and writing abilities, respectively, is a major concern for the educational system. Due to the complexity and uniqueness of the Sinhala language, these conditions are especially difficult for children who speak it. The traditional risk detection methods for Dyslexia and Dysgraphia frequently rely on subjective assessments, making it difficult to cover a wide range of risk detection and time-consuming. As a result, diagnoses may be delayed and opportunities for early intervention may be lost. The project was approached by developing a hybrid model that utilized various deep learning techniques for detecting risk of Dyslexia and Dysgraphia. Specifically, Resnet50, VGG16 and YOLOv8 were integrated to detect the handwriting issues, and their outputs were fed into an MLP model along with several other input data. The hyperparameters of the MLP model were fine-tuned using Grid Search CV, which allowed for the optimal values to be identified for the model. This approach proved to be effective in accurately predicting the risk of Dyslexia and Dysgraphia, providing a valuable tool for early detection and intervention of these conditions. The Resnet50 model achieved an accuracy of 0.9804 on the training data and 0.9653 on the validation data. The VGG16 model achieved an accuracy of 0.9991 on the training data and 0.9891 on the validation data. The MLP model achieved an impressive training accuracy of 0.99918 and a testing accuracy of 0.99223, with a loss of 0.01371. These results demonstrate that the proposed hybrid model achieved a high level of accuracy in predicting the risk of Dyslexia and Dysgraphia.

Keywords: neural networks, risk detection system, Dyslexia, Dysgraphia, deep learning, learning disabilities, data science

Procedia PDF Downloads 114
584 Identification of Risks Associated with Process Automation Systems

Authors: J. K. Visser, H. T. Malan

Abstract:

A need exists to identify the sources of risks associated with the process automation systems within petrochemical companies or similar energy related industries. These companies use many different process automation technologies in its value chain. A crucial part of the process automation system is the information technology component featuring in the supervisory control layer. The ever-changing technology within the process automation layers and the rate at which it advances pose a risk to safe and predictable automation system performance. The age of the automation equipment also provides challenges to the operations and maintenance managers of the plant due to obsolescence and unavailability of spare parts. The main objective of this research was to determine the risk sources associated with the equipment that is part of the process automation systems. A secondary objective was to establish whether technology managers and technicians were aware of the risks and share the same viewpoint on the importance of the risks associated with automation systems. A conceptual model for risk sources of automation systems was formulated from models and frameworks in literature. This model comprised six categories of risk which forms the basis for identifying specific risks. This model was used to develop a questionnaire that was sent to 172 instrument technicians and technology managers in the company to obtain primary data. 75 completed and useful responses were received. These responses were analyzed statistically to determine the highest risk sources and to determine whether there was difference in opinion between technology managers and technicians. The most important risks that were revealed in this study are: 1) the lack of skilled technicians, 2) integration capability of third-party system software, 3) reliability of the process automation hardware, 4) excessive costs pertaining to performing maintenance and migrations on process automation systems, and 5) requirements of having third-party communication interfacing compatibility as well as real-time communication networks.

Keywords: distributed control system, identification of risks, information technology, process automation system

Procedia PDF Downloads 139
583 Combining Transcriptomics, Bioinformatics, Biosynthesis Networks and Chromatographic Analyses for Cotton Gossypium hirsutum L. Defense Volatiles Study

Authors: Ronald Villamar-Torres, Michael Staudt, Christopher Viot

Abstract:

Cotton Gossypium hirsutum L. is one of the most important industrial crops, producing the world leading natural textile fiber, but is very prone to arthropod attacks that reduce crop yield and quality. Cotton cultivation, therefore, makes an outstanding use of chemical pesticides. In reaction to herbivorous arthropods, cotton plants nevertheless show natural defense reactions, in particular through volatile organic compounds (VOCs) emissions. These natural defense mechanisms are nowadays underutilized but have a very high potential for cotton cultivation, and elucidating their genetic bases will help to improve their use. Simulating herbivory attacks by mechanical wounding of cotton plants in greenhouse, we studied by qPCR the changes in gene expression for genes of the terpenoids biosynthesis pathway. Differentially expressed genes corresponded to higher levels of the terpenoids biosynthesis pathway and not to enzymes synthesizing particular terpenoids. The genes were mapped on the G. hirsutum L. reference genome; their global relationships inside the general metabolic pathways and the biosynthesis of secondary metabolites were visualized with iPath2. The chromatographic profiles of VOCs emissions indicated first monoterpenes and sesquiterpenes emissions, dominantly four molecules known to be involved in plant reactions to arthropod attacks. As a result, the study permitted to identify potential key genes for the emission of volatile terpenoids by cotton plants in reaction to an arthropod attack, opening possibilities for molecular-assisted cotton breeding in benefit of smallholder cotton growers.

Keywords: biosynthesis pathways, cotton, mechanisms of plant defense, terpenoids, volatile organic compounds

Procedia PDF Downloads 374
582 A Sentence-to-Sentence Relation Network for Recognizing Textual Entailment

Authors: Isaac K. E. Ampomah, Seong-Bae Park, Sang-Jo Lee

Abstract:

Over the past decade, there have been promising developments in Natural Language Processing (NLP) with several investigations of approaches focusing on Recognizing Textual Entailment (RTE). These models include models based on lexical similarities, models based on formal reasoning, and most recently deep neural models. In this paper, we present a sentence encoding model that exploits the sentence-to-sentence relation information for RTE. In terms of sentence modeling, Convolutional neural network (CNN) and recurrent neural networks (RNNs) adopt different approaches. RNNs are known to be well suited for sequence modeling, whilst CNN is suited for the extraction of n-gram features through the filters and can learn ranges of relations via the pooling mechanism. We combine the strength of RNN and CNN as stated above to present a unified model for the RTE task. Our model basically combines relation vectors computed from the phrasal representation of each sentence and final encoded sentence representations. Firstly, we pass each sentence through a convolutional layer to extract a sequence of higher-level phrase representation for each sentence from which the first relation vector is computed. Secondly, the phrasal representation of each sentence from the convolutional layer is fed into a Bidirectional Long Short Term Memory (Bi-LSTM) to obtain the final sentence representations from which a second relation vector is computed. The relations vectors are combined and then used in then used in the same fashion as attention mechanism over the Bi-LSTM outputs to yield the final sentence representations for the classification. Experiment on the Stanford Natural Language Inference (SNLI) corpus suggests that this is a promising technique for RTE.

Keywords: deep neural models, natural language inference, recognizing textual entailment (RTE), sentence-to-sentence relation

Procedia PDF Downloads 348