Search results for: road networks
871 Effective Supply Chain Coordination with Hybrid Demand Forecasting Techniques
Authors: Gurmail Singh
Abstract:
Effective supply chain is the main priority of every organization which is the outcome of strategic corporate investments with deliberate management action. Value-driven supply chain is defined through development, procurement and by configuring the appropriate resources, metrics and processes. However, responsiveness of the supply chain can be improved by proper coordination. So the Bullwhip effect (BWE) and Net stock amplification (NSAmp) values were anticipated and used for the control of inventory in organizations by both discrete wavelet transform-Artificial neural network (DWT-ANN) and Adaptive Network-based fuzzy inference system (ANFIS). This work presents a comparative methodology of forecasting for the customers demand which is non linear in nature for a multilevel supply chain structure using hybrid techniques such as Artificial intelligence techniques including Artificial neural networks (ANN) and Adaptive Network-based fuzzy inference system (ANFIS) and Discrete wavelet theory (DWT). The productiveness of these forecasting models are shown by computing the data from real world problems for Bullwhip effect and Net stock amplification. The results showed that these parameters were comparatively less in case of discrete wavelet transform-Artificial neural network (DWT-ANN) model and using Adaptive network-based fuzzy inference system (ANFIS).Keywords: bullwhip effect, hybrid techniques, net stock amplification, supply chain flexibility
Procedia PDF Downloads 127870 From Electroencephalogram to Epileptic Seizures Detection by Using Artificial Neural Networks
Authors: Gaetano Zazzaro, Angelo Martone, Roberto V. Montaquila, Luigi Pavone
Abstract:
Seizure is the main factor that affects the quality of life of epileptic patients. The diagnosis of epilepsy, and hence the identification of epileptogenic zone, is commonly made by using continuous Electroencephalogram (EEG) signal monitoring. Seizure identification on EEG signals is made manually by epileptologists and this process is usually very long and error prone. The aim of this paper is to describe an automated method able to detect seizures in EEG signals, using knowledge discovery in database process and data mining methods and algorithms, which can support physicians during the seizure detection process. Our detection method is based on Artificial Neural Network classifier, trained by applying the multilayer perceptron algorithm, and by using a software application, called Training Builder that has been developed for the massive extraction of features from EEG signals. This tool is able to cover all the data preparation steps ranging from signal processing to data analysis techniques, including the sliding window paradigm, the dimensionality reduction algorithms, information theory, and feature selection measures. The final model shows excellent performances, reaching an accuracy of over 99% during tests on data of a single patient retrieved from a publicly available EEG dataset.Keywords: artificial neural network, data mining, electroencephalogram, epilepsy, feature extraction, seizure detection, signal processing
Procedia PDF Downloads 188869 Using Nature-Based Solutions to Decarbonize Buildings in Canadian Cities
Authors: Zahra Jandaghian, Mehdi Ghobadi, Michal Bartko, Alex Hayes, Marianne Armstrong, Alexandra Thompson, Michael Lacasse
Abstract:
The Intergovernmental Panel on Climate Change (IPCC) report stated the urgent need to cut greenhouse gas emissions to avoid the adverse impacts of climatic changes. The United Nations has forecasted that nearly 70 percent of people will live in urban areas by 2050 resulting in a doubling of the global building stock. Given that buildings are currently recognised as emitting 40 percent of global carbon emissions, there is thus an urgent incentive to decarbonize existing buildings and to build net-zero carbon buildings. To attain net zero carbon emissions in communities in the future requires action in two directions: I) reduction of emissions; and II) removal of on-going emissions from the atmosphere once de-carbonization measures have been implemented. Nature-based solutions (NBS) have a significant role to play in achieving net zero carbon communities, spanning both emission reductions and removal of on-going emissions. NBS for the decarbonisation of buildings can be achieved by using green roofs and green walls – increasing vertical and horizontal vegetation on the building envelopes – and using nature-based materials that either emit less heat to the atmosphere thus decreasing photochemical reaction rates, or store substantial amount of carbon during the whole building service life within their structure. The NBS approach can also mitigate urban flooding and overheating, improve urban climate and air quality, and provide better living conditions for the urban population. For existing buildings, de-carbonization mostly requires retrofitting existing envelopes efficiently to use NBS techniques whereas for future construction, de-carbonization involves designing new buildings with low carbon materials as well as having the integrity and system capacity to effectively employ NBS. This paper presents the opportunities and challenges in respect to the de-carbonization of buildings using NBS for both building retrofits and new construction. This review documents the effectiveness of NBS to de-carbonize Canadian buildings, identifies the missing links to implement these techniques in cold climatic conditions, and determine a road map and immediate approaches to mitigate the adverse impacts of climate change such as urban heat islanding. Recommendations are drafted for possible inclusion in the Canadian building and energy codes.Keywords: decarbonization, nature-based solutions, GHG emissions, greenery enhancement, buildings
Procedia PDF Downloads 93868 Intelligent Agent-Based Model for the 5G mmWave O2I Technology Adoption
Authors: Robert Joseph M. Licup
Abstract:
The deployment of the fifth-generation (5G) mobile system through mmWave frequencies is the new solution in the requirement to provide higher bandwidth readily available for all users. The usage pattern of the mobile users has moved towards either the work from home or online classes set-up because of the pandemic. Previous mobile technologies can no longer meet the high speed, and bandwidth requirement needed, given the drastic shift of transactions to the home. The millimeter-wave (mmWave) underutilized frequency is utilized by the fifth-generation (5G) cellular networks that support multi-gigabit-per-second (Gbps) transmission. However, due to its short wavelengths, high path loss, directivity, blockage sensitivity, and narrow beamwidth are some of the technical challenges that need to be addressed. Different tools, technologies, and scenarios are explored to support network design, accurate channel modeling, implementation, and deployment effectively. However, there is a big challenge on how the consumer will adopt this solution and maximize the benefits offered by the 5G Technology. This research proposes to study the intricacies of technology diffusion, individual attitude, behaviors, and how technology adoption will be attained. The agent based simulation model shaped by the actual applications, technology solution, and related literature was used to arrive at a computational model. The research examines the different attributes, factors, and intricacies that can affect each identified agent towards technology adoption.Keywords: agent-based model, AnyLogic, 5G O21, 5G mmWave solutions, technology adoption
Procedia PDF Downloads 109867 Effectiveness of Opuntia ficus indica Cladodes Extract for Wound-Healing
Authors: Giuffrida Graziella, Pennisi Stefania, Coppa Federica, Iannello Giulia, Cartelli Simone, Lo Faro Riccardo, Ferruggia Greta, Brundo Maria Violetta
Abstract:
Cladode chemical composition may vary according to soil factors, cultivation season, and plant age. The primary metabolites of cladodes are water, carbohydrates, and proteins. The carbohydrates in cladodes are divided into two types: structural and storage. Polysaccharides from Opuntia ficus‐indica (L.) Mill plants build molecular networks with the capacity to retain water; thus, they act as mucoprotective agents. Mucilage is the main polysaccharide of cladodes; it contains polymers of β‐d‐galacturonic acid bound in positions (1–4) and traces of R‐linked l‐rhamnose (1-2). Mucilage regulates both the cell water content during prolonged drought and the calcium flux in the plant cells. The in vitro analysis of keratinocytes in monolayer, through the scratch-wound-healing assay, provided promising results. After 48 hours of exposure, the wound scratch was almost completely closed in cells treated with cladode extract. After 72 hours, the treated cells reached complete confluence, while in the untreated cells (negative control) the confluence was reached after 96 hours. We also added a positive control group of cells treated with colchicine, which inhibited wound closure for a more comprehensive analysis.Keywords: cladodes, metabolites, polysaccharide, scratch-wound-healing assay
Procedia PDF Downloads 54866 Reduce the Impact of Wildfires by Identifying Them Early from Space and Sending Location Directly to Closest First Responders
Authors: Gregory Sullivan
Abstract:
The evolution of global warming has escalated the number and complexity of forest fires around the world. As an example, the United States and Brazil combined generated more than 30,000 forest fires last year. The impact to our environment, structures and individuals is incalculable. The world has learned to try to take this in stride, trying multiple ways to contain fires. Some countries are trying to use cameras in limited areas. There are discussions of using hundreds of low earth orbit satellites and linking them together, and, interfacing them through ground networks. These are all truly noble attempts to defeat the forest fire phenomenon. But there is a better, simpler answer. A bigger piece of the solutions puzzle is to see the fires while they are small, soon after initiation. The approach is to see the fires while they are very small and report their location (latitude and longitude) to local first responders. This is done by placing a sensor at geostationary orbit (GEO: 26,000 miles above the earth). By placing this small satellite in GEO, we can “stare” at the earth, and sense temperature changes. We do not “see” fires, but “measure” temperature changes. This has already been demonstrated on an experimental scale. Fires were seen at close to initiation, and info forwarded to first responders. it were the first to identify the fires 7 out of 8 times. The goal is to have a small independent satellite at GEO orbit focused only on forest fire initiation. Thus, with one small satellite, focused only on forest fire initiation, we hope to greatly decrease the impact to persons, property and the environment.Keywords: space detection, wildfire early warning, demonstration wildfire detection and action from space, space detection to first responders
Procedia PDF Downloads 70865 The Use of Layered Neural Networks for Classifying Hierarchical Scientific Fields of Study
Authors: Colin Smith, Linsey S Passarella
Abstract:
Due to the proliferation and decentralized nature of academic publication, no widely accepted scheme exists for organizing papers by their scientific field of study (FoS) to the author’s best knowledge. While many academic journals require author provided keywords for papers, these keywords range wildly in scope and are not consistent across papers, journals, or field domains, necessitating alternative approaches to paper classification. Past attempts to perform field-of-study (FoS) classification on scientific texts have largely used a-hierarchical FoS schemas or ignored the schema’s inherently hierarchical structure, e.g. by compressing the structure into a single layer for multi-label classification. In this paper, we introduce an application of a Layered Neural Network (LNN) to the problem of performing supervised hierarchical classification of scientific fields of study (FoS) on research papers. In this approach, paper embeddings from a pretrained language model are fed into a top-down LNN. Beginning with a single neural network (NN) for the highest layer of the class hierarchy, each node uses a separate local NN to classify the subsequent subfield child node(s) for an input embedding of concatenated paper titles and abstracts. We compare our LNN-FOS method to other recent machine learning methods using the Microsoft Academic Graph (MAG) FoS hierarchy and find that the LNN-FOS offers increased classification accuracy at each FoS hierarchical level.Keywords: hierarchical classification, layer neural network, scientific field of study, scientific taxonomy
Procedia PDF Downloads 133864 Resilience of Infrastructure Networks: Maintenance of Bridges in Mountainous Environments
Authors: Lorenza Abbracciavento, Valerio De Biagi
Abstract:
Infrastructures are key elements to ensure the operational functionality of the transport system. The collapse of a single bridge or, equivalently, a tunnel can leads an entire motorway to be considered completely inaccessible. As a consequence, the paralysis of the communications network determines several important drawbacks for the community. Recent chronicle events have demonstrated that ensuring the functional continuity of the strategic infrastructures during and after a catastrophic event makes a significant difference in terms of life and economical losses. Moreover, it has been observed that RC structures located in mountain environments show a worst state of conservation compared to the same typology and aging structures located in temperate climates. Because of its morphology, in fact, the mountain environment is particularly exposed to severe collapse and deterioration phenomena, generally: natural hazards, e.g. rock falls, and meteorological hazards, e.g. freeze-thaw cycles or heavy snows. For these reasons, deep investigation on the characteristics of these processes becomes of fundamental importance to provide smart and sustainable solutions and make the infrastructure system more resilient. In this paper, the design of a monitoring system in mountainous environments is presented and analyzed in its parts. The method not only takes into account the peculiar climatic conditions, but it is integrated and interacts with the environment surrounding.Keywords: structural health monitoring, resilience of bridges, mountain infrastructures, infrastructural network, maintenance
Procedia PDF Downloads 77863 Elaboration and Physico-Chemical Characterization of Edible Films Made from Chitosan and Spray Dried Ethanolic Extracts of Propolis
Authors: David Guillermo Piedrahita Marquez, Hector Suarez Mahecha, Jairo Humberto Lopez
Abstract:
It was necessary to establish which formulation is suitable for the preservation of aquaculture products, that why edible films were made. These were to a characterization in order to meet their morphology physicochemical and mechanical properties, optical. Six Formulations of chitosan and propolis ethanolic extract encapsulated were developed because of their activity against pathogens and due to their properties, which allows the creation waterproof polymer networks against gasses, vapor, and physical damage. In the six Formulations, the concentration of comparison material (1% w/v, 2% pv) and the bioactive concentrations (0.5% w/v, 1% w/v, 1.5% pv) were changed and the results obtained were compared with statistical and multivariate analysis methods. It was observed that the matrices showed a mayor impermeability and thickness control samples and the samples reported in the literature. Also, these films showed a notorious uniformity of the films and a bigger resistance to the physical damage compared with other edible films made of other biopolymers. However the action of some compounds had a negative effect on the mechanical properties and changed drastically the optical properties, the bioactive has an effect on Polymer Matrix and it was determined that the films with 2% w / v of chitosan and 1.5% w/v encapsulated, exhibited the best properties and suffered to a lesser extent the negative impact of immiscible substances.Keywords: chitosan, edible films, ethanolic extract of propolis, mechanical properties, optical properties, physical characterization, scanning electron microscopy (SEM)
Procedia PDF Downloads 446862 An Efficient Subcarrier Scheduling Algorithm for Downlink OFDMA-Based Wireless Broadband Networks
Authors: Hassen Hamouda, Mohamed Ouwais Kabaou, Med Salim Bouhlel
Abstract:
The growth of wireless technology made opportunistic scheduling a widespread theme in recent research. Providing high system throughput without reducing fairness allocation is becoming a very challenging task. A suitable policy for resource allocation among users is of crucial importance. This study focuses on scheduling multiple streaming flows on the downlink of a WiMAX system based on orthogonal frequency division multiple access (OFDMA). In this paper, we take the first step in formulating and analyzing this problem scrupulously. As a result, we proposed a new scheduling scheme based on Round Robin (RR) Algorithm. Because of its non-opportunistic process, RR does not take in account radio conditions and consequently it affect both system throughput and multi-users diversity. Our contribution called MORRA (Modified Round Robin Opportunistic Algorithm) consists to propose a solution to this issue. MORRA not only exploits the concept of opportunistic scheduler but also takes into account other parameters in the allocation process. The first parameter is called courtesy coefficient (CC) and the second is called Buffer Occupancy (BO). Performance evaluation shows that this well-balanced scheme outperforms both RR and MaxSNR schedulers and demonstrate that choosing between system throughput and fairness is not required.Keywords: OFDMA, opportunistic scheduling, fairness hierarchy, courtesy coefficient, buffer occupancy
Procedia PDF Downloads 300861 Optimal Allocation of Oil Rents and Public Investment In Low-Income Developing Countries: A Computable General Equilibrium Analysis
Authors: Paule Olivia Akotto
Abstract:
The recent literature suggests spending between 50%-85% of oil rents. However, there are not yet clear guidelines for allocating this windfall in the public investment system, while most of the resource-rich countries fail to improve their intergenerational mobility. We study a design of the optimal spending system in Senegal, a low-income developing country featuring newly discovered oil fields and low intergenerational mobility. We build a dynamic general equilibrium model in which rural and urban (Dakar and other urban centers henceforth OUC) households face different health, education, and employment opportunities based on their location, affecting their intergenerational mobility. The model captures the relationship between oil rents, public investment, and multidimensional inequality of opportunity. The government invests oil rents in three broad sectors: health and education, road and industries, and agriculture. Through endogenous productivity externality and human capital accumulation, our model generates the predominant position of Dakar and OUC households in terms of access to health, education, and employment in line with Senegal data. Rural households are worse off in all dimensions. We compute the optimal spending policy under two sets of simulation scenarios. Under the current Senegal public investment strategy, which weighs more health and education investments, we find that the reform maximizing the decline in inequality of opportunity between households, frontloads investment during the first eight years of the oil exploitation and spends the perpetual value of oil wealth thereafter. We will then identify the marginal winners and losers associated with this policy and its redistributive implications. Under our second set of scenarios, we will test whether the Senegalese economy can reach better equality of opportunity outcomes under this frontloading reform, by allowing the sectoral shares of investment to vary. The trade-off will be between cutting human capital investment in favor of agricultural and productive infrastructure or increasing the former. We will characterize the optimal policy by specifying where the higher weight should be. We expect that the optimal policy of the second set strictly dominates in terms of equality of opportunity, the optimal policy computed under the current investment strategy. Finally, we will quantify this optimal policy's aggregate and distributional effects on poverty, well-being, and gender earning gaps.Keywords: developing countries, general equilibrium, inequality of opportunity, oil rents
Procedia PDF Downloads 237860 Transition from Linear to Circular Economy in Gypsum in India
Authors: Shanti Swaroop Gupta, Bibekananda Mohapatra, S. K. Chaturvedi, Anand Bohra
Abstract:
For sustainable development in India, there is an urgent need to follow the principles of industrial symbiosis in the industrial processes, under which the scraps, wastes, or by‐products of one industry can become the raw materials for another. This will not only help in reducing the dependence on natural resources but also help in gaining economic advantage to the industry. Gypsum is one such area in India, where the linear economy model of by-product gypsum utilization has resulted in unutilized legacy phosphogypsum stock of 64.65 million tonnes (mt) at phosphoric acid plants in 2020-21. In the future, this unutilized gypsum stock will increase further due to the expected generation of Flue Gas Desulphurization (FGD) gypsum in huge quantities from thermal power plants. Therefore, it is essential to transit from the linear to circular economy in Gypsum in India, which will result in huge environmental as well as ecological benefits. Gypsum is required in many sectors like Construction (Cement industry, gypsum boards, glass fiber reinforced gypsum panels, gypsum plaster, fly ash lime bricks, floor screeds, road construction), agriculture, in the manufacture of Plaster of Paris, pottery, ceramic industry, water treatment processes, manufacture of ammonium sulphate, paints, textiles, etc. The challenges faced in areas of quality, policy, logistics, lack of infrastructure, promotion, etc., for complete utilization of by-product gypsum have been discussed. The untapped potential of by-product gypsum utilization in various sectors like the use of gypsum in agriculture for sodic soil reclamation, utilization of legacy stock in cement industry on mission mode, improvement in quality of by-product gypsum by standardization and usage in building materials industry has been identified. Based on the measures required to tackle the various challenges and utilization of the untapped potential of gypsum, a comprehensive action plan for the transition from linear to the circular economy in gypsum in India has been formulated. The strategies and policy measures required to implement the action plan to achieve a circular economy in Gypsum have been recommended for various government departments. It is estimated that the focused implementation of the proposed action plan would result in a significant decrease in unutilized gypsum legacy stock in the next five years and it would cease to exist by 2027-28 if the proposed action plan is effectively implemented.Keywords: circular economy, FGD gypsum, India, phosphogypsum
Procedia PDF Downloads 268859 A Non-Destructive Estimation Method for Internal Time in Perilla Leaf Using Hyperspectral Data
Authors: Shogo Nagano, Yusuke Tanigaki, Hirokazu Fukuda
Abstract:
Vegetables harvested early in the morning or late in the afternoon are valued in plant production, and so the time of harvest is important. The biological functions known as circadian clocks have a significant effect on this harvest timing. The purpose of this study was to non-destructively estimate the circadian clock and so construct a method for determining a suitable harvest time. We took eight samples of green busil (Perilla frutescens var. crispa) every 4 hours, six times for 1 day and analyzed all samples at the same time. A hyperspectral camera was used to collect spectrum intensities at 141 different wavelengths (350–1050 nm). Calculation of correlations between spectrum intensity of each wavelength and harvest time suggested the suitability of the hyperspectral camera for non-destructive estimation. However, even the highest correlated wavelength had a weak correlation, so we used machine learning to raise the accuracy of estimation and constructed a machine learning model to estimate the internal time of the circadian clock. Artificial neural networks (ANN) were used for machine learning because this is an effective analysis method for large amounts of data. Using the estimation model resulted in an error between estimated and real times of 3 min. The estimations were made in less than 2 hours. Thus, we successfully demonstrated this method of non-destructively estimating internal time.Keywords: artificial neural network (ANN), circadian clock, green busil, hyperspectral camera, non-destructive evaluation
Procedia PDF Downloads 299858 Ontology Expansion via Synthetic Dataset Generation and Transformer-Based Concept Extraction
Authors: Andrey Khalov
Abstract:
The rapid proliferation of unstructured data in IT infrastructure management demands innovative approaches for extracting actionable knowledge. This paper presents a framework for ontology-based knowledge extraction that combines relational graph neural networks (R-GNN) with large language models (LLMs). The proposed method leverages the DOLCE framework as the foundational ontology, extending it with concepts from ITSMO for domain-specific applications in IT service management and outsourcing. A key component of this research is the use of transformer-based models, such as DeBERTa-v3-large, for automatic entity and relationship extraction from unstructured texts. Furthermore, the paper explores how transfer learning techniques can be applied to fine-tune large language models (LLaMA) for using to generate synthetic datasets to improve precision in BERT-based entity recognition and ontology alignment. The resulting IT Ontology (ITO) serves as a comprehensive knowledge base that integrates domain-specific insights from ITIL processes, enabling more efficient decision-making. Experimental results demonstrate significant improvements in knowledge extraction and relationship mapping, offering a cutting-edge solution for enhancing cognitive computing in IT service environments.Keywords: ontology expansion, synthetic dataset, transformer fine-tuning, concept extraction, DOLCE, BERT, taxonomy, LLM, NER
Procedia PDF Downloads 14857 The Need for Innovation Management in the Context of Integrated Management Systems
Authors: Adela Mariana Vadastreanu, Adrian Bot, Andreea Maier, Dorin Maier
Abstract:
This paper approaches the need for innovation management in the context of an existing integrated management system implemented in an organization. The road to success for companies in today’s economic environment is more demanding than ever and the capacity of adapting to the rapid changes is compensatory in order to resist on the market. The managers struggle, daily, with increasingly complex problems, caused by fierce competition in the market but also from the rising demands of customers. Innovation seems to be the solution for these problems. During the last decade almost all companies have been certificated according to various management systems, like quality management system, environmental management system, health and safety management system and others; furthermore many companies have implemented an integrated management system, by integrating two or more management systems. The problem rising today is how to integrate innovation in this integrated management systems. The challenge of the problem is that the development of an innovation management system is in the early phase. In this paper we have studied the possibility of integrating some of the innovation request in an existing management system, we have identify the innovation performance request and we proposed some recommendations regarding innovation management and its implementation as a part of an integrated management system. This paper lies down the bases for developing an model of integration management systems that include innovation as a main part of it. Organizations are becoming more aware of the importance of Integrated Management Systems (IMS). Integrating two or more management systems into an integrated management system can have much advantages.This paper examines various models of management systems integration in accordance with professional references ISO 9001, ISO 18001 and OHSAS 18001, highlighting strengths and weaknesses, creating a basis for future development of integrated management systems, and their involvement in various other processes within the organization, such as innovation management. The more and more demanding economic context emphasizes the awareness of the importance of innovation for organizations. This paper highlights the importance of the innovation for an organization and also gives some practical solution in order to improve the overall success of the business through a better approach of innovation. Various standards have been developed in order to certificate organizations that they respect the requirements. Applying an integrated standards model is shown to be a more effective way then applying the standards independently. The problem that arises is that in order to adopt the integrated version of standards there have to be made some changes at the organizational level. Every change that needs to be done has an effect on its activity, and in this sense the paper tries to deal with the changes needed for adopting an integrated management system and if those changes have an influence over the performance. After the analysis of the results, we can conclude that in order to improve the performance a necessary step is the implementation of innovation in the existing integrated management system.Keywords: innovation, integrated management systems, innovation management, quality
Procedia PDF Downloads 315856 Localization of Buried People Using Received Signal Strength Indication Measurement of Wireless Sensor
Authors: Feng Tao, Han Ye, Shaoyi Liao
Abstract:
City constructions collapse after earthquake and people will be buried under ruins. Search and rescue should be conducted as soon as possible to save them. Therefore, according to the complicated environment, irregular aftershocks and rescue allow of no delay, a kind of target localization method based on RSSI (Received Signal Strength Indication) is proposed in this article. The target localization technology based on RSSI with the features of low cost and low complexity has been widely applied to nodes localization in WSN (Wireless Sensor Networks). Based on the theory of RSSI transmission and the environment impact to RSSI, this article conducts the experiments in five scenes, and multiple filtering algorithms are applied to original RSSI value in order to establish the signal propagation model with minimum test error respectively. Target location can be calculated from the distance, which can be estimated from signal propagation model, through improved centroid algorithm. Result shows that the localization technology based on RSSI is suitable for large-scale nodes localization. Among filtering algorithms, mixed filtering algorithm (average of average, median and Gaussian filtering) performs better than any other single filtering algorithm, and by using the signal propagation model, the minimum error of distance between known nodes and target node in the five scene is about 3.06m.Keywords: signal propagation model, centroid algorithm, localization, mixed filtering, RSSI
Procedia PDF Downloads 300855 Cosmetic Surgery on the Rise: The Impact of Remote Communication
Authors: Bruno Di Pace, Roxanne H. Padley
Abstract:
Aims: The recent increase in remote video interaction has increased the number of requests for teleconsultations with plastic surgeons in private practice (70% in the UK and 64% in the USA). This study investigated the motivations for such an increase and the underlying psychological impact on patients. Method: An anonymous web-based poll of 8 questions was designed and distributed to patients seeking cosmetic surgery through social networks in both Italy and the UK. The questions gathered responses regarding 1. Reasons for pursuing cosmetic surgery; 2. The effects of delays caused by the SARS-COV-2 pandemic; 3. The effects on mood; 4. The influence of video conferencing on body-image perception. Results: 85 respondents completed the online poll. Overall, 68% of respondents stated that seeing themselves more frequently online had influenced their decision to seek cosmetic surgery. The types of surgeries indicated were predominantly to the upper body and face (82%). Delays and access to surgeons during the pandemic were perceived as negatively impacting patients' moods (95%). Body-image perception and self-esteem were lower than in the pre-pandemic, particularly during lockdown (72%). Patients were more inclined to undergo cosmetic surgery during the pandemic, both due to the wish to improve their “lockdown face” for video conferencing (77%) and also due to the benefits of home recovery while in smart working (58%). Conclusions: Overall, findings suggest that video conferencing has led to a significant increase in requests for cosmetic surgery and the so-called “Zoom Boom” effect.Keywords: cosmetic surgery, remote communication, telehealth, zoom boom
Procedia PDF Downloads 179854 A Bi-Objective Model to Optimize the Total Time and Idle Probability for Facility Location Problem Behaving as M/M/1/K Queues
Authors: Amirhossein Chambari
Abstract:
This article proposes a bi-objective model for the facility location problem subject to congestion (overcrowding). Motivated by implementations to locate servers in internet mirror sites, communication networks, one-server-systems, so on. This model consider for situations in which immobile (or fixed) service facilities are congested (or queued) by stochastic demand to behave as M/M/1/K queues. We consider for this problem two simultaneous perspectives; (1) Customers (desire to limit times of accessing and waiting for service) and (2) Service provider (desire to limit average facility idle-time). A bi-objective model is setup for facility location problem with two objective functions; (1) Minimizing sum of expected total traveling and waiting time (customers) and (2) Minimizing the average facility idle-time percentage (service provider). The proposed model belongs to the class of mixed-integer nonlinear programming models and the class of NP-hard problems. In addition, to solve the model, controlled elitist non-dominated sorting genetic algorithms (Controlled NSGA-II) and controlled elitist non-dominated ranking genetic algorithms (NRGA-I) are proposed. Furthermore, the two proposed metaheuristics algorithms are evaluated by establishing standard multiobjective metrics. Finally, the results are analyzed and some conclusions are given.Keywords: bi-objective, facility location, queueing, controlled NSGA-II, NRGA-I
Procedia PDF Downloads 583853 Redefining Identity of People with Disabilities Based on Content Analysis of Instagram Accounts
Authors: Grzegorz Kubinski
Abstract:
The proposed paper is focused on forms of identity expression in people with disabilities (PWD) in the social networks like Instagram. Theoretical analysis widely proposes using the new media as an assistive tool for improving wellbeing and labour activities of PWD. This kind of use is definitely important and plays a key role in all social inclusion processes. However, Instagram is not a place where PWD only express their own problems, but in the opposite, allows them to construct a new definition of disability. In the paper, the problem how this different than a classical approach to disability is created by PWD will be discussed. This issue will be scrutinized mainly in two points. Firstly, the question of how disability is changed by other everyday activities, like fashion or sport, will be described. Secondly, and this could be seen as more important, the point how PWD redefining their bodies creating a different form of aesthetic will be presented. The paper is based on content analysis of Instagram accounts. About 20 accounts created by PWD were analyzed for 6 month period, taking into account elements like photos, comments and discussions. All those information were studied in relation to 'everyday life' category and 'aesthetic' category. Works by T. Siebers, L. J. Davis or R. McRuer were used as theoretical background. Conclusions and interpretations presented in the proposed paper show that the Internet can be used by PWD not only as prosthetic and assistive tools. PWD willingly use them as modes of expression their independence, agency and identity. The paper proposes that in further research this way of using the Internet communication by PWD should be taken into account as an important part of the understanding of disability.Keywords: body, disability, identity, new media
Procedia PDF Downloads 138852 Machine Learning and Deep Learning Approach for People Recognition and Tracking in Crowd for Safety Monitoring
Authors: A. Degale Desta, Cheng Jian
Abstract:
Deep learning application in computer vision is rapidly advancing, giving it the ability to monitor the public and quickly identify potentially anomalous behaviour from crowd scenes. Therefore, the purpose of the current work is to improve the performance of safety of people in crowd events from panic behaviour through introducing the innovative idea of Aggregation of Ensembles (AOE), which makes use of the pre-trained ConvNets and a pool of classifiers to find anomalies in video data with packed scenes. According to the theory of algorithms that applied K-means, KNN, CNN, SVD, and Faster-CNN, YOLOv5 architectures learn different levels of semantic representation from crowd videos; the proposed approach leverages an ensemble of various fine-tuned convolutional neural networks (CNN), allowing for the extraction of enriched feature sets. In addition to the above algorithms, a long short-term memory neural network to forecast future feature values and a handmade feature that takes into consideration the peculiarities of the crowd to understand human behavior. On well-known datasets of panic situations, experiments are run to assess the effectiveness and precision of the suggested method. Results reveal that, compared to state-of-the-art methodologies, the system produces better and more promising results in terms of accuracy and processing speed.Keywords: action recognition, computer vision, crowd detecting and tracking, deep learning
Procedia PDF Downloads 161851 Understanding Cognitive Fatigue From FMRI Scans With Self-supervised Learning
Authors: Ashish Jaiswal, Ashwin Ramesh Babu, Mohammad Zaki Zadeh, Fillia Makedon, Glenn Wylie
Abstract:
Functional magnetic resonance imaging (fMRI) is a neuroimaging technique that records neural activations in the brain by capturing the blood oxygen level in different regions based on the task performed by a subject. Given fMRI data, the problem of predicting the state of cognitive fatigue in a person has not been investigated to its full extent. This paper proposes tackling this issue as a multi-class classification problem by dividing the state of cognitive fatigue into six different levels, ranging from no-fatigue to extreme fatigue conditions. We built a spatio-temporal model that uses convolutional neural networks (CNN) for spatial feature extraction and a long short-term memory (LSTM) network for temporal modeling of 4D fMRI scans. We also applied a self-supervised method called MoCo (Momentum Contrast) to pre-train our model on a public dataset BOLD5000 and fine-tuned it on our labeled dataset to predict cognitive fatigue. Our novel dataset contains fMRI scans from Traumatic Brain Injury (TBI) patients and healthy controls (HCs) while performing a series of N-back cognitive tasks. This method establishes a state-of-the-art technique to analyze cognitive fatigue from fMRI data and beats previous approaches to solve this problem.Keywords: fMRI, brain imaging, deep learning, self-supervised learning, contrastive learning, cognitive fatigue
Procedia PDF Downloads 190850 Develop a Conceptual Data Model of Geotechnical Risk Assessment in Underground Coal Mining Using a Cloud-Based Machine Learning Platform
Authors: Reza Mohammadzadeh
Abstract:
The major challenges in geotechnical engineering in underground spaces arise from uncertainties and different probabilities. The collection, collation, and collaboration of existing data to incorporate them in analysis and design for given prospect evaluation would be a reliable, practical problem solving method under uncertainty. Machine learning (ML) is a subfield of artificial intelligence in statistical science which applies different techniques (e.g., Regression, neural networks, support vector machines, decision trees, random forests, genetic programming, etc.) on data to automatically learn and improve from them without being explicitly programmed and make decisions and predictions. In this paper, a conceptual database schema of geotechnical risks in underground coal mining based on a cloud system architecture has been designed. A new approach of risk assessment using a three-dimensional risk matrix supported by the level of knowledge (LoK) has been proposed in this model. Subsequently, the model workflow methodology stages have been described. In order to train data and LoK models deployment, an ML platform has been implemented. IBM Watson Studio, as a leading data science tool and data-driven cloud integration ML platform, is employed in this study. As a Use case, a data set of geotechnical hazards and risk assessment in underground coal mining were prepared to demonstrate the performance of the model, and accordingly, the results have been outlined.Keywords: data model, geotechnical risks, machine learning, underground coal mining
Procedia PDF Downloads 274849 Enhancement of Underwater Haze Image with Edge Reveal Using Pixel Normalization
Authors: M. Dhana Lakshmi, S. Sakthivel Murugan
Abstract:
As light passes from source to observer in the water medium, it is scattered by the suspended particulate matter. This scattering effect will plague the captured images with non-uniform illumination, blurring details, halo artefacts, weak edges, etc. To overcome this, pixel normalization with an Amended Unsharp Mask (AUM) filter is proposed to enhance the degraded image. To validate the robustness of the proposed technique irrespective of atmospheric light, the considered datasets are collected on dual locations. For those images, the maxima and minima pixel intensity value is computed and normalized; then the AUM filter is applied to strengthen the blurred edges. Finally, the enhanced image is obtained with good illumination and contrast. Thus, the proposed technique removes the effect of scattering called de-hazing and restores the perceptual information with enhanced edge detail. Both qualitative and quantitative analyses are done on considering the standard non-reference metric called underwater image sharpness measure (UISM), and underwater image quality measure (UIQM) is used to measure color, sharpness, and contrast for both of the location images. It is observed that the proposed technique has shown overwhelming performance compared to other deep-based enhancement networks and traditional techniques in an adaptive manner.Keywords: underwater drone imagery, pixel normalization, thresholding, masking, unsharp mask filter
Procedia PDF Downloads 195848 R-Killer: An Email-Based Ransomware Protection Tool
Authors: B. Lokuketagoda, M. Weerakoon, U. Madushan, A. N. Senaratne, K. Y. Abeywardena
Abstract:
Ransomware has become a common threat in past few years and the recent threat reports show an increase of growth in Ransomware infections. Researchers have identified different variants of Ransomware families since 2015. Lack of knowledge of the user about the threat is a major concern. Ransomware detection methodologies are still growing through the industry. Email is the easiest method to send Ransomware to its victims. Uninformed users tend to click on links and attachments without much consideration assuming the emails are genuine. As a solution to this in this paper R-Killer Ransomware detection tool is introduced. Tool can be integrated with existing email services. The core detection Engine (CDE) discussed in the paper focuses on separating suspicious samples from emails and handling them until a decision is made regarding the suspicious mail. It has the capability of preventing execution of identified ransomware processes. On the other hand, Sandboxing and URL analyzing system has the capability of communication with public threat intelligence services to gather known threat intelligence. The R-Killer has its own mechanism developed in its Proactive Monitoring System (PMS) which can monitor the processes created by downloaded email attachments and identify potential Ransomware activities. R-killer is capable of gathering threat intelligence without exposing the user’s data to public threat intelligence services, hence protecting the confidentiality of user data.Keywords: ransomware, deep learning, recurrent neural networks, email, core detection engine
Procedia PDF Downloads 215847 A Bibliometric Analysis on Filter Bubble
Authors: Misbah Fatma, Anam Saiyeda
Abstract:
This analysis charts the introduction and expansion of research into the filter bubble phenomena over the last 10 years using a large dataset of academic publications. This bibliometric study demonstrates how interdisciplinary filter bubble research is. The identification of key authors and organizations leading the filter bubble study sheds information on collaborative networks and knowledge transfer. Relevant papers are organized based on themes including algorithmic bias, polarisation, social media, and ethical implications through a systematic examination of the literature. In order to shed light on how these patterns have changed over time, the study plots their historical history. The study also looks at how research is distributed globally, showing geographic patterns and discrepancies in scholarly output. The results of this bibliometric analysis let us fully comprehend the development and reach of filter bubble research. This study offers insights into the ongoing discussion surrounding information personalization and its implications for societal discourse, democratic participation, and the potential risks to an informed citizenry by exposing dominant themes, interdisciplinary collaborations, and geographic patterns. In order to solve the problems caused by filter bubbles and to advance a more diverse and inclusive information environment, this analysis is essential for scholars and researchers.Keywords: bibliometric analysis, social media, social networking, algorithmic personalization, self-selection, content moderation policies and limited access to information, recommender system and polarization
Procedia PDF Downloads 118846 Navigating Urban Childcare Challenges: Perspectives of Dhaka City Parents
Authors: Md. Shafiullah
Abstract:
This study delves into the evolving landscape of urban childcare in Bangladesh, focusing on the experiences and challenges faced by parents in Dhaka city. This paper argues that the traditional childcare arrangement of city families is inadequate to meet the development needs of children. The study aims to explore the childcare challenges faced by urban parents as they transition from traditional family-based childcare networks to alternative caregiving arrangements amidst urbanization, economic shifts, and social transformations. Utilizing a mixed-method research approach, combining quantitative surveys (n = 200) and four qualitative interviews, the research examines the parental viewpoints on childcare practices and the role of societal norms and values. The study finds childcare crises in both the family and daycare settings. In family care, caregiving suffers from the less availability of grandparents, a lack of skills of caregivers, and a lack of child interaction. As for the daycare, it is affected by the absence of appropriate policies, a lack of quality, health and safety concerns, affordability issues, and cultural concerns. Additionally, the study highlights inadequacies in childcare policies and regulatory frameworks, calling for comprehensive reforms to address the childcare vacuum in urban areas. By shifting the focus from developed to developing countries, this study contributes to the literature and suggests policy implications for Bangladesh and beyond.Keywords: childcare, child development, childcare policy, daycare, Bangladesh
Procedia PDF Downloads 56845 Estimation of Delay Due to Loading–Unloading of Passengers by Buses and Reduction of Number of Lanes at Selected Intersections in Dhaka City
Abstract:
One of the significant reasons that increase the delay time in the intersections at heterogeneous traffic condition is a sudden reduction of the capacity of the roads. In this study, the delay for this sudden capacity reduction is estimated. Two intersections at Dhaka city were brought in to thestudy, i.e., Kakrail intersection, and SAARC Foara intersection. At Kakrail intersection, the sudden reduction of capacity in the roads is seen at three downstream legs of the intersection, which are because of slowing down or stopping of buses for loading and unloading of passengers. At SAARC Foara intersection, sudden reduction of capacity was seen at two downstream legs. At one leg, it was due to loading and unloading of buses, and at another leg, it was for both loading and unloading of buses and reduction of the number of lanes. With these considerations, the delay due to intentional stoppage or slowing down of buses and reduction of the number of lanes for these two intersections are estimated. Here the delay was calculated by two approaches. The first approach came from the concept of shock waves in traffic streams. Here the delay was calculated by determining the flow, density, and speed before and after the sudden capacity reduction. The second approach came from the deterministic analysis of queues. Here the delay is calculated by determining the volume, capacity and reduced capacity of the road. After determining the delay from these two approaches, the results were compared. For this study, the video of each of the two intersections was recorded for one hour at the evening peak. Necessary geometric data were also taken to determine speed, flow, and density, etc. parameters. The delay was calculated for one hour with one-hour data at both intersections. In case of Kakrail intersection, the per hour delay for Kakrail circle leg was 5.79, and 7.15 minutes, for Shantinagar cross intersection leg they were 13.02 and 15.65 minutes, and for Paltan T intersection leg, they were 3 and 1.3 minutes for 1st and 2nd approaches respectively. In the case of SAARC Foara intersection, the delay at Shahbag leg was only due to intentional stopping or slowing down of busses, which were 3.2 and 3 minutes respectively for both approaches. For the Karwan Bazar leg, the delays for buses by both approaches were 5 and 7.5 minutes respectively, and for reduction of the number of lanes, the delays for both approaches were 2 and 1.78 minutes respectively. Measuring the delay per hour for the Kakrail leg at Kakrail circle, it is seen that, with consideration of the first approach of delay estimation, the intentional stoppage and lowering of speed by buses contribute to 26.24% of total delay at Kakrail circle. If the loading and unloading of buses at intersection is made forbidden near intersection, and any other measures for loading and unloading of passengers are established far enough from the intersections, then the delay at intersections can be reduced at significant scale, and the performance of the intersections can be enhanced.Keywords: delay, deterministic queue analysis, shock wave, passenger loading-unloading
Procedia PDF Downloads 178844 Managing City Pipe Leaks through Community Participation Using a Web and Mobile Application in South Africa
Authors: Mpai Mokoena, Nsenda Lukumwena
Abstract:
South Africa is one of the driest countries in the world and is facing a water crisis. In addition to inadequate infrastructure and poor planning, the country is experiencing high rates of water wastage due to pipe leaks. This study outlines the level of water wastage and develops a smart solution to efficiently manage and reduce the effects of pipe leaks, while monitoring the situation before and after fixing the pipe leaks. To understand the issue in depth, a literature review of journal papers and government reports was conducted. A questionnaire was designed and distributed to the general public. Additionally, the municipality office was contacted from a managerial perspective. The analysis from the study indicated that the majority of the citizens are aware of the water crisis and are willing to participate positively to decrease the level of water wasted. Furthermore, the response from the municipality acknowledged that more practical solutions are needed to reduce water wastage, and resources to attend to pipe leaks swiftly. Therefore, this paper proposes a specific solution for municipalities, local plumbers and citizens to minimize the effects of pipe leaks. The solution provides web and mobile application platforms to report and manage leaks swiftly. The solution is beneficial to the country in achieving water security and would promote a culture of responsibility toward water usage.Keywords: urban distribution networks, leak management, mobile application, responsible citizens, water crisis, water security
Procedia PDF Downloads 145843 Wolof Voice Response Recognition System: A Deep Learning Model for Wolof Audio Classification
Authors: Krishna Mohan Bathula, Fatou Bintou Loucoubar, FNU Kaleemunnisa, Christelle Scharff, Mark Anthony De Castro
Abstract:
Voice recognition algorithms such as automatic speech recognition and text-to-speech systems with African languages can play an important role in bridging the digital divide of Artificial Intelligence in Africa, contributing to the establishment of a fully inclusive information society. This paper proposes a Deep Learning model that can classify the user responses as inputs for an interactive voice response system. A dataset with Wolof language words ‘yes’ and ‘no’ is collected as audio recordings. A two stage Data Augmentation approach is adopted for enhancing the dataset size required by the deep neural network. Data preprocessing and feature engineering with Mel-Frequency Cepstral Coefficients are implemented. Convolutional Neural Networks (CNNs) have proven to be very powerful in image classification and are promising for audio processing when sounds are transformed into spectra. For performing voice response classification, the recordings are transformed into sound frequency feature spectra and then applied image classification methodology using a deep CNN model. The inference model of this trained and reusable Wolof voice response recognition system can be integrated with many applications associated with both web and mobile platforms.Keywords: automatic speech recognition, interactive voice response, voice response recognition, wolof word classification
Procedia PDF Downloads 116842 Load-Enabled Deployment and Sensing Range Optimization for Lifetime Enhancement of WSNs
Authors: Krishan P. Sharma, T. P. Sharma
Abstract:
Wireless sensor nodes are resource constrained battery powered devices usually deployed in hostile and ill-disposed areas to cooperatively monitor physical or environmental conditions. Due to their limited power supply, the major challenge for researchers is to utilize their battery power for enhancing the lifetime of whole network. Communication and sensing are two major sources of energy consumption in sensor networks. In this paper, we propose a deployment strategy for enhancing the average lifetime of a sensor network by effectively utilizing communication and sensing energy to provide full coverage. The proposed scheme is based on the fact that due to heavy relaying load, sensor nodes near to the sink drain energy at much faster rate than other nodes in the network and consequently die much earlier. To cover this imbalance, proposed scheme finds optimal communication and sensing ranges according to effective load at each node and uses a non-uniform deployment strategy where there is a comparatively high density of nodes near to the sink. Probable relaying load factor at particular node is calculated and accordingly optimal communication distance and sensing range for each sensor node is adjusted. Thus, sensor nodes are placed at locations that optimize energy during network operation. Formal mathematical analysis for calculating optimized locations is reported in present work.Keywords: load factor, network lifetime, non-uniform deployment, sensing range
Procedia PDF Downloads 383