Search results for: cohesion metrics
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 848

Search results for: cohesion metrics

278 Importance of Human Factors on Cybersecurity within Organizations: A Study of Attitudes and Behaviours

Authors: Elham Rajabian

Abstract:

The ascent of cybersecurity incidents is a rising threat to most organisations in general, while the impact of the incidents is unique to each of the organizations. It is a need for behavioural sciences to concentrate on employees’ behaviour in order to prepare key security mitigation opinions versus cybersecurity incidents. There are noticeable differences among users of a computer system in terms of complying with security behaviours. We can discuss the people's differences under several subjects such as delaying tactics on something that must be done, the tendency to act without thinking, future thinking about unexpected implications of present-day issues, and risk-taking behaviours in security policies compliance. In this article, we introduce high-profile cyber-attacks and their impacts on weakening cyber resiliency in organizations. We also give attention to human errors that influence network security. Human errors are discussed as a part of psychological matters to enhance compliance with the security policies. The organizational challenges are studied in order to shape a sustainable cyber risks management approach in the related work section. Insiders’ behaviours are viewed as a cyber security gap to draw proper cyber resiliency in section 3. We carry out the best cybersecurity practices by discussing four CIS challenges in section 4. In this regard, we provide a guideline and metrics to measure cyber resilience in organizations in section 5. In the end, we give some recommendations in order to build a cybersecurity culture based on individual behaviours.

Keywords: cyber resilience, human factors, cybersecurity behavior, attitude, usability, security culture

Procedia PDF Downloads 78
277 Improving Patient Journey in the Obstetrics and Gynecology Emergency Department: A Comprehensive Analysis of Patient Experience

Authors: Lolwa Alansari, Abdelhamid Azhaghdani, Sufia Athar, Hanen Mrabet, Annaliza Cruz, Tamara Alshadafat, Almunzer Zakaria

Abstract:

Introduction: Improving the patient experience is a fundamental pillar of healthcare's quadruple aims. Recognizing the importance of patient experiences and perceptions in healthcare interactions is pivotal for driving quality improvement. This abstract centers around the Patient Experience Program, an endeavor crafted with the purpose of comprehending and elevating the experiences of patients in the Obstetrics & Gynecology Emergency Department (OB/GYN ED). Methodology: This comprehensive endeavor unfolded through a structured sequence of phases following Plan-Do-Study-Act (PDSA) model, spanning over 12 months, focused on enhancing patient experiences in the Obstetrics & Gynecology Emergency Department (OB/GYN ED). The study meticulously examined the journeys of patients with acute obstetrics and gynecological conditions, collecting data from over 100 participants monthly. The inclusive approach covered patients of different priority levels (1-5) admitted for acute conditions, with no exclusions. Historical data from March and April 2022 serves as a benchmark for comparison, strengthening causality claims by providing a baseline understanding of OB/GYN ED performance before interventions. Additionally, the methodology includes the incorporation of staff engagement surveys to comprehensively understand the experiences of healthcare professionals with the implemented improvements. Data extraction involved administering open-ended questions and comment sections to gather rich qualitative insights. The survey covered various aspects of the patient journey, including communication, emotional support, timely access to care, care coordination, and patient-centered decision-making. The project's data analysis utilized a mixed-methods approach, combining qualitative techniques to identify recurring themes and extract actionable insights and quantitative methods to assess patient satisfaction scores and relevant metrics over time, facilitating the measurement of intervention impact and longitudinal tracking of changes. From the themes we discovered in both the online and in-person patient experience surveys, several key findings emerged that guided us in initiating improvements, including effective communication and information sharing, providing emotional support and empathy, ensuring timely access to care, fostering care coordination and continuity, and promoting patient-centered decision-making. Results: The project yielded substantial positive outcomes, significantly improving patient experiences in the OB/GYN ED. Patient satisfaction levels rose from 62% to a consistent 98%, with notable improvements in satisfaction with care plan information and physician care. Waiting time satisfaction increased from 68% to a steady 97%. The project positively impacted nurses' and midwives' job satisfaction, increasing from 64% to an impressive 94%. Operational metrics displayed positive trends, including a decrease in the "left without being seen" rate from 3% to 1%, the discharge against medical advice rate dropping from 8% to 1%, and the absconded rate reducing from 3% to 0%. These outcomes underscore the project's effectiveness in enhancing both patient and staff experiences in the healthcare setting. Conclusion: The use of a patient experience questionnaire has been substantiated by evidence-based research as an effective tool for improving the patient experience, guiding interventions, and enhancing overall healthcare quality in the OB/GYN ED. The project's interventions have resulted in a more efficient allocation of resources, reduced hospital stays, and minimized unnecessary resource utilization. This, in turn, contributes to cost savings for the healthcare facility.

Keywords: patient experience, patient survey, person centered care, quality initiatives

Procedia PDF Downloads 46
276 The Organizational Structure, Development Features, and Metadiscoursal Elements in the Expository Writing of College Freshman Students

Authors: Lota Largavista

Abstract:

This study entitled, ‘The Organizational Structure, Development Features, and Metadiscoursal Elements in the Expository Writing of Freshman College Writers’ aimed to examine essays written by college students. It seeks to examine the organizational structure and development features of the essays and describe their defining characteristics, the linguistic elements at both macrostructural and microstructural discourse levels and the types of textual and interpersonal metadiscourse markers that are employed in order to negotiate meanings with their prospective readers. The different frameworks used to analyze the essays include Toulmin’s ( 1984) model for argument structure, Olson’s ( 2003) three-part essay structure; Halliday and Matthiesen (2004) in Herriman (2011) notions of thematic structure, Danes (1974) thematic progression or method of development, Halliday’s (2004) concept of grammatical and lexical cohesion ;Hyland’s (2005) metadiscourse strategies; and Chung and Nation’s( 2003) four-step scale for technical vocabulary. This descriptive study analyzes qualitatively and quantitatively how freshman students generally express their written compositions. Coding of units is done to determine what linguistic features are present in the essays. Findings revealed that students’ expository essays observe a three-part structure having all three moves, the Introduction, the Body and the Conclusion. Stance assertion, stance support, and emerging moves/strategies are found to be employed in the essays. Students have more marked themes on the essays and also prefer constant theme progression as their method of development. The analysis of salient linguistic elements reveals frequently used cohesive devices and metadiscoursal strategies. Based on the findings, an instructional learning plan is being proposed. This plan is characterized by a genre approach that focuses on expository and linguistic conventions.

Keywords: metadiscourse, organization, theme progression, structure

Procedia PDF Downloads 214
275 Utilizing Computational Fluid Dynamics in the Analysis of Natural Ventilation in Buildings

Authors: A. W. J. Wong, I. H. Ibrahim

Abstract:

Increasing urbanisation has driven building designers to incorporate natural ventilation in the designs of sustainable buildings. This project utilises Computational Fluid Dynamics (CFD) to investigate the natural ventilation of an academic building, SIT@SP, using an assessment criterion based on daily mean temperature and mean velocity. The areas of interest are the pedestrian level of first and fourth levels of the building. A reference case recommended by the Architectural Institute of Japan was used to validate the simulation model. The validated simulation model was then used for coupled simulations on SIT@SP and neighbouring geometries, under two wind speeds. Both steady and transient simulations were used to identify differences in results. Steady and transient results are agreeable with the transient simulation identifying peak velocities during flow development. Under a lower wind speed, the first level was sufficiently ventilated while the fourth level was not. The first level has excessive wind velocities in the higher wind speed and the fourth level was adequately ventilated. Fourth level flow velocity was consistently lower than those of the first level. This is attributed to either simulation model error or poor building design. SIT@SP is concluded to have a sufficiently ventilated first level and insufficiently ventilated fourth level. Future works for this project extend to modifying the urban geometry, simulation model improvements, evaluation using other assessment metrics and extending the area of interest to the entire building.

Keywords: buildings, CFD Simulations, natural ventilation, urban airflow

Procedia PDF Downloads 204
274 Performance Evaluation of Soft RoCE over 1 Gigabit Ethernet

Authors: Gurkirat Kaur, Manoj Kumar, Manju Bala

Abstract:

Ethernet is the most influential and widely used technology in the world. With the growing demand of low latency and high throughput technologies like InfiniBand and RoCE, unique features viz. RDMA (Remote Direct Memory Access) have evolved. RDMA is an effective technology which is used for reducing system load and improving performance. InfiniBand is a well known technology which provides high-bandwidth and low-latency and makes optimal use of in-built features like RDMA. With the rapid evolution of InfiniBand technology and Ethernet lacking the RDMA and zero copy protocol, the Ethernet community has came out with a new enhancements that bridges the gap between InfiniBand and Ethernet. By adding the RDMA and zero copy protocol to the Ethernet a new networking technology is evolved, called RDMA over Converged Ethernet (RoCE). RoCE is a standard released by the IBTA standardization body to define RDMA protocol over Ethernet. With the emergence of lossless Ethernet, RoCE uses InfiniBand’s efficient transport to provide the platform for deploying RDMA technology in mainstream data centres over 10GigE, 40GigE and beyond. RoCE provide all of the InfiniBand benefits transport benefits and well established RDMA ecosystem combined with converged Ethernet. In this paper, we evaluate the heterogeneous Linux cluster, having multi nodes with fast interconnects i.e. gigabit Ethernet and Soft RoCE. This paper presents the heterogeneous Linux cluster configuration and evaluates its performance using Intel’s MPI Benchmarks. Our result shows that Soft RoCE is performing better than Ethernet in various performance metrics like bandwidth, latency and throughput.

Keywords: ethernet, InfiniBand, RoCE, RDMA, MPI, Soft RoCE

Procedia PDF Downloads 444
273 Socio-Economic Impact of Covid-19 in Ethiopia

Authors: Kebron Abich Asnake

Abstract:

The outbreak of COVID-19 has had far-reaching socio-economic consequences globally, and Ethiopia is no exception. This abstract provides a summary of a research study on the socio-economic impact of COVID-19 in Ethiopia. The study analyzes the health impact, economic repercussions, social consequences, government response measures, and opportunities for post-crisis recovery. In terms of health impact, the research explores the spread and transmission of the virus, the capacity and response of the healthcare system, and the mortality rate, with a focus on vulnerable populations. The economic impact analysis entails investigating the contraction of the GDP, employment and income loss, disruption in key sectors such as agriculture, tourism, and manufacturing, and the specific implications for small and medium-sized enterprises (SMEs), foreign direct investment, and remittances. The social impact section looks at the disruptions in education and the digital divide, food security and nutrition challenges, increased poverty and inequality, gender-based violence, and mental health issues. The research also examines the measures taken by the Ethiopian government, including health and safety regulations, economic stimulus packages, social protection programs, and support for vulnerable populations. Furthermore, the study outlines long-term recovery prospects, social cohesion, and community resilience challenges. It highlights the need to strengthen the healthcare system and finds a balance between health and economic priorities. The research concludes by presenting recommendations for policy-makers and stakeholders, emphasizing opportunities for post-crisis recovery such as diversification of the economy, enhanced healthcare infrastructure, investment in digital infrastructure and technology, and support for domestic tourism and local industries. This research provides valuable insights into the socio-economic impact of COVID-19 in Ethiopia, offering a comprehensive analysis of the challenges faced and potential pathways towards recovery.

Keywords: impact, covid, ethiopia, health

Procedia PDF Downloads 59
272 Progress in Combining Image Captioning and Visual Question Answering Tasks

Authors: Prathiksha Kamath, Pratibha Jamkhandi, Prateek Ghanti, Priyanshu Gupta, M. Lakshmi Neelima

Abstract:

Combining Image Captioning and Visual Question Answering (VQA) tasks have emerged as a new and exciting research area. The image captioning task involves generating a textual description that summarizes the content of the image. VQA aims to answer a natural language question about the image. Both these tasks include computer vision and natural language processing (NLP) and require a deep understanding of the content of the image and semantic relationship within the image and the ability to generate a response in natural language. There has been remarkable growth in both these tasks with rapid advancement in deep learning. In this paper, we present a comprehensive review of recent progress in combining image captioning and visual question-answering (VQA) tasks. We first discuss both image captioning and VQA tasks individually and then the various ways in which both these tasks can be integrated. We also analyze the challenges associated with these tasks and ways to overcome them. We finally discuss the various datasets and evaluation metrics used in these tasks. This paper concludes with the need for generating captions based on the context and captions that are able to answer the most likely asked questions about the image so as to aid the VQA task. Overall, this review highlights the significant progress made in combining image captioning and VQA, as well as the ongoing challenges and opportunities for further research in this exciting and rapidly evolving field, which has the potential to improve the performance of real-world applications such as autonomous vehicles, robotics, and image search.

Keywords: image captioning, visual question answering, deep learning, natural language processing

Procedia PDF Downloads 58
271 Refined Edge Detection Network

Authors: Omar Elharrouss, Youssef Hmamouche, Assia Kamal Idrissi, Btissam El Khamlichi, Amal El Fallah-Seghrouchni

Abstract:

Edge detection is represented as one of the most challenging tasks in computer vision, due to the complexity of detecting the edges or boundaries in real-world images that contains objects of different types and scales like trees, building as well as various backgrounds. Edge detection is represented also as a key task for many computer vision applications. Using a set of backbones as well as attention modules, deep-learning-based methods improved the detection of edges compared with the traditional methods like Sobel and Canny. However, images of complex scenes still represent a challenge for these methods. Also, the detected edges using the existing approaches suffer from non-refined results while the image output contains many erroneous edges. To overcome this, n this paper, by using the mechanism of residual learning, a refined edge detection network is proposed (RED-Net). By maintaining the high resolution of edges during the training process, and conserving the resolution of the edge image during the network stage, we make the pooling outputs at each stage connected with the output of the previous layer. Also, after each layer, we use an affined batch normalization layer as an erosion operation for the homogeneous region in the image. The proposed methods are evaluated using the most challenging datasets including BSDS500, NYUD, and Multicue. The obtained results outperform the designed edge detection networks in terms of performance metrics and quality of output images.

Keywords: edge detection, convolutional neural networks, deep learning, scale-representation, backbone

Procedia PDF Downloads 87
270 Personalized Social Resource Recommender Systems on Interest-Based Social Networks

Authors: C. L. Huang, J. J. Sia

Abstract:

The interest-based social networks, also known as social bookmark sharing systems, are useful platforms for people to conveniently read and collect internet resources. These platforms also providing function of social networks, and users can share and explore internet resources from the social networks. Providing personalized internet resources to users is an important issue on these platforms. This study uses two types of relationship on the social networks—following and follower and proposes a collaborative recommender system, consisting of two main steps. First, this study calculates the relationship strength between the target user and the target user's followings and followers to find top-N similar neighbors. Second, from the top-N similar neighbors, the articles (internet resources) that may interest the target user are recommended to the target user. In this system, users can efficiently obtain recent, related and diverse internet resources (knowledge) from the interest-based social network. This study collected the experimental dataset from Diigo, which is a famous bookmark sharing system. The experimental results show that the proposed recommendation model is more accurate than two traditional baseline recommendation models but slightly lower than the cosine model in accuracy. However, in the metrics of the diversity and executing time, our proposed model outperforms the cosine model.

Keywords: recommender systems, social networks, tagging, bookmark sharing systems, collaborative recommender systems, knowledge management

Procedia PDF Downloads 150
269 Time and Cost Prediction Models for Language Classification Over a Large Corpus on Spark

Authors: Jairson Barbosa Rodrigues, Paulo Romero Martins Maciel, Germano Crispim Vasconcelos

Abstract:

This paper presents an investigation of the performance impacts regarding the variation of five factors (input data size, node number, cores, memory, and disks) when applying a distributed implementation of Naïve Bayes for text classification of a large Corpus on the Spark big data processing framework. Problem: The algorithm's performance depends on multiple factors, and knowing before-hand the effects of each factor becomes especially critical as hardware is priced by time slice in cloud environments. Objectives: To explain the functional relationship between factors and performance and to develop linear predictor models for time and cost. Methods: the solid statistical principles of Design of Experiments (DoE), particularly the randomized two-level fractional factorial design with replications. This research involved 48 real clusters with different hardware arrangements. The metrics were analyzed using linear models for screening, ranking, and measurement of each factor's impact. Results: Our findings include prediction models and show some non-intuitive results about the small influence of cores and the neutrality of memory and disks on total execution time, and the non-significant impact of data input scale on costs, although notably impacts the execution time.

Keywords: big data, design of experiments, distributed machine learning, natural language processing, spark

Procedia PDF Downloads 96
268 Comprehensive Feature Extraction for Optimized Condition Assessment of Fuel Pumps

Authors: Ugochukwu Ejike Akpudo, Jank-Wook Hur

Abstract:

The increasing demand for improved productivity, maintainability, and reliability has prompted rapidly increasing research studies on the emerging condition-based maintenance concept- Prognostics and health management (PHM). Varieties of fuel pumps serve critical functions in several hydraulic systems; hence, their failure can have daunting effects on productivity, safety, etc. The need for condition monitoring and assessment of these pumps cannot be overemphasized, and this has led to the uproar in research studies on standard feature extraction techniques for optimized condition assessment of fuel pumps. By extracting time-based, frequency-based and the more robust time-frequency based features from these vibrational signals, a more comprehensive feature assessment (and selection) can be achieved for a more accurate and reliable condition assessment of these pumps. With the aid of emerging deep classification and regression algorithms like the locally linear embedding (LLE), we propose a method for comprehensive condition assessment of electromagnetic fuel pumps (EMFPs). Results show that the LLE as a comprehensive feature extraction technique yields better feature fusion/dimensionality reduction results for condition assessment of EMFPs against the use of single features. Also, unlike other feature fusion techniques, its capabilities as a fault classification technique were explored, and the results show an acceptable accuracy level using standard performance metrics for evaluation.

Keywords: electromagnetic fuel pumps, comprehensive feature extraction, condition assessment, locally linear embedding, feature fusion

Procedia PDF Downloads 104
267 A Medical Vulnerability Scoring System Incorporating Health and Data Sensitivity Metrics

Authors: Nadir A. Carreon, Christa Sonderer, Aakarsh Rao, Roman Lysecky

Abstract:

With the advent of complex software and increased connectivity, the security of life-critical medical devices is becoming an increasing concern, particularly with their direct impact on human safety. Security is essential, but it is impossible to develop completely secure and impenetrable systems at design time. Therefore, it is important to assess the potential impact on the security and safety of exploiting a vulnerability in such critical medical systems. The common vulnerability scoring system (CVSS) calculates the severity of exploitable vulnerabilities. However, for medical devices it does not consider the unique challenges of impacts to human health and privacy. Thus, the scoring of a medical device on which human life depends (e.g., pacemakers, insulin pumps) can score very low, while a system on which human life does not depend (e.g., hospital archiving systems) might score very high. In this paper, we propose a medical vulnerability scoring system (MVSS) that extends CVSS to address the health and privacy concerns of medical devices. We propose incorporating two new parameters, namely health impact, and sensitivity impact. Sensitivity refers to the type of information that can be stolen from the device, and health represents the impact on the safety of the patient if the vulnerability is exploited (e.g., potential harm, life-threatening). We evaluate fifteen different known vulnerabilities in medical devices and compare MVSS against two state-of-the-art medical device-oriented vulnerability scoring systems and the foundational CVSS.

Keywords: common vulnerability system, medical devices, medical device security, vulnerabilities

Procedia PDF Downloads 142
266 Highway Lighting of the 21st Century is Smart, but is it Cost Efficient?

Authors: Saurabh Gupta, Vanshdeep Parmar, Sri Harsha Reddy Yelly, Michele Baker, Elizabeth Bigler, Kunhee Choi

Abstract:

It is known that the adoption of solar powered LED highway lighting systems or sensory LED highway lighting systems can dramatically reduce energy consumption by 55 percent when compared to conventional on-grid High Pressure Sodium (HPS) lamps that are widely applied to most highways. However, an initial high installation cost for building the infrastructure of solar photovoltaic devices hampers a wider adoption of such technologies. This research aims to examine currently available state-of-the-art solar photovoltaic and sensory technologies, identify major obstacles, and analyze each technology to create a benchmarking metrics from the benefit-cost analysis perspective. The on-grid HPS lighting systems will serve as the baseline for this study to compare it with other lighting alternatives such as solar and sensory LED lighting systems. This research will test the validity of the research hypothesis that alternative LED lighting systems produce more favorable benefit-cost ratios and the added initial investment costs are recouped by the savings in the operation and maintenance cost. The payback period of the excess investment and projected savings over the life-cycle of the selected lighting systems will be analyzed by utilizing the concept of Net Present Value (NPV). Researchers believe that if this study validates the research hypothesis, it can promote a wider adoption of alternative lighting systems that will eventually save millions of taxpayer dollars in the long-run.

Keywords: lighting systems, sensory and solar PV, benefit cost analysis, net present value

Procedia PDF Downloads 335
265 Improving Similarity Search Using Clustered Data

Authors: Deokho Kim, Wonwoo Lee, Jaewoong Lee, Teresa Ng, Gun-Ill Lee, Jiwon Jeong

Abstract:

This paper presents a method for improving object search accuracy using a deep learning model. A major limitation to provide accurate similarity with deep learning is the requirement of huge amount of data for training pairwise similarity scores (metrics), which is impractical to collect. Thus, similarity scores are usually trained with a relatively small dataset, which comes from a different domain, causing limited accuracy on measuring similarity. For this reason, this paper proposes a deep learning model that can be trained with a significantly small amount of data, a clustered data which of each cluster contains a set of visually similar images. In order to measure similarity distance with the proposed method, visual features of two images are extracted from intermediate layers of a convolutional neural network with various pooling methods, and the network is trained with pairwise similarity scores which is defined zero for images in identical cluster. The proposed method outperforms the state-of-the-art object similarity scoring techniques on evaluation for finding exact items. The proposed method achieves 86.5% of accuracy compared to the accuracy of the state-of-the-art technique, which is 59.9%. That is, an exact item can be found among four retrieved images with an accuracy of 86.5%, and the rest can possibly be similar products more than the accuracy. Therefore, the proposed method can greatly reduce the amount of training data with an order of magnitude as well as providing a reliable similarity metric.

Keywords: visual search, deep learning, convolutional neural network, machine learning

Procedia PDF Downloads 198
264 Evolving Urban Landscapes: Smart Cities and Sustainable Futures

Authors: Mehrzad Soltani, Pegah Rezaei

Abstract:

In response to the escalating challenges posed by resource scarcity, urban congestion, and the dearth of green spaces, contemporary urban areas have undergone a remarkable transformation into smart cities. This evolution necessitates a strategic and forward-thinking approach to urban development, with the primary objective of diminishing and eventually eradicating dependence on non-renewable energy sources. This steadfast commitment to sustainable development is geared toward the continual enhancement of our global urban milieu, ensuring a healthier and more prosperous environment for forthcoming generations. This transformative vision has been meticulously shaped by an extensive research framework, incorporating in-depth field studies and investigations conducted at both neighborhood and city levels. Our holistic strategy extends its purview to encompass major cities and states, advocating for the realization of exceptional development firmly rooted in the principles of sustainable intelligence. At its core, this approach places a paramount emphasis on stringent pollution control measures, concurrently safeguarding ecological equilibrium and regional cohesion. Central to the realization of this vision is the widespread adoption of environmentally friendly materials and components, championing the cultivation of plant life and harmonious green spaces, and the seamless integration of intelligent lighting and irrigation systems. These systems, including solar panels and solar energy utilization, are deployed wherever feasible, effectively meeting the essential lighting and irrigation needs of these dynamic urban ecosystems. Overall, the transformation of urban areas into smart cities necessitates a holistic and innovative approach to urban development. By actively embracing sustainable intelligence and adhering to strict environmental standards, these cities pave the way for a brighter and more sustainable future, one that is marked by resilient, thriving, and eco-conscious urban communities.

Keywords: smart city, green urban, sustainability, urban management

Procedia PDF Downloads 57
263 Optimization of the Mechanical Performance of Fused Filament Fabrication Parts

Authors: Iván Rivet, Narges Dialami, Miguel Cervera, Michele Chiumenti

Abstract:

Process parameters in Additive Manufacturing (AM) play a critical role in the mechanical performance of the final component. In order to find the input configuration that guarantees the optimal performance of the printed part, the process-performance relationship must be found. Fused Filament Fabrication (FFF) is the selected demonstrative AM technology due to its great popularity in the industrial manufacturing world. A material model that considers the different printing patterns present in a FFF part is used. A voxelized mesh is built from the manufacturing toolpaths described in the G-Code file. An Adaptive Mesh Refinement (AMR) based on the octree strategy is used in order to reduce the complexity of the mesh while maintaining its accuracy. High-fidelity and cost-efficient Finite Element (FE) simulations are performed and the influence of key process parameters in the mechanical performance of the component is analyzed. A robust optimization process based on appropriate failure criteria is developed to find the printing direction that leads to the optimal mechanical performance of the component. The Tsai-Wu failure criterion is implemented due to the orthotropy and heterogeneity constitutive nature of FFF components and because of the differences between the strengths in tension and compression. The optimization loop implements a modified version of an Anomaly Detection (AD) algorithm and uses the computed metrics to obtain the optimal printing direction. The developed methodology is verified with a case study on an industrial demonstrator.

Keywords: additive manufacturing, optimization, printing direction, mechanical performance, voxelization

Procedia PDF Downloads 42
262 Optimization of Fused Deposition Modeling 3D Printing Process via Preprocess Calibration Routine Using Low-Cost Thermal Sensing

Authors: Raz Flieshman, Adam Michael Altenbuchner, Jörg Krüger

Abstract:

This paper presents an approach to optimizing the Fused Deposition Modeling (FDM) 3D printing process through a preprocess calibration routine of printing parameters. The core of this method involves the use of a low-cost thermal sensor capable of measuring tempera-tures within the range of -20 to 500 degrees Celsius for detailed process observation. The calibration process is conducted by printing a predetermined path while varying the process parameters through machine instructions (g-code). This enables the extraction of critical thermal, dimensional, and surface properties along the printed path. The calibration routine utilizes computer vision models to extract features and metrics from the thermal images, in-cluding temperature distribution, layer adhesion quality, surface roughness, and dimension-al accuracy and consistency. These extracted properties are then analyzed to optimize the process parameters to achieve the desired qualities of the printed material. A significant benefit of this calibration method is its potential to create printing parameter profiles for new polymer and composite materials, thereby enhancing the versatility and application range of FDM 3D printing. The proposed method demonstrates significant potential in enhancing the precision and reliability of FDM 3D printing, making it a valuable contribution to the field of additive manufacturing.

Keywords: FDM 3D printing, preprocess calibration, thermal sensor, process optimization, additive manufacturing, computer vision, material profiles

Procedia PDF Downloads 11
261 Downscaling Seasonal Sea Surface Temperature Forecasts over the Mediterranean Sea Using Deep Learning

Authors: Redouane Larbi Boufeniza, Jing-Jia Luo

Abstract:

This study assesses the suitability of deep learning (DL) for downscaling sea surface temperature (SST) over the Mediterranean Sea in the context of seasonal forecasting. We design a set of experiments that compare different DL configurations and deploy the best-performing architecture to downscale one-month lead forecasts of June–September (JJAS) SST from the Nanjing University of Information Science and Technology Climate Forecast System version 1.0 (NUIST-CFS1.0) for the period of 1982–2020. We have also introduced predictors over a larger area to include information about the main large-scale circulations that drive SST over the Mediterranean Sea region, which improves the downscaling results. Finally, we validate the raw model and downscaled forecasts in terms of both deterministic and probabilistic verification metrics, as well as their ability to reproduce the observed precipitation extreme and spell indicator indices. The results showed that the convolutional neural network (CNN)-based downscaling consistently improves the raw model forecasts, with lower bias and more accurate representations of the observed mean and extreme SST spatial patterns. Besides, the CNN-based downscaling yields a much more accurate forecast of extreme SST and spell indicators and reduces the significant relevant biases exhibited by the raw model predictions. Moreover, our results show that the CNN-based downscaling yields better skill scores than the raw model forecasts over most portions of the Mediterranean Sea. The results demonstrate the potential usefulness of CNN in downscaling seasonal SST predictions over the Mediterranean Sea, particularly in providing improved forecast products.

Keywords: Mediterranean Sea, sea surface temperature, seasonal forecasting, downscaling, deep learning

Procedia PDF Downloads 59
260 Normalized Enterprises Architectures: Portugal's Public Procurement System Application

Authors: Tiago Sampaio, André Vasconcelos, Bruno Fragoso

Abstract:

The Normalized Systems Theory, which is designed to be applied to software architectures, provides a set of theorems, elements and rules, with the purpose of enabling evolution in Information Systems, as well as ensuring that they are ready for change. In order to make that possible, this work’s solution is to apply the Normalized Systems Theory to the domain of enterprise architectures, using Archimate. This application is achieved through the adaptation of the elements of this theory, making them artifacts of the modeling language. The theorems are applied through the identification of the viewpoints to be used in the architectures, as well as the transformation of the theory’s encapsulation rules into architectural rules. This way, it is possible to create normalized enterprise architectures, thus fulfilling the needs and requirements of the business. This solution was demonstrated using the Portuguese Public Procurement System. The Portuguese government aims to make this system as fair as possible, allowing every organization to have the same business opportunities. The aim is for every economic operator to have access to all public tenders, which are published in any of the 6 existing platforms, independently of where they are registered. In order to make this possible, we applied our solution to the construction of two different architectures, which are able of fulfilling the requirements of the Portuguese government. One of those architectures, TO-BE A, has a Message Broker that performs the communication between the platforms. The other, TO-BE B, represents the scenario in which the platforms communicate with each other directly. Apart from these 2 architectures, we also represent the AS-IS architecture that demonstrates the current behavior of the Public Procurement Systems. Our evaluation is based on a comparison between the AS-IS and the TO-BE architectures, regarding the fulfillment of the rules and theorems of the Normalized Systems Theory and some quality metrics.

Keywords: archimate, architecture, broker, enterprise, evolvable systems, interoperability, normalized architectures, normalized systems, normalized systems theory, platforms

Procedia PDF Downloads 337
259 Relationship Between Health Coverage and Emergency Disease Burden

Authors: Karim Hajjar, Luis Lillo, Diego Martinez, Manuel Hermosilla, Nicholas Risko

Abstract:

Objectives: This study examines the relationship between universal health coverage (UCH) and the burden of emergency diseases at a global level. Methods: Data on Disability-Adjusted Life Years (DALYs) from emergency conditions were extracted from the Institute for Health Metrics and Evaluation (IHME) database for the years 2015 and 2019. Data on UHC, measured using two variables, 1) coverage of essential health services and 2) proportion of population spending more than 10% of household income on out-of-pocket health care expenditure, was extracted from the World Bank Database for years preceding our outcome of interest. Linear regression was performed, analyzing the effect of the UHC variables on the DALYs of emergency diseases, controlling for other variables. Results: A total of 133 countries were included. 44.4% of the analyzed countries had coverage of essential health services index of at least 70/100, and 35.3% had at least 10% of their population spend greater than 10% of their household income on healthcare. For every point increase in the coverage of essential health services index, there was a 13-point reduction in DALYs of emergency medical diseases (95% CI -16, -11). Conversely, for every percent decrease in the population with large household expenditure on healthcare, there was a 0.48 increase in DALYs of emergency medical diseases (95% CI -5.6, 4.7). Conclusions: After adjusting for multiple variables, an increase in coverage of essential health services was significantly associated with improvement in DALYs for emergency conditions. There was, however, no association between catastrophic health expenditure and DALYs.

Keywords: emergency medicine, universal healthcare, global health, health economics

Procedia PDF Downloads 80
258 Simulation of Utility Accrual Scheduling and Recovery Algorithm in Multiprocessor Environment

Authors: A. Idawaty, O. Mohamed, A. Z. Zuriati

Abstract:

This paper presents the development of an event based Discrete Event Simulation (DES) for a recovery algorithm known Backward Recovery Global Preemptive Utility Accrual Scheduling (BR_GPUAS). This algorithm implements the Backward Recovery (BR) mechanism as a fault recovery solution under the existing Time/Utility Function/ Utility Accrual (TUF/UA) scheduling domain for multiprocessor environment. The BR mechanism attempts to take the faulty tasks back to its initial safe state and then proceeds to re-execute the affected section of the faulty tasks to enable recovery. Considering that faults may occur in the components of any system; a fault tolerance system that can nullify the erroneous effect is necessary to be developed. Current TUF/UA scheduling algorithm uses the abortion recovery mechanism and it simply aborts the erroneous task as their fault recovery solution. None of the existing algorithm in TUF/UA scheduling domain in multiprocessor scheduling environment have considered the transient fault and implement the BR mechanism as a fault recovery mechanism to nullify the erroneous effect and solve the recovery problem in this domain. The developed BR_GPUAS simulator has derived the set of parameter, events and performance metrics according to a detailed analysis of the base model. Simulation results revealed that BR_GPUAS algorithm can saved almost 20-30% of the accumulated utilities making it reliable and efficient for the real-time application in the multiprocessor scheduling environment.

Keywords: real-time system (RTS), time utility function/ utility accrual (TUF/UA) scheduling, backward recovery mechanism, multiprocessor, discrete event simulation (DES)

Procedia PDF Downloads 289
257 Performance Evaluation of Parallel Surface Modeling and Generation on Actual and Virtual Multicore Systems

Authors: Nyeng P. Gyang

Abstract:

Even though past, current and future trends suggest that multicore and cloud computing systems are increasingly prevalent/ubiquitous, this class of parallel systems is nonetheless underutilized, in general, and barely used for research on employing parallel Delaunay triangulation for parallel surface modeling and generation, in particular. The performances, of actual/physical and virtual/cloud multicore systems/machines, at executing various algorithms, which implement various parallelization strategies of the incremental insertion technique of the Delaunay triangulation algorithm, were evaluated. T-tests were run on the data collected, in order to determine whether various performance metrics differences (including execution time, speedup and efficiency) were statistically significant. Results show that the actual machine is approximately twice faster than the virtual machine at executing the same programs for the various parallelization strategies. Results, which furnish the scalability behaviors of the various parallelization strategies, also show that some of the differences between the performances of these systems, during different runs of the algorithms on the systems, were statistically significant. A few pseudo superlinear speedup results, which were computed from the raw data collected, are not true superlinear speedup values. These pseudo superlinear speedup values, which arise as a result of one way of computing speedups, disappear and give way to asymmetric speedups, which are the accurate kind of speedups that occur in the experiments performed.

Keywords: cloud computing systems, multicore systems, parallel Delaunay triangulation, parallel surface modeling and generation

Procedia PDF Downloads 188
256 Effect of Dual Wavelength Light Exposure on Regeneration of Dugesia dorotocephala

Authors: Zayedali Shaikh

Abstract:

Increasingly now more than ever, UV damage brings with it a litany of minor deformities that can range from mild lesions and discoloring to cataracts and blindness. Pluripotent stem cells in planaria and human skin can be used to treat wounds and skin damage, with the primary limitations being inadequate growth factors. Photobiomodulation therapy in the form of low-intensity red light therapy has been proven to provide helpful benefits in the healing of skin that displays some of the symptoms of UV damage, such as burns and lesions, along with stimulating the proliferation of stem cells in recellularizing tissue. This paper puts forth an alternate means by which to treat the effects of UV damage using the freshwater planarian model system, Dugesia dorotocephala, known for its regenerative abilities and abundance of pluripotent stem cells, which allow for the rapid growth and repair of missing or damaged structures. Our work consisted of exposing planaria to different types of light: red light, blue light, white light, darkness, red and blue light together, UV light, and finally, red and UV light together. The primary focus of this research was on the red and UV lights, with six controls acting as metrics to compare our findings. Through computer-assisted morphological analysis, the results show that there is no significant difference in the rates of regeneration of planaria treated with simultaneous exposure to red and UV light versus planaria in darkness (p > .05), a representation of their preferred natural habitat. Our research suggests the viability of red-light therapy in actively combating UV damage and expediting the growth of epidermal stem cells by acting as another growth factor.

Keywords: regenerative medicine, stem cells, planaria, photobiomodulation

Procedia PDF Downloads 55
255 Using Environmental Life Cycle Assessment to Design Sustainable Packaging

Authors: Timothy Francis Grant

Abstract:

There are conflicting purposes at play with the design of sustainable packaging which include material reduction, recycling compatibility, use of secondary content and performance of the package in protecting and delivering the product. Life Cycle Assessment (LCA) is able to evaluate these different strategies against environmental metrics such as climate change, land and water use and marine litter pollution. However, LCA has traditionally been too time consuming and expensive to be used effectively in packaging design process. To make LCA practical for packaging technologist and designers a simplified tool is needed to make LCA possible for non-environmental specialists. The Packaging Quick Evaluation Tool (PIQET) is a web-based solution for undertaking LCA of new and existing packaging designs considering the global supply chain and impacts from cradle to grave. PIQET is based on a pre-calculated LCA database covering the materials and processes involved in the packaging lifecycle from cradle to grave. This includes both virgin materials and recycled content, conversion of materials into packaging, and the transportation of packaging to the product filling. In addition, PIQET assesses the impacts once the package is filled looking at storage, transport and product loss through the supply chain. When applied to consumer packaging light weight packages which are note recyclable have lower impacts than more recyclable packages which have a higher mass. Its also apparent that for many products the impacts of product failure and product loss are more important environmentally compared to packaging material efficiency.

Keywords: Climate change, Life Cycle Assessment, Marine litter, Packaging sustainability

Procedia PDF Downloads 110
254 Identification of Indices to Quantify Gentrification

Authors: Sophy Ann Xavier, Lakshmi A

Abstract:

Gentrification is the process of altering a neighborhood's character through the influx of wealthier people and establishments. This idea has subsequently been expanded to encompass brand-new, high-status construction projects that involve regenerating brownfield sites or demolishing and rebuilding residential neighborhoods. Inequality is made worse by Gentrification in ways that go beyond socioeconomic position. The elderly, members of racial and ethnic minorities, individuals with disabilities, and mental health all suffer disproportionately when they are displaced. Cities must cultivate openness, diversity, and inclusion in their collaborations, as well as cooperation on objectives and results. The papers compiled in this issue concentrate on the new gentrification discussions, the rising residential allure of central cities, and the indices to measure this process according to its various varieties. The study makes an effort to fill the research gap in the area of gentrification studies, which is the absence of a set of indices for measuring Gentrification in a specific area. Studies on Gentrification that contain maps of historical change highlight trends that will aid in the production of displacement risk maps, which will guide future interventions by allowing residents and policymakers to extrapolate into the future. Additionally, these maps give locals a glimpse into the future of their communities and serve as a political call to action in areas where residents are expected to be displaced. This study intends to pinpoint metrics and approaches for measuring Gentrification that can then be applied to create a spatiotemporal map of a region and tactics for its inclusive planning. An understanding of various approaches will enable planners and policymakers to select the best approach and create the appropriate plans.

Keywords: gentrification, indices, methods, quantification

Procedia PDF Downloads 64
253 Machine Learning Classification of Fused Sentinel-1 and Sentinel-2 Image Data Towards Mapping Fruit Plantations in Highly Heterogenous Landscapes

Authors: Yingisani Chabalala, Elhadi Adam, Khalid Adem Ali

Abstract:

Mapping smallholder fruit plantations using optical data is challenging due to morphological landscape heterogeneity and crop types having overlapped spectral signatures. Furthermore, cloud covers limit the use of optical sensing, especially in subtropical climates where they are persistent. This research assessed the effectiveness of Sentinel-1 (S1) and Sentinel-2 (S2) data for mapping fruit trees and co-existing land-use types by using support vector machine (SVM) and random forest (RF) classifiers independently. These classifiers were also applied to fused data from the two sensors. Feature ranks were extracted using the RF mean decrease accuracy (MDA) and forward variable selection (FVS) to identify optimal spectral windows to classify fruit trees. Based on RF MDA and FVS, the SVM classifier resulted in relatively high classification accuracy with overall accuracy (OA) = 0.91.6% and kappa coefficient = 0.91% when applied to the fused satellite data. Application of SVM to S1, S2, S2 selected variables and S1S2 fusion independently produced OA = 27.64, Kappa coefficient = 0.13%; OA= 87%, Kappa coefficient = 86.89%; OA = 69.33, Kappa coefficient = 69. %; OA = 87.01%, Kappa coefficient = 87%, respectively. Results also indicated that the optimal spectral bands for fruit tree mapping are green (B3) and SWIR_2 (B10) for S2, whereas for S1, the vertical-horizontal (VH) polarization band. Including the textural metrics from the VV channel improved crop discrimination and co-existing land use cover types. The fusion approach proved robust and well-suited for accurate smallholder fruit plantation mapping.

Keywords: smallholder agriculture, fruit trees, data fusion, precision agriculture

Procedia PDF Downloads 33
252 Women and Terrorism in Nigeria: Policy Templates for Addressing Complex Challenges in a Changing Democratic State

Authors: Godiya Pius Atsiya

Abstract:

One of the most devastating impacts of terrorism on the Nigerian state is the danger it has posed on women, children and other vulnerable groups. The complexity of terrorism in Nigeria, especially in most parts of Northern Nigeria has entrenched unprecedented security challenges such as refugee crisis, kidnapping, food shortages, increase in death tolls, malnutrition, fear, rape and several other psychological factors. Of particular interest in this paper as it relates to terrorism is the high rate of Internally Displaced Persons(IDPs), with women, children and the aged being the most affected. Empirical evidence arising from recent development in Nigeria’s North-East geo-political zone shows that large numbers of refugees fleeing the Boko Haram attacks have doubled. The attendant consequences of this mass exodus of people in the affected areas are that the victims now suffer untold and unwarranted economic hardship. In another dimension, recent findings have it that most powerless women and young teenage girls have been forcefully conscripted into the Islamic extremist groups and used as shields. In some respect, these groups of people have been used as available tools for suicide bombing and other criminal tendencies, the result of which can be detrimental to social cohesion and integration. This work is a theoretical insight into terrorism discourses; hence, the paper relies on existing works of scholars in carrying out the research. The paper argues that the implications of terrorism on women gender have grounding effects on the moral psyche of women who are supposed to be home managers and custodians of morality in society. The burden of terrorism and all it tends to propagate has literally upturned social lives and hence, Nigeria is gradually being plunged into the Hobesian state of nature. As a panacea to resolving this social malaise, the paper submits that government and indeed, all stakeholders in the nation’s democratic project must expedite action to nip this trend in the bud. The paper sums up with conclusion and other alternative policy measures to mitigate the challenges of terrorism in Nigeria.

Keywords: changing democratic state, policy measures, terrorism, women

Procedia PDF Downloads 210
251 Load Balancing Technique for Energy - Efficiency in Cloud Computing

Authors: Rani Danavath, V. B. Narsimha

Abstract:

Cloud computing is emerging as a new paradigm of large scale distributed computing. Cloud computing is a model for enabling ubiquitous, convenient, on-demand network access to a shared pool of configurable computing resources (e.g., three service models, and four deployment networks, servers, storage, applications, and services) that can be rapidly provisioned and released with minimal management effort or service provider interaction. This cloud model is composed of five essential characteristics models. Load balancing is one of the main challenges in cloud computing, which is required to distribute the dynamic workload across multiple nodes, to ensure that no single node is overloaded. It helps in optimal utilization of resources, enhancing the performance of the system. The goal of the load balancing is to minimize the resource consumption and carbon emission rate, that is the direct need of cloud computing. This determined the need of new metrics energy consumption and carbon emission for energy-efficiency load balancing techniques in cloud computing. Existing load balancing techniques mainly focuses on reducing overhead, services, response time and improving performance etc. In this paper we introduced a Technique for energy-efficiency, but none of the techniques have considered the energy consumption and carbon emission. Therefore, our proposed work will go towards energy – efficiency. So this energy-efficiency load balancing technique can be used to improve the performance of cloud computing by balancing the workload across all the nodes in the cloud with the minimum resource utilization, in turn, reducing energy consumption, and carbon emission to an extent, which will help to achieve green computing.

Keywords: cloud computing, distributed computing, energy efficiency, green computing, load balancing, energy consumption, carbon emission

Procedia PDF Downloads 429
250 Nanda Ways of Knowing, Being and Doing: Our Process of Research Engagement and Research Impacts

Authors: Steven Kelly

Abstract:

A fundament role of the researcher is research engagement, that is, the interaction between researchers and research end-users outside of academia for the mutually beneficial transfer of knowledge, technologies, methods, or resources. While research impact is the contribution that research makes to the economy, society, environment, or culture beyond the contribution to academic research. Ironically, traditional impact metrics in the academy are designed to focus on the outputs; it dismisses the important role engagement plays in fostering a collaborative process that leads to meaningful, ethical, and useful impacts. Dr. Kelly, aNanda (First Nations) man himself, has worked closely with the Nanda community over the past decade, ensuring cultural protocols are upheld and implemented while doing research engagement. The focus was on the process, which was essential to foster a positive research impact culture. The contributions that flowed from this process were the naming of a new species of squat lobster in the Nanda language, a poster design in collaboration with The University of Melbourne, Museums Victoria and Bundiyarra - IrraWanga language centre, media coverage, and the formation of the “Nanda language, Nanda country project”. The Nanda language, Nanda country project is a language revitalization project that focused on reconnecting Nanda people with the language & culture on Nanda Country. Such outcomes are imperative on the eve of the United Nations International Decade of Indigenous Languages. In this paperDr, Kellywill discuss howNanda cultural practicesinformed research engagement to foster a collaborative processthat, in turn, ledto meaningful, ethical, and useful impacts within and outside of the academy.

Keywords: community collaboration, indigenous, nanda, research engagement, research impacts

Procedia PDF Downloads 98
249 Securing Wireless Sensor Network From Rank Attack Using Fast Sensor Data Encryption and Decryption Protocol

Authors: Eden Teshome Hunde

Abstract:

Wireless sensor and actuator networks (WSANs) are of great significance in the realm of industrial automation systems. However, the aspect of security in WSANs has been somewhat overlooked. One particular security concern is the rank attack, where malicious actors actively manipulate the transmission of messages from neighboring nodes. This undermines the entire network's data collection and routing operations, resulting in a significant degradation of network performance. This attack adversely affects crucial metrics such as packet delivery ratio (PDR), latency, and power consumption, ultimately reducing the network's overall lifespan. In order to foster trust among nodes, ensure accurate delivery of data to end users, safeguard shared data in the cloud from security breaches, and prevent rank attacks within the network, it is crucial to protect the network against such malicious activities. This research paper aims to introduce an enhanced version of the Routing Protocol for Low-Power and Lossy Networks (RPL) protocol, specifically tailored to identify and eliminate rank attacks within existing WSANs. The effectiveness of the new protocol will be assessed through experimentation using Zolertia (Z1) sensors in the Cooja network simulator. To minimize network overhead on the sensors' side, the proposed scheme limits cryptographic operations to symmetric key-based mechanisms such as XORing, hash functions, and encryption. These operations will be implemented using a C-compiler and verified through the ModelSIM Altera SE edition 11.0 simulator.

Keywords: ModelSIM Altera SE, RPL, WSANs, Zolertia

Procedia PDF Downloads 17