Search results for: computer dynamics modelling
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 6538

Search results for: computer dynamics modelling

628 A Neurofeedback Learning Model Using Time-Frequency Analysis for Volleyball Performance Enhancement

Authors: Hamed Yousefi, Farnaz Mohammadi, Niloufar Mirian, Navid Amini

Abstract:

Investigating possible capacities of visual functions where adapted mechanisms can enhance the capability of sports trainees is a promising area of research, not only from the cognitive viewpoint but also in terms of unlimited applications in sports training. In this paper, the visual evoked potential (VEP) and event-related potential (ERP) signals of amateur and trained volleyball players in a pilot study were processed. Two groups of amateur and trained subjects are asked to imagine themselves in the state of receiving a ball while they are shown a simulated volleyball field. The proposed method is based on a set of time-frequency features using algorithms such as Gabor filter, continuous wavelet transform, and a multi-stage wavelet decomposition that are extracted from VEP signals that can be indicative of being amateur or trained. The linear discriminant classifier achieves the accuracy, sensitivity, and specificity of 100% when the average of the repetitions of the signal corresponding to the task is used. The main purpose of this study is to investigate the feasibility of a fast, robust, and reliable feature/model determination as a neurofeedback parameter to be utilized for improving the volleyball players’ performance. The proposed measure has potential applications in brain-computer interface technology where a real-time biomarker is needed.

Keywords: visual evoked potential, time-frequency feature extraction, short-time Fourier transform, event-related spectrum potential classification, linear discriminant analysis

Procedia PDF Downloads 128
627 Numerical Simulation of Free Surface Water Wave for the Flow Around NACA 0012 Hydrofoil and Wigley Hull Using VOF Method

Authors: Omar Imine, Mohammed Aounallah, Mustapha Belkadi

Abstract:

Steady three-dimensional and two free surface waves generated by moving bodies are presented, the flow problem to be simulated is rich in complexity and poses many modeling challenges because of the existence of breaking waves around the ship hull, and because of the interaction of the two-phase flow with the turbulent boundary layer. The results of several simulations are reported. The first study was performed for NACA0012 of hydrofoil with different meshes, this section is analyzed at h/c= 1, 0345 for 2D. In the second simulation, a mathematically defined Wigley hull form is used to investigate the application of a commercial CFD code in prediction of the total resistance and its components from tangential and normal forces on the hull wetted surface. The computed resistance and wave profiles are used to estimate the coefficient of the total resistance for Wigley hull advancing in calm water under steady conditions. The commercial CFD software FLUENT version 12 is used for the computations in the present study. The calculated grid is established using the code computer GAMBIT 2.3.26. The shear stress k-ωSST model is used for turbulence modeling and the volume of the fluid technique is employed to simulate the free-surface motion. The second order upwind scheme is used for discretizing the convection terms in the momentum transport equations, the Modified HRICscheme for VOF discretization. The results obtained compare well with the experimental data.

Keywords: free surface flows, breaking waves, boundary layer, Wigley hull, volume of fluid

Procedia PDF Downloads 367
626 A Study to Explore the Views of Students regarding E-Learning as an Instructional Tool at University Level

Authors: Zafar Iqbal

Abstract:

This study involved students of 6th semester enrolled in a Bachelor of Computer Science Program at university level. In this era of science and technology, e-learning can be helpful for grassroots in providing them access to education tenant in less developed areas. It is a potential substitute of face-to-face teaching being used in different countries. The purpose of the study was to explore the views of students about e-learning (Facebook) as an instructional tool. By using purposive sampling technique an intact class of 30 students included both male and female were selected where e-learning was used as an instructional tool. The views of students were explored through qualitative approach by using focus group interviews. The approach was helpful to develop comprehensive understanding of students’ views towards e- learning. In addition, probing questions were also asked and recorded. Data was transcribed, generated nodes and then coded text against these nodes. For this purpose and further analysis, NVivo 10 software was used. Themes were generated and tangibly presented through cluster analysis. Findings were interesting and provide sufficient evidence that face book is a subsequent e-learning source for students of higher education. Students acknowledged it as best source of learning and it was aligned with their academic and social behavior. It was not time specific and therefore, feasible for students who work day time and can get on line access to the material when they got free time. There were some distracters (time wasters) reported by the students but can be minimized by little effort. In short, e-learning is need of the day and potential learning source for every individual who have access to internet living at any part of the globe.

Keywords: e-learning, facebook, instructional tool, higher education

Procedia PDF Downloads 368
625 Deciphering Information Quality: Unraveling the Impact of Information Distortion in the UK Aerospace Supply Chains

Authors: Jing Jin

Abstract:

The incorporation of artificial intelligence (AI) and machine learning (ML) in aircraft manufacturing and aerospace supply chains leads to the generation of a substantial amount of data among various tiers of suppliers and OEMs. Identifying the high-quality information challenges decision-makers. The application of AI/ML models necessitates access to 'high-quality' information to yield desired outputs. However, the process of information sharing introduces complexities, including distortion through various communication channels and biases introduced by both human and AI entities. This phenomenon significantly influences the quality of information, impacting decision-makers engaged in configuring supply chain systems. Traditionally, distorted information is categorized as 'low-quality'; however, this study challenges this perception, positing that distorted information, contributing to stakeholder goals, can be deemed high-quality within supply chains. The main aim of this study is to identify and evaluate the dimensions of information quality crucial to the UK aerospace supply chain. Guided by a central research question, "What information quality dimensions are considered when defining information quality in the UK aerospace supply chain?" the study delves into the intricate dynamics of information quality in the aerospace industry. Additionally, the research explores the nuanced impact of information distortion on stakeholders' decision-making processes, addressing the question, "How does the information distortion phenomenon influence stakeholders’ decisions regarding information quality in the UK aerospace supply chain system?" This study employs deductive methodologies rooted in positivism, utilizing a cross-sectional approach and a mono-quantitative method -a questionnaire survey. Data is systematically collected from diverse tiers of supply chain stakeholders, encompassing end-customers, OEMs, Tier 0.5, Tier 1, and Tier 2 suppliers. Employing robust statistical data analysis methods, including mean values, mode values, standard deviation, one-way analysis of variance (ANOVA), and Pearson’s correlation analysis, the study interprets and extracts meaningful insights from the gathered data. Initial analyses challenge conventional notions, revealing that information distortion positively influences the definition of information quality, disrupting the established perception of distorted information as inherently low-quality. Further exploration through correlation analysis unveils the varied perspectives of different stakeholder tiers on the impact of information distortion on specific information quality dimensions. For instance, Tier 2 suppliers demonstrate strong positive correlations between information distortion and dimensions like access security, accuracy, interpretability, and timeliness. Conversely, Tier 1 suppliers emphasise strong negative influences on the security of accessing information and negligible impact on information timeliness. Tier 0.5 suppliers showcase very strong positive correlations with dimensions like conciseness and completeness, while OEMs exhibit limited interest in considering information distortion within the supply chain. Introducing social network analysis (SNA) provides a structural understanding of the relationships between information distortion and quality dimensions. The moderately high density of ‘information distortion-by-information quality’ underscores the interconnected nature of these factors. In conclusion, this study offers a nuanced exploration of information quality dimensions in the UK aerospace supply chain, highlighting the significance of individual perspectives across different tiers. The positive influence of information distortion challenges prevailing assumptions, fostering a more nuanced understanding of information's role in the Industry 4.0 landscape.

Keywords: information distortion, information quality, supply chain configuration, UK aerospace industry

Procedia PDF Downloads 50
624 Interaction Evaluation of Silver Ion and Silver Nanoparticles with Dithizone Complexes Using DFT Calculations and NMR Analysis

Authors: W. Nootcharin, S. Sujittra, K. Mayuso, K. Kornphimol, M. Rawiwan

Abstract:

Silver has distinct antibacterial properties and has been used as a component of commercial products with many applications. An increasing number of commercial products cause risks of silver effects for human and environment such as the symptoms of Argyria and the release of silver to the environment. Therefore, the detection of silver in the aquatic environment is important. The colorimetric chemosensor is designed by the basic of ligand interactions with a metal ion, leading to the change of signals for the naked-eyes which are very useful method to this application. Dithizone ligand is considered as one of the effective chelating reagents for metal ions due to its high selectivity and sensitivity of a photochromic reaction for silver as well as the linear backbone of dithizone affords the rotation of various isomeric forms. The present study is focused on the conformation and interaction of silver ion and silver nanoparticles (AgNPs) with dithizone using density functional theory (DFT). The interaction parameters were determined in term of binding energy of complexes and the geometry optimization, frequency of the structures and calculation of binding energies using density functional approaches B3LYP and the 6-31G(d,p) basis set. Moreover, the interaction of silver–dithizone complexes was supported by UV–Vis spectroscopy, FT-IR spectrum that was simulated by using B3LYP/6-31G(d,p) and 1H NMR spectra calculation using B3LYP/6-311+G(2d,p) method compared with the experimental data. The results showed the ion exchange interaction between hydrogen of dithizone and silver atom, with minimized binding energies of silver–dithizone interaction. However, the result of AgNPs in the form of complexes with dithizone. Moreover, the AgNPs-dithizone complexes were confirmed by using transmission electron microscope (TEM). Therefore, the results can be the useful information for determination of complex interaction using the analysis of computer simulations.

Keywords: silver nanoparticles, dithizone, DFT, NMR

Procedia PDF Downloads 199
623 Memristor-A Promising Candidate for Neural Circuits in Neuromorphic Computing Systems

Authors: Juhi Faridi, Mohd. Ajmal Kafeel

Abstract:

The advancements in the field of Artificial Intelligence (AI) and technology has led to an evolution of an intelligent era. Neural networks, having the computational power and learning ability similar to the brain is one of the key AI technologies. Neuromorphic computing system (NCS) consists of the synaptic device, neuronal circuit, and neuromorphic architecture. Memristor are a promising candidate for neuromorphic computing systems, but when it comes to neuromorphic computing, the conductance behavior of the synaptic memristor or neuronal memristor needs to be studied thoroughly in order to fathom the neuroscience or computer science. Furthermore, there is a need of more simulation work for utilizing the existing device properties and providing guidance to the development of future devices for different performance requirements. Hence, development of NCS needs more simulation work to make use of existing device properties. This work aims to provide an insight to build neuronal circuits using memristors to achieve a Memristor based NCS.  Here we throw a light on the research conducted in the field of memristors for building analog and digital circuits in order to motivate the research in the field of NCS by building memristor based neural circuits for advanced AI applications. This literature is a step in the direction where we describe the various Key findings about memristors and its analog and digital circuits implemented over the years which can be further utilized in implementing the neuronal circuits in the NCS. This work aims to help the electronic circuit designers to understand how the research progressed in memristors and how these findings can be used in implementing the neuronal circuits meant for the recent progress in the NCS.

Keywords: analog circuits, digital circuits, memristors, neuromorphic computing systems

Procedia PDF Downloads 163
622 Large-Scale Simulations of Turbulence Using Discontinuous Spectral Element Method

Authors: A. Peyvan, D. Li, J. Komperda, F. Mashayek

Abstract:

Turbulence can be observed in a variety fluid motions in nature and industrial applications. Recent investment in high-speed aircraft and propulsion systems has revitalized fundamental research on turbulent flows. In these systems, capturing chaotic fluid structures with different length and time scales is accomplished through the Direct Numerical Simulation (DNS) approach since it accurately simulates flows down to smallest dissipative scales, i.e., Kolmogorov’s scales. The discontinuous spectral element method (DSEM) is a high-order technique that uses spectral functions for approximating the solution. The DSEM code has been developed by our research group over the course of more than two decades. Recently, the code has been improved to run large cases in the order of billions of solution points. Running big simulations requires a considerable amount of RAM. Therefore, the DSEM code must be highly parallelized and able to start on multiple computational nodes on an HPC cluster with distributed memory. However, some pre-processing procedures, such as determining global element information, creating a global face list, and assigning global partitioning and element connection information of the domain for communication, must be done sequentially with a single processing core. A separate code has been written to perform the pre-processing procedures on a local machine. It stores the minimum amount of information that is required for the DSEM code to start in parallel, extracted from the mesh file, into text files (pre-files). It packs integer type information with a Stream Binary format in pre-files that are portable between machines. The files are generated to ensure fast read performance on different file-systems, such as Lustre and General Parallel File System (GPFS). A new subroutine has been added to the DSEM code to read the startup files using parallel MPI I/O, for Lustre, in a way that each MPI rank acquires its information from the file in parallel. In case of GPFS, in each computational node, a single MPI rank reads data from the file, which is specifically generated for the computational node, and send them to other ranks on the node using point to point non-blocking MPI communication. This way, communication takes place locally on each node and signals do not cross the switches of the cluster. The read subroutine has been tested on Argonne National Laboratory’s Mira (GPFS), National Center for Supercomputing Application’s Blue Waters (Lustre), San Diego Supercomputer Center’s Comet (Lustre), and UIC’s Extreme (Lustre). The tests showed that one file per node is suited for GPFS and parallel MPI I/O is the best choice for Lustre file system. The DSEM code relies on heavily optimized linear algebra operation such as matrix-matrix and matrix-vector products for calculation of the solution in every time-step. For this, the code can either make use of its matrix math library, BLAS, Intel MKL, or ATLAS. This fact and the discontinuous nature of the method makes the DSEM code run efficiently in parallel. The results of weak scaling tests performed on Blue Waters showed a scalable and efficient performance of the code in parallel computing.

Keywords: computational fluid dynamics, direct numerical simulation, spectral element, turbulent flow

Procedia PDF Downloads 125
621 Managing Fake News for Sustainable Democracy in Enugu State, Nigeria

Authors: Gloria Ebere Amadi, Emeka Promise Ugwunwotti

Abstract:

The study was carried out to determine the strategies for managing fake news for sustainable democracy in Enugu State, Nigeria. Two research questions and two null hypotheses guided the study. A survey research design was used for the study. The population for the study consisted of 100 respondents (from Enugu state House of Assembly). Of the entire population, 24 elected law makers and 76 staff were used; hence there was no sampling since the population was manageable. A 28-item structured questionnaire developed by the researcher was used for data collection. The instrument entitled Managing Fake News Questionnaire (MFNQ) was validated by three experts, two from the Department of Computer Science and one from the Department of Maths and Statistics, all from Enugu State University of Science and Technology. Cronbach Alpha was used to determine the reliability coefficient of the two sections of the instrument, and they are 0.67 and 0.82, while the reliability coefficient of the whole instrument gave a value of 0.81. Mean with standard deviation was used to answer research questions, while the null hypotheses at 0.5 level of significance at 98 degrees of freedom were tested with a t-test. The findings of the study revealed that the respondents agreed that government and citizens-related strategies improve the management of fake news for sustainable democracy in Enugu State. Again, there was no significant difference between the mean response of the lawmakers and staff on government and citizens-related strategies for managing fake news for sustainable democracy in Enugu State. Based on the findings, it was recommended, among others, that there should be regular workshops on the management of fake news for citizens.

Keywords: fake news, sustainability, democracy, management

Procedia PDF Downloads 60
620 Effect of Dual Wavelength Light Exposure on Regeneration of Dugesia dorotocephala

Authors: Zayedali Shaikh

Abstract:

Increasingly now more than ever, UV damage brings with it a litany of minor deformities that can range from mild lesions and discoloring to cataracts and blindness. Pluripotent stem cells in planaria and human skin can be used to treat wounds and skin damage, with the primary limitations being inadequate growth factors. Photobiomodulation therapy in the form of low-intensity red light therapy has been proven to provide helpful benefits in the healing of skin that displays some of the symptoms of UV damage, such as burns and lesions, along with stimulating the proliferation of stem cells in recellularizing tissue. This paper puts forth an alternate means by which to treat the effects of UV damage using the freshwater planarian model system, Dugesia dorotocephala, known for its regenerative abilities and abundance of pluripotent stem cells, which allow for the rapid growth and repair of missing or damaged structures. Our work consisted of exposing planaria to different types of light: red light, blue light, white light, darkness, red and blue light together, UV light, and finally, red and UV light together. The primary focus of this research was on the red and UV lights, with six controls acting as metrics to compare our findings. Through computer-assisted morphological analysis, the results show that there is no significant difference in the rates of regeneration of planaria treated with simultaneous exposure to red and UV light versus planaria in darkness (p > .05), a representation of their preferred natural habitat. Our research suggests the viability of red-light therapy in actively combating UV damage and expediting the growth of epidermal stem cells by acting as another growth factor.

Keywords: regenerative medicine, stem cells, planaria, photobiomodulation

Procedia PDF Downloads 64
619 An Extensible Software Infrastructure for Computer Aided Custom Monitoring of Patients in Smart Homes

Authors: Ritwik Dutta, Marylin Wolf

Abstract:

This paper describes the trade-offs and the design from scratch of a self-contained, easy-to-use health dashboard software system that provides customizable data tracking for patients in smart homes. The system is made up of different software modules and comprises a front-end and a back-end component. Built with HTML, CSS, and JavaScript, the front-end allows adding users, logging into the system, selecting metrics, and specifying health goals. The back-end consists of a NoSQL Mongo database, a Python script, and a SimpleHTTPServer written in Python. The database stores user profiles and health data in JSON format. The Python script makes use of the PyMongo driver library to query the database and displays formatted data as a daily snapshot of user health metrics against target goals. Any number of standard and custom metrics can be added to the system, and corresponding health data can be fed automatically, via sensor APIs or manually, as text or picture data files. A real-time METAR request API permits correlating weather data with patient health, and an advanced query system is implemented to allow trend analysis of selected health metrics over custom time intervals. Available on the GitHub repository system, the project is free to use for academic purposes of learning and experimenting, or practical purposes by building on it.

Keywords: flask, Java, JavaScript, health monitoring, long-term care, Mongo, Python, smart home, software engineering, webserver

Procedia PDF Downloads 381
618 Factors Influencing the Enjoyment and Performance of Students in Statistics Service Courses: A Mixed-Method Study

Authors: Wilma Coetzee

Abstract:

Statistics lecturers experience that many students who are taking a service course in statistics do not like statistics. Students in these courses tend to struggle and do not perform well. This research takes a look at the student’s perspective, with the aim to determine how to change the teaching of statistics so that students will enjoy it more and perform better. Questionnaires were used to determine the perspectives of first year service statistics students at a South African university. Factors addressed included motivation to study, attitude toward statistics, statistical anxiety, mathematical abilities and tendency to procrastinate. Logistic regression was used to determine what contributes to students performing badly in statistics. The results show that the factors that contribute the most to students performing badly are: statistical anxiety, not being motivated and having had mathematical literacy instead of mathematics in secondary school. Two open ended questions were included in the questionnaire: 'I will enjoy statistics more if…' and 'I will perform better in statistics if…'. The answers to these questions were analyzed using qualitative methods. Frequent themes were identified for each of the questions. A simulation study incorporating bootstrapping was done to determine the saturation of the themes. The majority of the students indicated that they would perform better in statistics if they studied more, managed their time better, had a flare for mathematics and if the lecturer was able to explain difficult concepts better. They also want more active learning. To ensure that students enjoy statistics more, they want an active learning experience. They want fun activities, more interaction with the lecturer and with one another, more computer based problems, and more challenges. They want a better understanding of the subject, want to understand the relevance of statistics to their future career and want excellent lecturers. These findings can be used to direct the improvement of the tuition of statistics.

Keywords: active learning, performance in statistics, statistical anxiety, statistics education

Procedia PDF Downloads 143
617 A Study of Seismic Design Approaches for Steel Sheet Piles: Hydrodynamic Pressures and Reduction Factors Using CFD and Dynamic Calculations

Authors: Helena Pera, Arcadi Sanmartin, Albert Falques, Rafael Rebolo, Xavier Ametller, Heiko Zillgen, Cecile Prum, Boris Even, Eric Kapornyai

Abstract:

Sheet piles system can be an interesting solution when dealing with harbors or quays designs. However, current design methods lead to conservative approaches due to the lack of specific basis of design. For instance, some design features still deal with pseudo-static approaches, although being a dynamic problem. Under this concern, the study particularly focuses on hydrodynamic water pressure definition and stability analysis of sheet pile system under seismic loads. During a seismic event, seawater produces hydrodynamic pressures on structures. Currently, design methods introduce hydrodynamic forces by means of Westergaard formulation and Eurocodes recommendations. They apply constant hydrodynamic pressure on the front sheet pile during the entire earthquake. As a result, the hydrodynamic load may represent 20% of the total forces produced on the sheet pile. Nonetheless, some studies question that approach. Hence, this study assesses the soil-structure-fluid interaction of sheet piles under seismic action in order to evaluate if current design strategies overestimate hydrodynamic pressures. For that purpose, this study performs various simulations by Plaxis 2D, a well-known geotechnical software, and CFD models, which treat fluid dynamic behaviours. Knowing that neither Plaxis nor CFD can resolve a soil-fluid coupled problem, the investigation imposes sheet pile displacements from Plaxis as input data for the CFD model. Then, it provides hydrodynamic pressures under seismic action, which fit theoretical Westergaard pressures if calculated using the acceleration at each moment of the earthquake. Thus, hydrodynamic pressures fluctuate during seismic action instead of remaining constant, as design recommendations propose. Additionally, these findings detect that hydrodynamic pressure contributes a 5% to the total load applied on sheet pile due to its instantaneous nature. These results are in line with other studies that use added masses methods for hydrodynamic pressures. Another important feature in sheet pile design is the assessment of the geotechnical overall stability. It uses pseudo-static analysis since the dynamic analysis cannot provide a safety calculation. Consequently, it estimates the seismic action. One of its relevant factors is the selection of the seismic reduction factor. A huge amount of studies discusses the importance of it but also about all its uncertainties. Moreover, current European standards do not propose a clear statement on that, and they recommend using a reduction factor equal to 1. This leads to conservative requirements when compared with more advanced methods. Under this situation, the study calibrates seismic reduction factor by fitting results from pseudo-static to dynamic analysis. The investigation concludes that pseudo-static analyses could reduce seismic action by 40-50%. These results are in line with some studies from Japanese and European working groups. In addition, it seems suitable to account for the flexibility of the sheet pile-soil system. Nevertheless, the calibrated reduction factor is subjected to particular conditions of each design case. Further research would contribute to specifying recommendations for selecting reduction factor values in the early stages of the design. In conclusion, sheet pile design still has chances for improving its design methodologies and approaches. Consequently, design could propose better seismic solutions thanks to advanced methods such as findings of this study.

Keywords: computational fluid dynamics, hydrodynamic pressures, pseudo-static analysis, quays, seismic design, steel sheet pile

Procedia PDF Downloads 134
616 Regulatory and Economic Challenges of AI Integration in Cyber Insurance

Authors: Shreyas Kumar, Mili Shangari

Abstract:

Integrating artificial intelligence (AI) in the cyber insurance sector represents a significant advancement, offering the potential to revolutionize risk assessment, fraud detection, and claims processing. However, this integration introduces a range of regulatory and economic challenges that must be addressed to ensure responsible and effective deployment of AI technologies. This paper examines the multifaceted regulatory landscape governing AI in cyber insurance and explores the economic implications of compliance, innovation, and market dynamics. AI's capabilities in processing vast amounts of data and identifying patterns make it an invaluable tool for insurers in managing cyber risks. Yet, the application of AI in this domain is subject to stringent regulatory scrutiny aimed at safeguarding data privacy, ensuring algorithmic transparency, and preventing biases. Regulatory bodies, such as the European Union with its General Data Protection Regulation (GDPR), mandate strict compliance requirements that can significantly impact the deployment of AI systems. These regulations necessitate robust data protection measures, ethical AI practices, and clear accountability frameworks, all of which entail substantial compliance costs for insurers. The economic implications of these regulatory requirements are profound. Insurers must invest heavily in upgrading their IT infrastructure, implementing robust data governance frameworks, and training personnel to handle AI systems ethically and effectively. These investments, while essential for regulatory compliance, can strain financial resources, particularly for smaller insurers, potentially leading to market consolidation. Furthermore, the cost of regulatory compliance can translate into higher premiums for policyholders, affecting the overall affordability and accessibility of cyber insurance. Despite these challenges, the potential economic benefits of AI integration in cyber insurance are significant. AI-enhanced risk assessment models can provide more accurate pricing, reduce the incidence of fraudulent claims, and expedite claims processing, leading to overall cost savings and increased efficiency. These efficiencies can improve the competitiveness of insurers and drive innovation in product offerings. However, balancing these benefits with regulatory compliance is crucial to avoid legal penalties and reputational damage. The paper also explores the potential risks associated with AI integration, such as algorithmic biases that could lead to unfair discrimination in policy underwriting and claims adjudication. Regulatory frameworks need to evolve to address these issues, promoting fairness and transparency in AI applications. Policymakers play a critical role in creating a balanced regulatory environment that fosters innovation while protecting consumer rights and ensuring market stability. In conclusion, the integration of AI in cyber insurance presents both regulatory and economic challenges that require a coordinated approach involving regulators, insurers, and other stakeholders. By navigating these challenges effectively, the industry can harness the transformative potential of AI, driving advancements in risk management and enhancing the resilience of the cyber insurance market. This paper provides insights and recommendations for policymakers and industry leaders to achieve a balanced and sustainable integration of AI technologies in cyber insurance.

Keywords: artificial intelligence (AI), cyber insurance, regulatory compliance, economic impact, risk assessment, fraud detection, cyber liability insurance, risk management, ransomware

Procedia PDF Downloads 18
615 Simulation Aided Life Cycle Sustainability Assessment Framework for Manufacturing Design and Management

Authors: Mijoh A. Gbededo, Kapila Liyanage, Ilias Oraifige

Abstract:

Decision making for sustainable manufacturing design and management requires critical considerations due to the complexity and partly conflicting issues of economic, social and environmental factors. Although there are tools capable of assessing the combination of one or two of the sustainability factors, the frameworks have not adequately integrated all the three factors. Case study and review of existing simulation applications also shows the approach lacks integration of the sustainability factors. In this paper we discussed the development of a simulation based framework for support of a holistic assessment of sustainable manufacturing design and management. To achieve this, a strategic approach is introduced to investigate the strengths and weaknesses of the existing decision supporting tools. Investigation reveals that Discrete Event Simulation (DES) can serve as a rock base for other Life Cycle Analysis frameworks. Simio-DES application optimizes systems for both economic and competitive advantage, Granta CES EduPack and SimaPro collate data for Material Flow Analysis and environmental Life Cycle Assessment, while social and stakeholders’ analysis is supported by Analytical Hierarchy Process, a Multi-Criteria Decision Analysis method. Such a common and integrated framework creates a platform for companies to build a computer simulation model of a real system and assess the impact of alternative solutions before implementing a chosen solution.

Keywords: discrete event simulation, life cycle sustainability analysis, manufacturing, sustainability

Procedia PDF Downloads 270
614 Analyzing the Usage of Social Media: A Study on Elderly in Malaysia

Authors: Chan Eang Teng, Tang Mui Joo

Abstract:

In the beginning of the prevalence of social media, it would be an obvious trend that the young adult age group has the highest population among the users on social media. However, apart from the age group of the users are becoming younger and younger, the elderly group has become a new force on social media, and this age group has increased rapidly. On top of that, the influence of social media towards the elderly is becoming more significant and it is even trending among them. This is because basic computer knowledge is not instilled into their life when they were young. This age group tends to be engrossed more than the young as this is something new for them, and they have the mindset that it is a new platform to approach things, and they tend to be more engrossed when they start getting in touch with the social media. Generally, most of the social media has been accepted and accessed by teenagers and young adult, but it is reasonable to believe that the social media is not really accepted among the elderly. Surprisingly, the elderlies are more addicted to the social media than the teenagers. Therefore, this study is to determine and understand the relationship between the elderly and social media, and how they employ social media in their lives. An online survey on 200 elderly aged 45-80 and an interview with a media expert are conducted to answer the main questions in the research paper. Uses and Gratification Approach is employed in theoretical framework. Finding revealed that majority of the respondents use social media to connect with family, friends, and for leisure purposes. The finding concluded that the elderly use social media differently according to their needs and wants which is in par with the highlight of Uses and Gratification theory. Considering the significantly large role social media plays in our culture and daily life today, the finding will shed some light on the effect of social media on the elderly or senior citizens who are usually relegated into a minority group in today’s age where the internet and social media are of great importance to our society and humanity in general. This may also serve to be useful in understanding behavioral patterns and preference in terms of social media usage among the elderly.

Keywords: elderly, Facebook, Malaysia, social media

Procedia PDF Downloads 353
613 A Comparative Study of Morphine and Clonidine as an Adjunct to Ropivacaine in Paravertebral Block for Modified Radical Mastectomy

Authors: Mukesh K., Siddiqui A. K., Abbas H., Gupta R.

Abstract:

Background: General Anesthesia is a standard for breast onco-surgery. The issue of postoperative pain and the occurrence of nausea and vomiting has prompted the quest for a superior methodology with fewer complications. Over the recent couple of years, paravertebral block (PVB) has acquired huge fame either in combination with GA or alone for anesthetic management. In this study, we aim to evaluate the efficacy of morphine and clonidine as an adjunct to ropivacaine in a paravertebral block in breast cancer patients undergoing modified radical mastectomy. Methods: In this study, total 90 patients were divided into three groups (30 each) on the basis of computer-generated randomization. Group C (Control): Paravertebral block with 0.25% ropivacaine (19ml) and 1 ml saline; Group M- Paravertebral block with 0.25% ropivacaine(19ml) + 20 microgram/kg body weight morphine; Group N: Paravertebral block with 0.25% ropivacaine(19ml) +1.0 microgram/kg body weight clonidine. The postoperative pain intensity was recorded using the visual analog scale (VAS) and Sedation was observed by the Ramsay Sedation score (RSS). Results: The VAS was similar at 0hr, 2hr and 4 hr in the postoperative period among all the groups. There was a significant (p=0.003) difference in VAS from 6 hr to 20 hr in the postoperative period among the groups. A significant (p<0.05) difference was observed among the groups at 8 hr to 20 hr). The first requirement of analgesia was significantly (p=0.001) higher in Group N (7.70±1.74) than in Group C (4.43±1.43) and Group M (7.33±2.21). Conclusion: The morphine in the paravertebral block provides better postoperative analgesia. The consumption of rescue analgesia was significantly reduced in the morphine group as compared to the clonidine group. The procedure also proved to be safe as no complication was encountered in the paravertebral block in our study.

Keywords: ropivacaine, morphine, clonidine, paravertebral block

Procedia PDF Downloads 112
612 The Impact of E-Learning on the Performance of History Learners in Eswatini General Certificate of Secondary Education

Authors: Joseph Osodo, Motsa Thobekani Phila

Abstract:

The study investigated the impact of e-learning on the performance of history learners in Eswatini general certificate of secondary education in the Manzini region of Eswatini. The study was guided by the theory of connectivism. The study had three objectives which were to find out the significance of e-learning during the COVID-19 era in learning History subject; challenges faced by history teachers’ and learners’ in e-learning; and how the challenges were mitigated. The study used a qualitative research approach and descriptive research design. Purposive sampling was used to select eight History teachers and eight History learners from four secondary schools in the Manzini region. Data were collected using face to face interviews. The collected data were analyzed and presented in thematically. The findings showed that history teachers had good knowledge on what e-learning was, while students had little understanding of e-learning. Some of the forms of e-learning that were used during the pandemic in teaching history in secondary schools included TV, radio, computer, projectors, and social media especially WhatsApp. E-learning enabled the continuity of teaching and learning of history subject. The use of e-learning through the social media was more convenient to the teacher and the learners. It was concluded that in some secondary school in the Manzini region, history teacher and learners encountered challenges such as lack of finances to purchase e-learning gadgets and data bundles, lack of skills as well as access to the Internet. It was recommended that History teachers should create more time to offer additional learning support to students whose performance was affected by the COVID-19 pandemic effects.

Keywords: e-learning, performance, COVID-19, history, connectivism

Procedia PDF Downloads 68
611 Ergosterol Biosynthesis: Non-Conventional Method for Improving Process

Authors: Madalina Postaru, Alexandra Tucaliuc, Dan Cascaval, Anca Irina Galaction

Abstract:

Ergosterol (ergosta-5,7,22-trien-3β-ol) is the precursor of vitamin D2 (ergocalciferol), known as provitamin D2 as it is converted under UV radiation to this vitamin. The natural sources of ergosterol are mainly the yeasts (Saccharomyces sp., Candida sp.), but it can be also found in fungus (Claviceps sp.) or plants (orchids). As ergosterol is mainly accumulated in yeast cell membranes, especially in free form in the plasma-membrane, and the chemical synthesis of ergosterol does not represent an efficient method for its production, this study aimed to analyze the influence of aeration efficiency on ergosterol production by S. cerevisiae in batch and fed-batch fermentations, by considering different levels of mixing intensity, aeration rate, and n-dodecane concentration. Our previous studies on ergosterol production by S. cerevisiae in batch and fed-batch fermentation systems indicated that the addition of n-dodecane led to the increase of almost 50% of this sterol concentration, the highest productivity being reached for the fed-batch process. The experiments were carried out in a laboratory stirred bioreactor, provided with computer-controlled and recorded parameters. In batch fermentation system, the study indicated that the oxygen mass transfer coefficient, kLa, is amplified for about 3 times by increasing the volumetric concentration of n-dodecane from 0 to 15%. Moreover, the increase of dissolved oxygen concentration by adding n-dodecane leads to the diminution for 3.5 times of the produced alcohol amount. In fed-batch fermentation process, the positive influence of hydrocarbon on oxygen transfer rate is amplified mainly at its higher concentration level, as the result of the increased yeasts cells amount. Thus, by varying n-dodecane concentration from 0 to 15% vol., the kLa value increase becomes more important than for the batch fermentation, being of 4 times

Keywords: ergosterol, yeast fermentation, n-dodecane, oxygen-vector

Procedia PDF Downloads 109
610 Multimedia Technologies Utilisation as Predictors of Lecturers’ Teaching Effectiveness in Colleges of Education in South-West, Nigeria

Authors: Abel Olusegun Egunjobi, Olusegun Oyeleye Adesanya

Abstract:

Teaching effectiveness of lecturers in a tertiary institution in Nigeria is one of the determinants of the lecturer’s productivity. In this study, therefore, lecturers’ teaching effectiveness was examined vis-à-vis their multimedia technologies utilisation in Colleges of Education (CoE) in South-West, Nigeria. This is for the purpose of ascertaining the relationship and contribution of multimedia technologies utilisation to lecturers’ teaching effectiveness in Nigerian colleges of education. The descriptive survey research design was adopted in the study, while a multi-stage sampling procedure was used in the study. A stratified sampling technique was used to select colleges of education, and a simple random sampling method was employed to select lecturers from the selected colleges of education. A total of 862 lecturers (627 males and 235 females) were selected from the colleges of education used for the study. The instrument used was lecturers’ questionnaire on multimedia technologies utilisation and teaching effectiveness with a reliability coefficient of 0.85 at 0.05 level of significance. The data collected were analysed using descriptive statistics, multiple regression, and t-test. The findings showed that the level of multimedia technologies utilisation in colleges of education was low, whereas lecturers’ teaching effectiveness was high. Findings also revealed that the lecturers used multimedia technologies purposely for personal and professional developments, so also for up to date news on economic and political matters. Also, findings indicated that laptop, Ipad, CD-ROMs, and computer instructional software were the multimedia technologies frequently utilised by the lecturers. There was also a significant difference in the teaching effectiveness between lecturers in the Federal and State COE. The government should, therefore, make adequate provision for multimedia technologies in the COE in Nigeria for lecturers’ utilisation in their instructions so as to boost their students’ learning outcomes.

Keywords: colleges of education, lecturers’ teaching effectiveness, multimedia technologies utilisation, Southwest Nigeria

Procedia PDF Downloads 129
609 Two-Dimensional Observation of Oil Displacement by Water in a Petroleum Reservoir through Numerical Simulation and Application to a Petroleum Reservoir

Authors: Ahmad Fahim Nasiry, Shigeo Honma

Abstract:

We examine two-dimensional oil displacement by water in a petroleum reservoir. The pore fluid is immiscible, and the porous media is homogenous and isotropic in the horizontal direction. Buckley-Leverett theory and a combination of Laplacian and Darcy’s law are used to study the fluid flow through porous media, and the Laplacian that defines the dispersion and diffusion of fluid in the sand using heavy oil is discussed. The reservoir is homogenous in the horizontal direction, as expressed by the partial differential equation. Two main factors which are observed are the water saturation and pressure distribution in the reservoir, and they are evaluated for predicting oil recovery in two dimensions by a physical and mathematical simulation model. We review the numerical simulation that solves difficult partial differential reservoir equations. Based on the numerical simulations, the saturation and pressure equations are calculated by the iterative alternating direction implicit method and the iterative alternating direction explicit method, respectively, according to the finite difference assumption. However, to understand the displacement of oil by water and the amount of water dispersion in the reservoir better, an interpolated contour line of the water distribution of the five-spot pattern, that provides an approximate solution which agrees well with the experimental results, is also presented. Finally, a computer program is developed to calculate the equation for pressure and water saturation and to draw the pressure contour line and water distribution contour line for the reservoir.

Keywords: numerical simulation, immiscible, finite difference, IADI, IDE, waterflooding

Procedia PDF Downloads 323
608 Effective Student Engaging Strategies to Enhance Academic Learning in Middle Eastern Classrooms: An Action Research Approach

Authors: Anjum Afrooze

Abstract:

The curriculum at General Sciences department in Prince Sultan University includes ‘Physical science’ for Computer Science, Information Technology and Business courses. Students are apathetic towards Physical Science and question, as to, ‘How this course is related to their majors?’ English is not a native language for the students and also for many instructors. More than sixty percent of the students come from institutions where English is not the medium of instruction, which makes student learning and academic achievement challenging. After observing the less enthusiastic student cohort for two consecutive semesters, the instructor was keen to find effective strategies to enhance learning and further encourage deep learning by engaging students in different tasks to empower them with necessary skills and motivate them. This study is participatory action research, in which instructor designs effective tasks to engage students in their learning. The study is conducted through two semesters with a total of 200 students. The effectiveness of this approach is studied using questionnaire at the end of each semester and teacher observation. Major outcomes of this study were overall improvement in students attitude towards science learning, enhancement of multiple skills like note taking, problem solving, language proficiency and also fortifying confidence. This process transformed instructor into engaging and reflecting practitioner. Also, these strategies were implemented by other instructors teaching the course and proved effective in opening a path to changes in related areas of the course curriculum. However, refinement in the strategies could be done based on student evaluation and instructors observation.

Keywords: group activity, language proficiency, reasoning skills, science learning

Procedia PDF Downloads 134
607 Investigating the Effectiveness of a 3D Printed Composite Mold

Authors: Peng Hao Wang, Garam Kim, Ronald Sterkenburg

Abstract:

In composite manufacturing, the fabrication of tooling and tooling maintenance contributes to a large portion of the total cost. However, as the applications of composite materials continue to increase, there is also a growing demand for more tooling. The demand for more tooling places heavy emphasis on the industry’s ability to fabricate high quality tools while maintaining the tool’s cost effectiveness. One of the popular techniques of tool fabrication currently being developed utilizes additive manufacturing technology known as 3D printing. The popularity of 3D printing is due to 3D printing’s ability to maintain low material waste, low cost, and quick fabrication time. In this study, a team of Purdue University School of Aviation and Transportation Technology (SATT) faculty and students investigated the effectiveness of a 3D printed composite mold. A steel valve cover from an aircraft reciprocating engine was modeled utilizing 3D scanning and computer-aided design (CAD) to create a 3D printed composite mold. The mold was used to fabricate carbon fiber versions of the aircraft reciprocating engine valve cover. The carbon fiber valve covers were evaluated for dimensional accuracy and quality while the 3D printed composite mold was evaluated for durability and dimensional stability. The data collected from this study provided valuable information in the understanding of 3D printed composite molds, potential improvements for the molds, and considerations for future tooling design.

Keywords: additive manufacturing, carbon fiber, composite tooling, molds

Procedia PDF Downloads 108
606 Reactive X Proactive Searches on Internet After Leprosy Institutional Campaigns in Brazil: A Google Trends Analysis

Authors: Paulo Roberto Vasconcellos-Silva

Abstract:

The "Janeiro Roxo" (Purple January) campaign in Brazil aims to promote awareness of leprosy and its early symptoms. The COVID-19 pandemic has adversely affected institutional campaigns, mostly considering leprosy a neglected disease by the media. Google Trends (GT) is a tool that tracks user searches on Google, providing insights into the popularity of specific search terms. Our prior research has categorized online searches into two types: "Reactive searches," driven by transient campaign-related stimuli, and "Proactive searches," driven by personal interest in early symptoms and self-diagnosis. Using GT we studied: (i) the impact of "Janeiro Roxo" on public interest in leprosy (assessed through reactive searches) and its early symptoms (evaluated through proactive searches) over the past five years; (ii) changes in public interest during and after the COVID-19 pandemic; (iii) patterns in the dynamics of reactive and proactive searches Methods: We used GT's "Relative Search Volume" (RSV) to gauge public interest on a scale from 0 to 100. "HANSENÍASE" (HAN) was a proxy for reactive searches, and "HANSENÍASE SINTOMAS" (leprosy symptoms) (H.SIN) for proactive searches (interest in leprosy or in self-diagnosis). We analyzed 261 weeks of data from 2018 to 2023, using polynomial trend lines to model trends over this period. Analysis of Variance (ANOVA) was used to compare weekly RSV, monthly (MM) and annual means (AM). Results: Over a span of 261 weeks, there was consistently higher Relative Search Volume (RSV) for HAN compared to H.SIN. Both search terms exhibited their highest (MM) in January months during all periods. COVID-19 pandemic: a decline was observed during the pandemic years (2020-2021). There was a 24% decrease in RSV for HAN and a 32.5% decrease for H.SIN. Both HAN and H.SIN regained their pre-pandemic search levels in January 2022-2023. Breakpoints indicated abrupt changes - in the 26th week (February 2019), 55th and 213th weeks (September 2019 and 2022) related to September regional campaigns (interrupted in 2020-2021). Trend lines for HAN exhibited an upward curve between 33rd-45th week (April to June 2019), a pandemic-related downward trend between 120th-136th week (December 2020 to March 2021), and an upward trend between 220th-240th week (November 2022 to March 2023). Conclusion: The "Janeiro Roxo" campaign, along with other media-driven activities, exerts a notable influence on both reactive and proactive searches related to leprosy topics. Reactive searches, driven by campaign stimuli, significantly outnumber proactive searches. Despite the interruption of the campaign due to the pandemic, there was a subsequent resurgence in both types of searches. The recovery observed in reactive and proactive searches post-campaign interruption underscores the effectiveness of such initiatives, particularly at the national level. This suggests that regional campaigns aimed at leprosy awareness can be considered highly successful in stimulating proactive public engagement. The evaluation of internet-based campaign programs proves valuable not only for assessing their impact but also for identifying the needs of vulnerable regions. These programs can play a crucial role in integrating regions and highlighting their needs for assistance services in the context of leprosy awareness.

Keywords: health communication, leprosy, health campaigns, information seeking behavior, Google Trends, reactive searches, proactive searches, leprosy early identification

Procedia PDF Downloads 49
605 Error Detection and Correction for Onboard Satellite Computers Using Hamming Code

Authors: Rafsan Al Mamun, Md. Motaharul Islam, Rabana Tajrin, Nabiha Noor, Shafinaz Qader

Abstract:

In an attempt to enrich the lives of billions of people by providing proper information, security and a way of communicating with others, the need for efficient and improved satellites is constantly growing. Thus, there is an increasing demand for better error detection and correction (EDAC) schemes, which are capable of protecting the data onboard the satellites. The paper is aimed towards detecting and correcting such errors using a special algorithm called the Hamming Code, which uses the concept of parity and parity bits to prevent single-bit errors onboard a satellite in Low Earth Orbit. This paper focuses on the study of Low Earth Orbit satellites and the process of generating the Hamming Code matrix to be used for EDAC using computer programs. The most effective version of Hamming Code generated was the Hamming (16, 11, 4) version using MATLAB, and the paper compares this particular scheme with other EDAC mechanisms, including other versions of Hamming Codes and Cyclic Redundancy Check (CRC), and the limitations of this scheme. This particular version of the Hamming Code guarantees single-bit error corrections as well as double-bit error detections. Furthermore, this version of Hamming Code has proved to be fast with a checking time of 5.669 nanoseconds, that has a relatively higher code rate and lower bit overhead compared to the other versions and can detect a greater percentage of errors per length of code than other EDAC schemes with similar capabilities. In conclusion, with the proper implementation of the system, it is quite possible to ensure a relatively uncorrupted satellite storage system.

Keywords: bit-flips, Hamming code, low earth orbit, parity bits, satellite, single error upset

Procedia PDF Downloads 120
604 Modeling Palm Oil Quality During the Ripening Process of Fresh Fruits

Authors: Afshin Keshvadi, Johari Endan, Haniff Harun, Desa Ahmad, Farah Saleena

Abstract:

Experiments were conducted to develop a model for analyzing the ripening process of oil palm fresh fruits in relation to oil yield and oil quality of palm oil produced. This research was carried out on 8-year-old Tenera (Dura × Pisifera) palms planted in 2003 at the Malaysian Palm Oil Board Research Station. Fresh fruit bunches were harvested from designated palms during January till May of 2010. The bunches were divided into three regions (top, middle and bottom), and fruits from the outer and inner layers were randomly sampled for analysis at 8, 12, 16 and 20 weeks after anthesis to establish relationships between maturity and oil development in the mesocarp and kernel. Computations on data related to ripening time, oil content and oil quality were performed using several computer software programs (MSTAT-C, SAS and Microsoft Excel). Nine nonlinear mathematical models were utilized using MATLAB software to fit the data collected. The results showed mean mesocarp oil percent increased from 1.24 % at 8 weeks after anthesis to 29.6 % at 20 weeks after anthesis. Fruits from the top part of the bunch had the highest mesocarp oil content of 10.09 %. The lowest kernel oil percent of 0.03 % was recorded at 12 weeks after anthesis. Palmitic acid and oleic acid comprised of more than 73 % of total mesocarp fatty acids at 8 weeks after anthesis, and increased to more than 80 % at fruit maturity at 20 weeks. The Logistic model with the highest R2 and the lowest root mean square error was found to be the best fit model.

Keywords: oil palm, oil yield, ripening process, anthesis, fatty acids, modeling

Procedia PDF Downloads 303
603 Synthesis and Characterization of Some Novel Carbazole Schiff Bases (OLED)

Authors: Baki Cicek, Umit Calisir

Abstract:

Carbazoles have been replaced lots of studies from 1960's to present and also still continues. In 1987, the first diode device had been developed. Thanks to that study, light emitting devices have been investigated and developed and also have been used on commercial applications. Nowadays, OLED (Organic Light Emitting Diodes) technology is using on lots of electronic screen such as (mobile phone, computer monitors, televisions, etc.) Carbazoles were subject a lot of study as a semiconductor material. Although this technology is used commen and widely, it is still development stage. Metal complexes of these compounds are using at pigment dyes because of colored substances, polymer technology, medicine industry, agriculture area, preparing rocket fuel-oil, determine some of biological events, etc. Becides all of these to preparing of schiff base synthesis is going on intensely. In this study, some of novel carbazole schiff bases were synthesized starting from carbazole. For that purpose, firstly, carbazole was alkylated. After purification of N-substituted-carbazole was nitrated to sythesized 3-nitro-N-substituted and 3,6-dinitro-N-substituted carbazoles. At next step, nitro group/groups were reduced to amines. Purified with using a type of silica gel-column chromatography. At the last step of our study, with sythesized 3,6-diamino-N-substituted carbazoles and 3-amino-N-substituted carbazoles were reacted with aldehydes to condensation reactions. 3-(imino-p-hydroxybenzyl)-N-isobutyl -carbazole, 3-(imino-2,3,4-trimethoxybenzene)-N-butylcarbazole, 3-(imino-3,4-dihydroxybenzene)-N-octylcarbazole, 3-(imino-2,3-dihydroxybenzene)-N-octylkarbazole and 3,6-di(α-imino-β-naphthol) -N-hexylcarbazole compounds were synthesized. All of synthesized compounds were characterized with FT-IR, 1H-NMR, 13C-NMR, and LC-MS.

Keywords: carbazole, carbazol schiff base, condensation reactions, OLED

Procedia PDF Downloads 430
602 Thresholding Approach for Automatic Detection of Pseudomonas aeruginosa Biofilms from Fluorescence in situ Hybridization Images

Authors: Zonglin Yang, Tatsuya Akiyama, Kerry S. Williamson, Michael J. Franklin, Thiruvarangan Ramaraj

Abstract:

Pseudomonas aeruginosa is an opportunistic pathogen that forms surface-associated microbial communities (biofilms) on artificial implant devices and on human tissue. Biofilm infections are difficult to treat with antibiotics, in part, because the bacteria in biofilms are physiologically heterogeneous. One measure of biological heterogeneity in a population of cells is to quantify the cellular concentrations of ribosomes, which can be probed with fluorescently labeled nucleic acids. The fluorescent signal intensity following fluorescence in situ hybridization (FISH) analysis correlates to the cellular level of ribosomes. The goals here are to provide computationally and statistically robust approaches to automatically quantify cellular heterogeneity in biofilms from a large library of epifluorescent microscopy FISH images. In this work, the initial steps were developed toward these goals by developing an automated biofilm detection approach for use with FISH images. The approach allows rapid identification of biofilm regions from FISH images that are counterstained with fluorescent dyes. This methodology provides advances over other computational methods, allowing subtraction of spurious signals and non-biological fluorescent substrata. This method will be a robust and user-friendly approach which will enable users to semi-automatically detect biofilm boundaries and extract intensity values from fluorescent images for quantitative analysis of biofilm heterogeneity.

Keywords: image informatics, Pseudomonas aeruginosa, biofilm, FISH, computer vision, data visualization

Procedia PDF Downloads 127
601 An Image Processing Scheme for Skin Fungal Disease Identification

Authors: A. A. M. A. S. S. Perera, L. A. Ranasinghe, T. K. H. Nimeshika, D. M. Dhanushka Dissanayake, Namalie Walgampaya

Abstract:

Nowadays, skin fungal diseases are mostly found in people of tropical countries like Sri Lanka. A skin fungal disease is a particular kind of illness caused by fungus. These diseases have various dangerous effects on the skin and keep on spreading over time. It becomes important to identify these diseases at their initial stage to control it from spreading. This paper presents an automated skin fungal disease identification system implemented to speed up the diagnosis process by identifying skin fungal infections in digital images. An image of the diseased skin lesion is acquired and a comprehensive computer vision and image processing scheme is used to process the image for the disease identification. This includes colour analysis using RGB and HSV colour models, texture classification using Grey Level Run Length Matrix, Grey Level Co-Occurrence Matrix and Local Binary Pattern, Object detection, Shape Identification and many more. This paper presents the approach and its outcome for identification of four most common skin fungal infections, namely, Tinea Corporis, Sporotrichosis, Malassezia and Onychomycosis. The main intention of this research is to provide an automated skin fungal disease identification system that increase the diagnostic quality, shorten the time-to-diagnosis and improve the efficiency of detection and successful treatment for skin fungal diseases.

Keywords: Circularity Index, Grey Level Run Length Matrix, Grey Level Co-Occurrence Matrix, Local Binary Pattern, Object detection, Ring Detection, Shape Identification

Procedia PDF Downloads 221
600 Design Optimization of Miniature Mechanical Drive Systems Using Tolerance Analysis Approach

Authors: Eric Mxolisi Mkhondo

Abstract:

Geometrical deviations and interaction of mechanical parts influences the performance of miniature systems.These deviations tend to cause costly problems during assembly due to imperfections of components, which are invisible to a naked eye.They also tend to cause unsatisfactory performance during operation due to deformation cause by environmental conditions.One of the effective tools to manage the deviations and interaction of parts in the system is tolerance analysis.This is a quantitative tool for predicting the tolerance variations which are defined during the design process.Traditional tolerance analysis assumes that the assembly is static and the deviations come from the manufacturing discrepancies, overlooking the functionality of the whole system and deformation of parts due to effect of environmental conditions. This paper presents an integrated tolerance analysis approach for miniature system in operation.In this approach, a computer-aided design (CAD) model is developed from system’s specification.The CAD model is then used to specify the geometrical and dimensional tolerance limits (upper and lower limits) that vary component’s geometries and sizes while conforming to functional requirements.Worst-case tolerances are analyzed to determine the influenced of dimensional changes due to effects of operating temperatures.The method is used to evaluate the nominal conditions, and worse case conditions in maximum and minimum dimensions of assembled components.These three conditions will be evaluated under specific operating temperatures (-40°C,-18°C, 4°C, 26°C, 48°C, and 70°C). A case study on the mechanism of a zoom lens system is used to illustrate the effectiveness of the methodology.

Keywords: geometric dimensioning, tolerance analysis, worst-case analysis, zoom lens mechanism

Procedia PDF Downloads 159
599 Analyzing the Effect of Materials’ Selection on Energy Saving and Carbon Footprint: A Case Study Simulation of Concrete Structure Building

Authors: M. Kouhirostamkolaei, M. Kouhirostami, M. Sam, J. Woo, A. T. Asutosh, J. Li, C. Kibert

Abstract:

Construction is one of the most energy consumed activities in the urban environment that results in a significant amount of greenhouse gas emissions around the world. Thus, the impact of the construction industry on global warming is undeniable. Thus, reducing building energy consumption and mitigating carbon production can slow the rate of global warming. The purpose of this study is to determine the amount of energy consumption and carbon dioxide production during the operation phase and the impact of using new shells on energy saving and carbon footprint. Therefore, a residential building with a re-enforced concrete structure is selected in Babolsar, Iran. DesignBuilder software has been used for one year of building operation to calculate the amount of carbon dioxide production and energy consumption in the operation phase of the building. The primary results show the building use 61750 kWh of energy each year. Computer simulation analyzes the effect of changing building shells -using XPS polystyrene and new electrochromic windows- as well as changing the type of lighting on energy consumption reduction and subsequent carbon dioxide production. The results show that the amount of energy and carbon production during building operation has been reduced by approximately 70% by applying the proposed changes. The changes reduce CO2e to 11345 kg CO2/yr. The result of this study helps designers and engineers to consider material selection’s process as one of the most important stages of design for improving energy performance of buildings.

Keywords: construction materials, green construction, energy simulation, carbon footprint, energy saving, concrete structure, designbuilder

Procedia PDF Downloads 190