Search results for: common platform for automated programming
8200 The AU Culture Platform Approach to Measure the Impact of Cultural Participation on Individuals
Authors: Sendy Ghirardi, Pau Rausell Köster
Abstract:
The European Commission increasingly pushes cultural policies towards social outcomes and local and regional authorities also call for culture-driven strategies for local development and prosperity and therefore, the measurement of cultural participation becomes increasingly more significant for evidence-based policy-making processes. Cultural participation involves various kinds of social and economic spillovers that combine social and economic objectives of value creation, including social sustainability and respect for human values. Traditionally, from the economic perspective, cultural consumption is measured by the value of financial transactions in purchasing, subscribing to, or renting cultural equipment and content, addressing the market value of cultural products and services. The main sources of data are the household spending survey and merchandise trade survey, among others. However, what characterizes the cultural consumption is that it is linked with the hedonistic and affective dimension rather than the utilitarian one. In fact, nowadays, more and more attention is being paid to the social and psychological dimensions of culture. The aim of this work is to present a comprehensive approach to measure the impacts of cultural participation and cultural users’ behaviour, combining both socio-psychological and economic approaches. The model combines contingent evaluation techniques with the individual characteristic and perception analysis of the cultural experiences to evaluate the cognitive, aesthetic, emotive and social impacts of cultural participation. To investigate the comprehensive approach to measure the impact of the cultural events on individuals, the research has been designed on the basis of prior theoretical development. A deep literature methodology has been done to develop the theoretical model applied to the web platform to measure the impacts of cultural experience on individuals. The developed framework aims to become a democratic tool for evaluating the services that cultural or policy institutions can adopt through the use of an interacting platform that produces big data benefiting academia, cultural management and policies. The Au Culture is a prototype based on an application that can be used on mobile phones or any other digital platform. The development of the AU Culture Platform has been funded by the Valencian Innovation Agency (Government of the Region of Valencia) and it is part of the Horizon 2020 project MESOC.Keywords: comprehensive approach, cultural participation, economic dimension, socio-psychological dimension
Procedia PDF Downloads 1158199 Knowledge Diffusion via Automated Organizational Cartography (Autocart)
Authors: Mounir Kehal
Abstract:
The post-globalization epoch has placed businesses everywhere in new and different competitive situations where knowledgeable, effective and efficient behavior has come to provide the competitive and comparative edge. Enterprises have turned to explicit - and even conceptualizing on tacit - knowledge management to elaborate a systematic approach to develop and sustain the intellectual capital needed to succeed. To be able to do that, you have to be able to visualize your organization as consisting of nothing but knowledge and knowledge flows, whilst being presented in a graphical and visual framework, referred to as automated organizational cartography. Hence, creating the ability of further actively classifying existing organizational content evolving from and within data feeds, in an algorithmic manner, potentially giving insightful schemes and dynamics by which organizational know-how is visualized. It is discussed and elaborated on most recent and applicable definitions and classifications of knowledge management, representing a wide range of views from mechanistic (systematic, data driven) to a more socially (psychologically, cognitive/metadata driven) orientated. More elaborate continuum models, for knowledge acquisition and reasoning purposes, are being used for effectively representing the domain of information that an end user may contain in their decision making process for utilization of available organizational intellectual resources (i.e. Autocart). In this paper, we present an empirical research study conducted previously to try and explore knowledge diffusion in a specialist knowledge domain.Keywords: knowledge management, knowledge maps, knowledge diffusion, organizational cartography
Procedia PDF Downloads 3098198 Verification of a Simple Model for Rolling Isolation System Response
Authors: Aarthi Sridhar, Henri Gavin, Karah Kelly
Abstract:
Rolling Isolation Systems (RISs) are simple and effective means to mitigate earthquake hazards to equipment in critical and precious facilities, such as hospitals, network collocation facilities, supercomputer centers, and museums. The RIS works by isolating components acceleration the inertial forces felt by the subsystem. The RIS consists of two platforms with counter-facing concave surfaces (dishes) in each corner. Steel balls lie inside the dishes and allow the relative motion between the top and bottom platform. Formerly, a mathematical model for the dynamics of RISs was developed using Lagrange’s equations (LE) and experimentally validated. A new mathematical model was developed using Gauss’s Principle of Least Constraint (GPLC) and verified by comparing impulse response trajectories of the GPLC model and the LE model in terms of the peak displacements and accelerations of the top platform. Mathematical models for the RIS are tedious to derive because of the non-holonomic rolling constraints imposed on the system. However, using Gauss’s Principle of Least constraint to find the equations of motion removes some of the obscurity and yields a system that can be easily extended. Though the GPLC model requires more state variables, the equations of motion are far simpler. The non-holonomic constraint is enforced in terms of accelerations and therefore requires additional constraint stabilization methods in order to avoid the possibility that numerical integration methods can cause the system to go unstable. The GPLC model allows the incorporation of more physical aspects related to the RIS, such as contribution of the vertical velocity of the platform to the kinetic energy and the mass of the balls. This mathematical model for the RIS is a tool to predict the motion of the isolation platform. The ability to statistically quantify the expected responses of the RIS is critical in the implementation of earthquake hazard mitigation.Keywords: earthquake hazard mitigation, earthquake isolation, Gauss’s Principle of Least Constraint, nonlinear dynamics, rolling isolation system
Procedia PDF Downloads 2518197 Development of an Automatic Monitoring System Based on the Open Architecture Concept
Authors: Andrii Biloshchytskyi, Serik Omirbayev, Alexandr Neftissov, Sapar Toxanov, Svitlana Biloshchytska, Adil Faizullin
Abstract:
Kazakhstan has adopted a carbon neutrality strategy until 2060. In accordance with this strategy, it is necessary to introduce various tools to maintain the environmental safety of the environment. The use of IoT, in combination with the characteristics and requirements of Kazakhstan's environmental legislation, makes it possible to develop a modern environmental monitoring system. The article proposes a solution for developing an example of an automated system for the continuous collection of data on the concentration of pollutants in the atmosphere based on an open architecture. The Audino-based device acts as a microcontroller. It should be noted that the transmission of measured values is carried out via an open wireless communication protocol. The architecture of the system, which was used to build a prototype based on sensors, an Arduino microcontroller, and a wireless data transmission module, is presented. The selection of elementary components may change depending on the requirements of the system; the introduction of new units is limited by the number of ports. The openness of solutions allows you to change the configuration depending on the conditions. The advantages of the solutions are openness, low cost, versatility and mobility. However, there is no comparison of the working processes of the proposed solution with traditional ones.Keywords: environmental monitoring, greenhouse gases emissions, environmental pollution, Industry 4.0, IoT, microcontroller, automated monitoring system.
Procedia PDF Downloads 488196 Art Beyond Borders: Virtual School Field Trips
Authors: Audrey Hudson
Abstract:
In 2020, educational field trips went virtual for all students. At the Art Gallery of Ontario (AGO) in Canada, our solution was to create a virtual school program that addressed three pillars of access—economic, geographic and cultural—with art at the center. Now, at the close of three years, we have reached 1.6 million students! Exponentially more than we have ever welcomed for onsite school visits. In 2022, we partnered with the Museum of Modern Art (MoMA), the Hong Kong University Museum and the National Gallery of Singapore, which has pushed the boundaries of art education into the expanse of the global community. Looking forward to our fourth year of the program, we are using the platform of technology to expand our program of art, access and learning to a global platform. In 2023/24, we intend to connect across more borders to expand the pedagogical benefits of art education for a global community. We invite you to listen to how you can join this journey.Keywords: technology, museums, art education, pedagogy
Procedia PDF Downloads 658195 Automated Human Balance Assessment Using Contactless Sensors
Authors: Justin Tang
Abstract:
Balance tests are frequently used to diagnose concussions on the sidelines of sporting events. Manual scoring, however, is labor intensive and subjective, and many concussions go undetected. This study institutes a novel approach to conducting the Balance Error Scoring System (BESS) more quantitatively using Microsoft’s gaming system Kinect, which uses a contactless sensor and several cameras to receive data and estimate body limb positions. Using a machine learning approach, Visual Gesture Builder, and a deterministic approach, MATLAB, we tested whether the Kinect can differentiate between “correct” and erroneous stances of the BESS. We created the two separate solutions by recording test videos to teach the Kinect correct stances and by developing a code using Java. Twenty-two subjects were asked to perform a series of BESS tests while the Kinect was collecting data. The Kinect recorded the subjects and mapped key joints onto their bodies to obtain angles and measurements that are interpreted by the software. Through VGB and MATLAB, the videos are analyzed to enumerate the number of errors committed during testing. The resulting statistics demonstrate a high correlation between manual scoring and the Kinect approaches, indicating the viability of the use of remote tracking devices in conducting concussion tests.Keywords: automated, concussion detection, contactless sensors, microsoft kinect
Procedia PDF Downloads 3178194 Automated Multisensory Data Collection System for Continuous Monitoring of Refrigerating Appliances Recycling Plants
Authors: Georgii Emelianov, Mikhail Polikarpov, Fabian Hübner, Jochen Deuse, Jochen Schiemann
Abstract:
Recycling refrigerating appliances plays a major role in protecting the Earth's atmosphere from ozone depletion and emissions of greenhouse gases. The performance of refrigerator recycling plants in terms of material retention is the subject of strict environmental certifications and is reviewed periodically through specialized audits. The continuous collection of Refrigerator data required for the input-output analysis is still mostly manual, error-prone, and not digitalized. In this paper, we propose an automated data collection system for recycling plants in order to deduce expected material contents in individual end-of-life refrigerating appliances. The system utilizes laser scanner measurements and optical data to extract attributes of individual refrigerators by applying transfer learning with pre-trained vision models and optical character recognition. Based on Recognized features, the system automatically provides material categories and target values of contained material masses, especially foaming and cooling agents. The presented data collection system paves the way for continuous performance monitoring and efficient control of refrigerator recycling plants.Keywords: automation, data collection, performance monitoring, recycling, refrigerators
Procedia PDF Downloads 1648193 The Synopsis of the AI-Powered Therapy Web Platform ‘Free AI Therapist'
Authors: Arwa Alnowaiser, Hala Shoukri
Abstract:
The ‘FreeAITherapist’ is an artificial intelligence application that uses the power of AI to offer advice and mental health counseling to its users through its chatbot services. The AI therapist is designed to understand users' issues, concerns, and problems and respond appropriately; it provides empathy and guidance and uses evidence-based therapeutic techniques. With its user-friendly platform, it ensures accessibility for individuals in need, regardless of their geographical location. This website was created in direct response to the growing demand for mental health support, aiming to provide a cost-effective and confidential solution. Through promising confidentiality, it considers user privacy and data security. The ‘FreeAITherapist’ strives to bridge the gap in mental health services, offering a reliable resource for individuals seeking guidance and counseling to improve their overall well-being.Keywords: artificial intelligence, mental health, AI therapist, website, counseling
Procedia PDF Downloads 458192 Experimenting with Error Performance of Systems Employing Pulse Shaping Filters on a Software-Defined-Radio Platform
Authors: Chia-Yu Yao
Abstract:
This paper presents experimental results on testing the symbol-error-rate (SER) performance of quadrature amplitude modulation (QAM) systems employing symmetric pulse-shaping square-root (SR) filters designed by minimizing the roughness function and by minimizing the peak-to-average power ratio (PAR). The device used in the experiments is the 'bladeRF' software-defined-radio platform. PAR is a well-known measurement, whereas the roughness function is a concept for measuring the jitter-induced interference. The experimental results show that the system employing minimum-roughness pulse-shaping SR filters outperforms the system employing minimum-PAR pulse-shaping SR filters in the sense of SER performance.Keywords: pulse-shaping filters, FIR filters, jittering, QAM
Procedia PDF Downloads 3418191 Removal of Perchloroethylene, a Common Pollutant, in Groundwater Using Activated Carbon
Authors: Marianne Miguet, Gaël Plantard, Yves Jaeger, Vincent Goetz
Abstract:
The contamination of groundwater is a major concern. A common pollutant, the perchloroethylene, is the target contaminant. Water treatment process as Granular Activated Carbons are very efficient but requires pilot-scale testing to determine the full-scale GAC performance. First, the batch mode was used to get a reliable experimental method to estimate the adsorption capacity of a common volatile compound is settled. The Langmuir model is acceptable to fit the isotherms. Dynamic tests were performed with three columns and different operating conditions. A database of concentration profiles and breakthroughs were obtained. The resolution of the set of differential equations is acceptable to fit the dynamics tests and could be used for a full-scale adsorber.Keywords: activated carbon, groundwater, perchloroethylene, full-scale
Procedia PDF Downloads 4268190 The Role of Twitter Bots in Political Discussion on 2019 European Elections
Authors: Thomai Voulgari, Vasilis Vasilopoulos, Antonis Skamnakis
Abstract:
The aim of this study is to investigate the effect of the European election campaigns (May 23-26, 2019) on Twitter achieving with artificial intelligence tools such as troll factories and automated inauthentic accounts. Our research focuses on the last European Parliamentary elections that took place between 23 and 26 May 2019 specifically in Italy, Greece, Germany and France. It is difficult to estimate how many Twitter users are actually bots (Echeverría, 2017). Detection for fake accounts is becoming even more complicated as AI bots are made more advanced. A political bot can be programmed to post comments on a Twitter account for a political candidate, target journalists with manipulated content or engage with politicians and artificially increase their impact and popularity. We analyze variables related to 1) the scope of activity of automated bots accounts and 2) degree of coherence and 3) degree of interaction taking into account different factors, such as the type of content of Twitter messages and their intentions, as well as the spreading to the general public. For this purpose, we collected large volumes of Twitter accounts of party leaders and MEP candidates between 10th of May and 26th of July based on content analysis of tweets based on hashtags while using an innovative network analysis tool known as MediaWatch.io (https://mediawatch.io/). According to our findings, one of the highest percentage (64.6%) of automated “bot” accounts during 2019 European election campaigns was in Greece. In general terms, political bots aim to proliferation of misinformation on social media. Targeting voters is a way that it can be achieved contribute to social media manipulation. We found that political parties and individual politicians create and promote purposeful content on Twitter using algorithmic tools. Based on this analysis, online political advertising play an important role to the process of spreading misinformation during elections campaigns. Overall, inauthentic accounts and social media algorithms are being used to manipulate political behavior and public opinion.Keywords: artificial intelligence tools, human-bot interactions, political manipulation, social networking, troll factories
Procedia PDF Downloads 1388189 An Efficient Hardware/Software Workflow for Multi-Cores Simulink Applications
Authors: Asma Rebaya, Kaouther Gasmi, Imen Amari, Salem Hasnaoui
Abstract:
Over these last years, applications such as telecommunications, signal processing, digital communication with advanced features (Multi-antenna, equalization..) witness a rapid evaluation accompanied with an increase of user exigencies in terms of latency, the power of computation… To satisfy these requirements, the use of hardware/software systems is a common solution; where hardware is composed of multi-cores and software is represented by models of computation, synchronous data flow (SDF) graph for instance. Otherwise, the most of the embedded system designers utilize Simulink for modeling. The issue is how to simplify the c code generation, for a multi-cores platform, of an application modeled by Simulink. To overcome this problem, we propose a workflow allowing an automatic transformation from the Simulink model to the SDF graph and providing an efficient schedule permitting to optimize the number of cores and to minimize latency. This workflow goes from a Simulink application and a hardware architecture described by IP.XACT language. Based on the synchronous and hierarchical behavior of both models, the Simulink block diagram is automatically transformed into an SDF graph. Once this process is successfully achieved, the scheduler calculates the optimal cores’ number needful by minimizing the maximum density of the whole application. Then, a core is chosen to execute a specific graph task in a specific order and, subsequently, a compatible C code is generated. In order to perform this proposal, we extend Preesm, a rapid prototyping tool, to take the Simulink model as entry input and to support the optimal schedule. Afterward, we compared our results to this tool results, using a simple illustrative application. The comparison shows that our results strictly dominate the Preesm results in terms of number of cores and latency. In fact, if Preesm needs m processors and latency L, our workflow need processors and latency L'< L.Keywords: hardware/software system, latency, modeling, multi-cores platform, scheduler, SDF graph, Simulink model, workflow
Procedia PDF Downloads 2698188 A New Computational Tool for Noise Prediction of Rotating Surfaces (FACT)
Authors: Ana Vieira, Fernando Lau, João Pedro Mortágua, Luís Cruz, Rui Santos
Abstract:
The air transport impact on environment is more than ever a limitative obstacle to the aeronautical industry continuous growth. Over the last decades, considerable effort has been carried out in order to obtain quieter aircraft solutions, whether by changing the original design or investigating more silent maneuvers. The noise propagated by rotating surfaces is one of the most important sources of annoyance, being present in most aerial vehicles. Bearing this is mind, CEIIA developed a new computational chain for noise prediction with in-house software tools to obtain solutions in relatively short time without using excessive computer resources. This work is based on the new acoustic tool, which aims to predict the rotor noise generated during steady and maneuvering flight, making use of the flexibility of the C language and the advantages of GPU programming in terms of velocity. The acoustic tool is based in the Formulation 1A of Farassat, capable of predicting two important types of noise: the loading and thickness noise. The present work describes the most important features of the acoustic tool, presenting its most relevant results and framework analyses for helicopters and UAV quadrotors.Keywords: rotor noise, acoustic tool, GPU Programming, UAV noise
Procedia PDF Downloads 4018187 The Comparative Effect of Neuro-Linguistic Programming (NLP), Critical Thinking and a Combination of Both On EFL Learners' Reading Comprehension
Authors: Mona Khabiri, Fahimeh Farahani
Abstract:
The present study was an attempt to investigate the comparative effect of teaching NLP, critical thinking, and a combination of NLP and critical thinking on EFL learners' reading comprehension. To fulfill the purpose of this study, a group of 82 female and male intermediate EFL learners at a Language School in Iran took a piloted sample PET as a proficiency test and 63 of them were selected as homogenous learners and were randomly assigned to three experimental groups. Within a treatment process of 10 sessions the teacher/researcher provided the participants of each group with handouts, explanations, practices, homework, and questionnaires on techniques of NLP, critical thinking, and a combination of both. During these 10 sessions, 10 same reading comprehension texts extracted from the multi-skill course book suggested by the language school where thought to the participants of each experimental group using skills and strategies of NLP, critical thinking, and a combination of both. On the eleventh session, the participants sat for a reading posttest. The results of one-way ANOVA showed no significant difference among the three groups in terms of reading comprehension. Justifications and implications for the findings of the study and suggestions for further research are presented.Keywords: neuro-linguistic programming (NLP), critical thinking, reading comprehension
Procedia PDF Downloads 4128186 Technology and Terror
Authors: Janet D. Fish
Abstract:
This paper will analyze how advanced information technology communications platform’s such as you tube, twitter, Facebook, and websites are being used in marketing cultural diversity on a global scale. The scope of this topic will encompass the use of marketing terror as a tool of educational understanding, accepting, and incorporating other ethnic groups into extremist Islamic cultural practices with an end goal of cultural assimilation. This paper will examine the impacts of various influences, such as globalism and technology on common public values and cultural diversity. Additionally, multiculturalism in public administration settings will be examined across cultures. Communications are a primary focus of review for this paper, the purpose of this review is to see how different technological platforms are currently being used as major tools of influence within the public sector. Technology and terror must become a primary concern for new public administrators in a modern world. While its existence is acknowledged, boundaries of legal recourse are currently few. Public administrators must understand the depth and reach of the future consequences of an unchecked process in the realm of technology and terror on a global scale.Keywords: inclusionism, exclusionism, technology, terror
Procedia PDF Downloads 3868185 From Ride-Hailing App to Diversified and Sustainable Platform Business Model
Authors: Ridwan Dewayanto Rusli
Abstract:
We show how prisoner's dilemma-type competition problems can be mitigated through rapid platform diversification and ecosystem expansion. We analyze a ride-hailing company in Southeast Asia, Gojek, whose network grew to more than 170 million users comprising consumers, partner drivers, merchants, and complementors within a few years and has already achieved higher contribution margins than ride-hailing peers Uber and Lyft. Its ecosystem integrates ride-hailing, food delivery and logistics, merchant solutions, e-commerce, marketplace and advertising, payments, and fintech offerings. The company continues growing its network of complementors and App developers, expanding content and gaining critical mass in consumer data analytics and advertising. We compare the company's growth and diversification trajectory with those of its main international rivals and peers. The company's rapid growth and future potential are analyzed using Cusumano's (2012) Staying Power and Six Principles, Hax and Wilde's (2003) and Hax's (2010) The Delta Model as well as Santos' (2016) home-market advantages frameworks. The recently announced multi-billion-dollar merger with one of Southeast Asia's largest e-commerce majors lends additional support to the above arguments.Keywords: ride-hailing, prisoner's dilemma, platform and ecosystem strategy, digital applications, diversification, home market advantages, e-commerce
Procedia PDF Downloads 948184 Optimization of Structures with Mixed Integer Non-linear Programming (MINLP)
Authors: Stojan Kravanja, Andrej Ivanič, Tomaž Žula
Abstract:
This contribution focuses on structural optimization in civil engineering using mixed integer non-linear programming (MINLP). MINLP is characterized as a versatile method that can handle both continuous and discrete optimization variables simultaneously. Continuous variables are used to optimize parameters such as dimensions, stresses, masses, or costs, while discrete variables represent binary decisions to determine the presence or absence of structural elements within a structure while also calculating discrete materials and standard sections. The optimization process is divided into three main steps. First, a mechanical superstructure with a variety of different topology-, material- and dimensional alternatives. Next, a MINLP model is formulated to encapsulate the optimization problem. Finally, an optimal solution is searched in the direction of the defined objective function while respecting the structural constraints. The economic or mass objective function of the material and labor costs of a structure is subjected to the constraints known from structural analysis. These constraints include equations for the calculation of internal forces and deflections, as well as equations for the dimensioning of structural components (in accordance with the Eurocode standards). Given the complex, non-convex and highly non-linear nature of optimization problems in civil engineering, the Modified Outer-Approximation/Equality-Relaxation (OA/ER) algorithm is applied. This algorithm alternately solves subproblems of non-linear programming (NLP) and main problems of mixed-integer linear programming (MILP), in this way gradually refines the solution space up to the optimal solution. The NLP corresponds to the continuous optimization of parameters (with fixed topology, discrete materials and standard dimensions, all determined in the previous MILP), while the MILP involves a global approximation to the superstructure of alternatives, where a new topology, materials, standard dimensions are determined. The optimization of a convex problem is stopped when the MILP solution becomes better than the best NLP solution. Otherwise, it is terminated when the NLP solution can no longer be improved. While the OA/ER algorithm, like all other algorithms, does not guarantee global optimality due to the presence of non-convex functions, various modifications, including convexity tests, are implemented in OA/ER to mitigate these difficulties. The effectiveness of the proposed MINLP approach is demonstrated by its application to various structural optimization tasks, such as mass optimization of steel buildings, cost optimization of timber halls, composite floor systems, etc. Special optimization models have been developed for the optimization of these structures. The MINLP optimizations, facilitated by the user-friendly software package MIPSYN, provide insights into a mass or cost-optimal solutions, optimal structural topologies, optimal material and standard cross-section choices, confirming MINLP as a valuable method for the optimization of structures in civil engineering.Keywords: MINLP, mixed-integer non-linear programming, optimization, structures
Procedia PDF Downloads 468183 SEAWIZARD-Multiplex AI-Enabled Graphene Based Lab-On-Chip Sensing Platform for Heavy Metal Ions Monitoring on Marine Water
Authors: M. Moreno, M. Alique, D. Otero, C. Delgado, P. Lacharmoise, L. Gracia, L. Pires, A. Moya
Abstract:
Marine environments are increasingly threatened by heavy metal contamination, including mercury (Hg), lead (Pb), and cadmium (Cd), posing significant risks to ecosystems and human health. Traditional monitoring techniques often fail to provide the spatial and temporal resolution needed for real-time detection of these contaminants, especially in remote or harsh environments. SEAWIZARD addresses these challenges by leveraging the flexibility, adaptability, and cost-effectiveness of printed electronics, with the integration of microfluidics to develop a compact, portable, and reusable sensor platform designed specifically for real-time monitoring of heavy metal ions in seawater. The SEAWIZARD sensor is a multiparametric Lab-on-Chip (LoC) device, a miniaturized system that integrates several laboratory functions into a single chip, drastically reducing sample volumes and improving adaptability. This platform integrates three printed graphene electrodes for the simultaneous detection of Hg, Cd and Pb via square wave voltammetry. These electrodes share the reference and the counter electrodes to improve space efficiency. Additionally, it integrates printed pH and temperature sensors to correct environmental interferences that may impact the accuracy of metal detection. The pH sensor is based on a carbon electrode with iridium oxide electrodeposited while the temperature sensor is graphene based. A protective dielectric layer is printed on top of the sensor to safeguard it in harsh marine conditions. The use of flexible polyethylene terephthalate (PET) as the substrate enables the sensor to conform to various surfaces and operate in challenging environments. One of the key innovations of SEAWIZARD is its integrated microfluidic layer, fabricated from cyclic olefin copolymer (COC). This microfluidic component allows a controlled flow of seawater over the sensing area, allowing for significant improved detection limits compared to direct water sampling. The system’s dual-channel design separates the detection of heavy metals from the measurement of pH and temperature, ensuring that each parameter is measured under optimal conditions. In addition, the temperature sensor is finely tuned with a serpentine-shaped microfluidic channel to ensure precise thermal measurements. SEAWIZARD also incorporates custom electronics that allow for wireless data transmission via Bluetooth, facilitating rapid data collection and user interface integration. Embedded artificial intelligence further enhances the platform by providing an automated alarm system, capable of detecting predefined metal concentration thresholds and issuing warnings when limits are exceeded. This predictive feature enables early warnings of potential environmental disasters, such as industrial spills or toxic levels of heavy metal pollutants, making SEAWIZARD not just a detection tool, but a comprehensive monitoring and early intervention system. In conclusion, SEAWIZARD represents a significant advancement in printed electronics applied to environmental sensing. By combining flexible, low-cost materials with advanced microfluidics, custom electronics, and AI-driven intelligence, SEAWIZARD offers a highly adaptable and scalable solution for real-time, high-resolution monitoring of heavy metals in marine environments. Its compact and portable design makes it an accessible, user-friendly tool with the potential to transform water quality monitoring practices and provide critical data to protect marine ecosystems from contamination-related risks.Keywords: lab-on-chip, printed electronics, real-time monitoring, microfluidics, heavy metal contamination
Procedia PDF Downloads 318182 Wasting Human and Computer Resources
Authors: Mária Csernoch, Piroska Biró
Abstract:
The legends about “user-friendly” and “easy-to-use” birotical tools (computer-related office tools) have been spreading and misleading end-users. This approach has led us to the extremely high number of incorrect documents, causing serious financial losses in the creating, modifying, and retrieving processes. Our research proved that there are at least two sources of this underachievement: (1) The lack of the definition of the correctly edited, formatted documents. Consequently, end-users do not know whether their methods and results are correct or not. They are not aware of their ignorance. They are so ignorant that their ignorance does not allow them to realize their lack of knowledge. (2) The end-users’ problem-solving methods. We have found that in non-traditional programming environments end-users apply, almost exclusively, surface approach metacognitive methods to carry out their computer related activities, which are proved less effective than deep approach methods. Based on these findings we have developed deep approach methods which are based on and adapted from traditional programming languages. In this study, we focus on the most popular type of birotical documents, the text-based documents. We have provided the definition of the correctly edited text, and based on this definition, adapted the debugging method known in programming. According to the method, before the realization of text editing, a thorough debugging of already existing texts and the categorization of errors are carried out. With this method in advance to real text editing users learn the requirements of text-based documents and also of the correctly formatted text. The method has been proved much more effective than the previously applied surface approach methods. The advantages of the method are that the real text handling requires much less human and computer sources than clicking aimlessly in the GUI (Graphical User Interface), and the data retrieval is much more effective than from error-prone documents.Keywords: deep approach metacognitive methods, error-prone birotical documents, financial losses, human and computer resources
Procedia PDF Downloads 3828181 Flexible Arm Manipulator Control for Industrial Tasks
Authors: Mircea Ivanescu, Nirvana Popescu, Decebal Popescu, Dorin Popescu
Abstract:
This paper addresses the control problem of a class of hyper-redundant arms. In order to avoid discrepancy between the mathematical model and the actual dynamics, the dynamic model with uncertain parameters of this class of manipulators is inferred. A procedure to design a feedback controller which stabilizes the uncertain system has been proposed. A PD boundary control algorithm is used in order to control the desired position of the manipulator. This controller is easy to implement from the point of view of measuring techniques and actuation. Numerical simulations verify the effectiveness of the presented methods. In order to verify the suitability of the control algorithm, a platform with a 3D flexible manipulator has been employed for testing. Experimental tests on this platform illustrate the applications of the techniques developed in the paper.Keywords: distributed model, flexible manipulator, observer, robot control
Procedia PDF Downloads 3218180 Creation of Processes for a Safety Element Out of Context for an Actuator Circuit Control Module
Authors: Hassan Noun, Christian Urban-Seelmann, Mohamed Abdelfattah, Guillaume Zeller, Rajesh G., Iryna Mozgova, Roland Lachmayer
Abstract:
Several modules in automotive are usually modified and adapted for various project-specific applications. Due to a standardized safety concept, high reusability is accessible. A safety element out of context (SEooC) according to ISO 26262 can be a suitable approach. Based on the same safety concept and analysis, common modules can reach high usability. For developing according to a module out of context, an appropriate and detailed development approach is required. This paper shows how to derive these development processes for platform modules. Therefore, the detailed approach to the safety element out of context is derived. The aim is to create a detailed workflow for all phases of the development and integration of any kind of system modules. As an application example, an automotive project for an actuator control module is considered.Keywords: functional safety, engineering processes, system engineering, electronic engineering
Procedia PDF Downloads 1458179 Development of Automated Quality Management System for the Management of Heat Networks
Authors: Nigina Toktasynova, Sholpan Sagyndykova, Zhanat Kenzhebayeva, Maksat Kalimoldayev, Mariya Ishimova, Irbulat Utepbergenov
Abstract:
Any business needs a stable operation and continuous improvement, therefore it is necessary to constantly interact with the environment, to analyze the work of the enterprise in terms of employees, executives and consumers, as well as to correct any inconsistencies of certain types of processes and their aggregate. In the case of heat supply organizations, in addition to suppliers, local legislation must be considered which often is the main regulator of pricing of services. In this case, the process approach used to build a functional organizational structure in these types of businesses in Kazakhstan is a challenge not only in the implementation, but also in ways of analyzing the employee's salary. To solve these problems, we investigated the management system of heating enterprise, including strategic planning based on the balanced scorecard (BSC), quality management in accordance with the standards of the Quality Management System (QMS) ISO 9001 and analysis of the system based on expert judgment using fuzzy inference. To carry out our work we used the theory of fuzzy sets, the QMS in accordance with ISO 9001, BSC according to the method of Kaplan and Norton, method of construction of business processes according to the notation IDEF0, theory of modeling using Matlab software simulation tools and graphical programming LabVIEW. The results of the work are as follows: We determined possibilities of improving the management of heat-supply plant-based on QMS; after the justification and adaptation of software tool it has been used to automate a series of functions for the management and reduction of resources and for the maintenance of the system up to date; an application for the analysis of the QMS based on fuzzy inference has been created with novel organization of communication software with the application enabling the analysis of relevant data of enterprise management system.Keywords: balanced scorecard, heat supply, quality management system, the theory of fuzzy sets
Procedia PDF Downloads 3678178 Viability of Irrigation Water Conservation Practices in the Low Desert of California
Authors: Ali Montazar
Abstract:
California and the Colorado River Basin are facing increasing uncertainty concerning water supplies. The Colorado River is the main source of irrigation water in the low desert of California. Currently, due to an increasing water-use competition and long-term drought at the Colorado River Basin, efficient use of irrigation water is one of the highest conservation priorities in the region. This study aims to present some of current irrigation technologies and management approaches in the low desert and assess the viability and potential of these water management practices. The results of several field experiments are used to assess five water conservation practices of sub-surface drip irrigation, automated surface irrigation, sprinkler irrigation, tail-water recovery system, and deficit irrigation strategy. The preliminary results of several ongoing studies at commercial fields are presented, particularly researches in alfalfa, sugar beets, kliengrass, sunflower, and spinach fields. The findings indicate that all these practices have significant potential to conserve water (an average of 1 ac-ft/ac) and enhance the efficiency of water use (15-25%). Further work is needed to better understand the feasibility of each of these applications and to help maintain profitable and sustainable agricultural production system in the low desert as water and labor costs, and environmental issues increase.Keywords: automated surface irrigation, deficit irrigation, low desert of California, sprinkler irrigation, sub-surface drip irrigation, tail-water recovery system
Procedia PDF Downloads 1588177 Platform Virtual for Joint Amplitude Measurement Based in MEMS
Authors: Mauro Callejas-Cuervo, Andrea C. Alarcon-Aldana, Andres F. Ruiz-Olaya, Juan C. Alvarez
Abstract:
Motion capture (MC) is the construction of a precise and accurate digital representation of a real motion. Systems have been used in the last years in a wide range of applications, from films special effects and animation, interactive entertainment, medicine, to high competitive sport where a maximum performance and low injury risk during training and competition is seeking. This paper presents an inertial and magnetic sensor based technological platform, intended for particular amplitude monitoring and telerehabilitation processes considering an efficient cost/technical considerations compromise. Our platform particularities offer high social impact possibilities by making telerehabilitation accessible to large population sectors in marginal socio-economic sector, especially in underdeveloped countries that in opposition to developed countries specialist are scarce, and high technology is not available or inexistent. This platform integrates high-resolution low-cost inertial and magnetic sensors with adequate user interfaces and communication protocols to perform a web or other communication networks available diagnosis service. The amplitude information is generated by sensors then transferred to a computing device with adequate interfaces to make it accessible to inexperienced personnel, providing a high social value. Amplitude measurements of the platform virtual system presented a good fit to its respective reference system. Analyzing the robotic arm results (estimation error RMSE 1=2.12° and estimation error RMSE 2=2.28°), it can be observed that during arm motion in any sense, the estimation error is negligible; in fact, error appears only during sense inversion what can easily be explained by the nature of inertial sensors and its relation to acceleration. Inertial sensors present a time constant delay which acts as a first order filter attenuating signals at large acceleration values as is the case for a change of sense in motion. It can be seen a damped response of platform virtual in other images where error analysis show that at maximum amplitude an underestimation of amplitude is present whereas at minimum amplitude estimations an overestimation of amplitude is observed. This work presents and describes the platform virtual as a motion capture system suitable for telerehabilitation with the cost - quality and precision - accessibility relations optimized. These particular characteristics achieved by efficiently using the state of the art of accessible generic technology in sensors and hardware, and adequate software for capture, transmission analysis and visualization, provides the capacity to offer good telerehabilitation services, reaching large more or less marginal populations where technologies and specialists are not available but accessible with basic communication networks.Keywords: inertial sensors, joint amplitude measurement, MEMS, telerehabilitation
Procedia PDF Downloads 2598176 Analysis of the AZF Region in Slovak Men with Azoospermia
Authors: J. Bernasovská, R. Lohajová Behulová, E. Petrejčiková, I. Boroňová, I. Bernasovský
Abstract:
Y chromosome microdeletions are the most common genetic cause of male infertility and screening for these microdeletions in azoospermic or severely oligospermic men is now standard practice. Analysis of the Y chromosome in men with azoospermia or severe oligozoospermia has resulted in the identification of three regions in the euchromatic part of the long arm of the human Y chromosome (Yq11) that are frequently deleted in men with otherwise unexplained spermatogenic failure. PCR analysis of microdeletions in the AZFa, AZFb and AZFc regions of the human Y chromosome is an important screening tool. The aim of this study was to analyse the type of microdeletions in men with fertility disorders in Slovakia. We evaluated 227 patients with azoospermia and with normal karyotype. All patient samples were analyzed cytogenetically. For PCR amplification of sequence-tagged sites (STS) of the AZFa, AZFb and AZFc regions of the Y chromosome was used Devyser AZF set. Fluorescently labeled primers for all markers in one multiplex PCR reaction were used and for automated visualization and identification of the STS markers we used genetic analyzer ABi 3500xl (Life Technologies). We reported 13 cases of deletions in the AZF region 5,73%. Particular types of deletions were recorded in each region AZFa,b,c .The presence of microdeletions in the AZFc region was the most frequent. The study confirmed that percentage of microdeletions in the AZF region is low in Slovak azoospermic patients, but important from a prognostic view.Keywords: AZF, male infertility, microdeletions, Y chromosome
Procedia PDF Downloads 3738175 Examining the Critical Factors for Success and Failure of Common Ticketing Systems
Authors: Tam Viet Hoang
Abstract:
With a plethora of new mobility services and payment systems found in our cities and across modern public transportation systems, several cities globally have turned to common ticketing systems to help navigate this complexity. Helping to create time and space-differentiated fare structures and tariff schemes, common ticketing systems can optimize transport utilization rates, achieve cost efficiencies, and provide key incentives to specific target groups. However, not all cities and transportation systems have enjoyed a smooth journey towards the adoption, roll-out, and servicing of common ticketing systems, with both the experiences of success and failure being attributed to a wide variety of critical factors. Using case study research as a methodology and cities as the main unit of analysis, this research will seek to address the fundamental question of “what are the critical factors for the success and failure of common ticketing systems?” Using rail/train systems as the entry point for this study will start by providing a background to the evolution of transport ticketing and justify the improvements in operational efficiency that can be achieved through common ticketing systems. Examining the socio-economic benefits of common ticketing, the research will also help to articulate the value derived for different key identified stakeholder groups. By reviewing case studies of the implementation of common ticketing systems in different cities, the research will explore lessons learned from cities with the aim to elicit factors to ensure seamless connectivity integrated e-ticketing platforms. In an increasingly digital age and where cities are now coming online, this paper seeks to unpack these critical factors, undertaking case study research drawing from literature and lived experiences. Offering us a better understanding of the enabling environment and ideal mixture of ingredients to facilitate the successful roll-out of a common ticketing system, interviews will be conducted with transport operators from several selected cities to better appreciate the challenges and strategies employed to overcome those challenges in relation to common ticketing systems. Meanwhile, as we begin to see the introduction of new mobile applications and user interfaces to facilitate ticketing and payment as part of the transport journey, we take stock of numerous policy challenges ahead and implications on city-wide and system-wide urban planning. It is hoped that this study will help to identify the critical factors for the success and failure of common ticketing systems for cities set to embark on their implementation while serving to fine-tune processes in those cities where common ticketing systems are already in place. Outcomes from the study will help to facilitate an improved understanding of common pitfalls and essential milestones towards the roll-out of a common ticketing system for railway systems, especially for emerging countries where mass rapid transit transport systems are being considered or in the process of construction.Keywords: common ticketing, public transport, urban strategies, Bangkok, Fukuoka, Sydney
Procedia PDF Downloads 898174 System Identification in Presence of Outliers
Authors: Chao Yu, Qing-Guo Wang, Dan Zhang
Abstract:
The outlier detection problem for dynamic systems is formulated as a matrix decomposition problem with low-rank, sparse matrices and further recast as a semidefinite programming (SDP) problem. A fast algorithm is presented to solve the resulting problem while keeping the solution matrix structure and it can greatly reduce the computational cost over the standard interior-point method. The computational burden is further reduced by proper construction of subsets of the raw data without violating low rank property of the involved matrix. The proposed method can make exact detection of outliers in case of no or little noise in output observations. In case of significant noise, a novel approach based on under-sampling with averaging is developed to denoise while retaining the saliency of outliers and so-filtered data enables successful outlier detection with the proposed method while the existing filtering methods fail. Use of recovered “clean” data from the proposed method can give much better parameter estimation compared with that based on the raw data.Keywords: outlier detection, system identification, matrix decomposition, low-rank matrix, sparsity, semidefinite programming, interior-point methods, denoising
Procedia PDF Downloads 3078173 Optimization of a Convolutional Neural Network for the Automated Diagnosis of Melanoma
Authors: Kemka C. Ihemelandu, Chukwuemeka U. Ihemelandu
Abstract:
The incidence of melanoma has been increasing rapidly over the past two decades, making melanoma a current public health crisis. Unfortunately, even as screening efforts continue to expand in an effort to ameliorate the death rate from melanoma, there is a need to improve diagnostic accuracy to decrease misdiagnosis. Artificial intelligence (AI) a new frontier in patient care has the ability to improve the accuracy of melanoma diagnosis. Convolutional neural network (CNN) a form of deep neural network, most commonly applied to analyze visual imagery, has been shown to outperform the human brain in pattern recognition. However, there are noted limitations with the accuracy of the CNN models. Our aim in this study was the optimization of convolutional neural network algorithms for the automated diagnosis of melanoma. We hypothesized that Optimal selection of the momentum and batch hyperparameter increases model accuracy. Our most successful model developed during this study, showed that optimal selection of momentum of 0.25, batch size of 2, led to a superior performance and a faster model training time, with an accuracy of ~ 83% after nine hours of training. We did notice a lack of diversity in the dataset used, with a noted class imbalance favoring lighter vs. darker skin tone. Training set image transformations did not result in a superior model performance in our study.Keywords: melanoma, convolutional neural network, momentum, batch hyperparameter
Procedia PDF Downloads 1018172 Prevalence of Dactylogyrus spp. and Butriocephalus spp. in Common Carp (Cyprinus carpio) from Mazandaran Province, Iran
Authors: Mohamad Bolandmrtabeh, Nafiseh Hasani, Saeid Abdi darakeh, Amir Heydari, Maryam Asghari
Abstract:
Dactylogyrus spp. and Butriocephalus spp. are among the most common parasites infecting farmed fish in Iran and worldwide. These parasites infect fish of all ages and are responsible for many lesions and mortalities in fingerlings and occasionally in grow-out and brood fish. Understanding the prevalence and intensity of these parasites appears necessary for their effective management and control. In this study, 60 common carp (Cyprinus carpio) were collected from 6 stations in Mazandaran Province, Iran, during summer 2023 and examined parasitologically. Common carp is a freshwater fish of the family Cyprinidae. The fish were transferred live to the laboratory after capture and were examined for parasites after anaesthetization and biometrics. The identified parasitic species included a Dactylogyrus sp. and a Butriocephalus sp. A total of 11 Dactylogyrus spp. were recovered from the gills of 60 fish examined. Additionally, 6 Butriocephalus spp. were collected from the intestine of infected fish. Based on taxonomic keys, the parasites were identified to genus level. The overall prevalence of infection was 23.3%, with approximately one quarter of all fish infected. A significant association was found between infections in different sampling stations. These results highlight the need for further research and planning for treatment and control in order to mitigate the impact of these parasitic diseases.Keywords: dactylogyrus, butriocephalus, common carp, cyprinus carpio
Procedia PDF Downloads 718171 Microfluidic Synthesis of Chlorophyll Extraction–Loaded PCL Composite Microparticles Developed as Health Food
Authors: Ching-Ju Hsiao, Mao-Chen Huang, Pei-Fan Chen, Ruo-Yun Chung, Jiun-Hua Chou, Chih-Hui Yang, Keng-Shiang Huang, Jei-Fu Shaw
Abstract:
Chlorophyll has many benefits for human body. It is known to improve the health of the circulatory, digestive, immune and detoxification systems of the body. However, Chl can’t be preserved at the environment of high temperature and light exposure for a long time due to it is chemical structure is easily degradable. This characteristic causes that human body is difficult to absorb Chl effective components. In order to solve this problem, we utilize polycaprolactone (PCL) polymer encapsulation technology to increase the stability of Chl. In particular, we also established a microfluidic platform provide the control of composite beads diameter. The new composite beads is potential to be a health food. Result show that Chl effective components via the microfludic platform can be encapsulated effectively and still preserve its effective components.Keywords: chlorophyll, PCL, PVA, microfluidic
Procedia PDF Downloads 557