Search results for: customized thresholding
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 299

Search results for: customized thresholding

89 Innovation Management in State-Owned-Enterprises in the Digital Transformation: An Empirical Case Study of Swiss Post

Authors: Jiayun Shen, Lorenz Wyss, Thierry Golliard, Matthias Finger

Abstract:

Innovation is widely recognized as the key for private enterprises to win the market competition. The state-owned-enterprises need to be innovative to compete in the market after the privatization as well. However, it is a lack of research to study how state-owned-enterprises manage innovation to create new products and services. Swiss Post, a Swiss state-owned-enterprises, has established a department to transform the corporate culture and foster innovation to achieve digital transformation. This paper describes the innovation management process at the Swiss Post and analyzes the impacts of the instruments, the organizational structure, and explores the barriers of innovation. This study used qualitative methods based on a review of the literature on innovation management and semi-structured interviews. Being established for over five years, the Swiss Post’s innovation management department has established a software-assisted modularized platform with systematic instruments to help the internal employees with the different innovation processes. It guides the innovators from idea creation to piloting in markets and supports with a separate financing source, with knowledge inputs and coaching, as well as with connections to external partners through the open innovation and venturing team. The platform also adapts to different business units within the corporate with a customized tailor for the various operational business units. The separate financing instruments enabled the creation and further development of new ideas; the coaching services contribute greatly to the transformation of teams’ innovation culture by providing new knowledge, thinking methods, and use cases for inspiration. It also facilitates organizational learning to help the whole corporate with the digital transformation. However, it is also confronted with a big challenge in twofold. Internally, the disruptive projects often hardly overcome the obstacles of long-established operational processes in the traditional business units; externally, the expectations of the public and restrictions from the federal government have become high hurdles for the company to stay and compete in the innovation track.

Keywords: empirical case study, innovation management, state-owned-enterprise, Swiss Post

Procedia PDF Downloads 98
88 Including Local Economic and Anthropometric Parameters in the Design of an Stand up Wheelchair

Authors: Urrutia Fernando, López Jessica, Sánchez Carlos, San Antonio Thalía

Abstract:

Ecuador, as a signatory country of the convention of the rights of persons with disabilities (CRPD) has, in the recent years, strengthened the structures and legal framework required to protect this minority comprised of 13.2% of its total population. However, the reality is that this group has disproportionately low earnings and low educational attainment in comparison with the general population. The main struggles, to promote job placement of wheelchairs users, are environmental discrimination caused by accessibility in structures and transportation, this mainly due to the cost, for private and public entities, of performing the reasonable accommodation they require. It is widely known that product development and production is needed to support effective implementation of the CRPD and that walking and standing are the major life activities, in this context the objective of this investigation is to promote job placement of wheelchair user in the province of Tungurahua by means of the design, production and marketing of a customized stand up wheelchair. Exploratory interviews and measurements were performed in a representative sample of working age wheelchairs users that develop their disability after achieving their physical maturity and that are capable of performing professional activities with their upper limbs, this in order to detect the user’s preference and determine the local economic and anthropometric parameters to be included in the wheelchair design. The findings reveal factors that uniquely impact quality of life and development for people with a mobility disability within the context of the province, first that transportation is a big issue since public buses does not have accessibility for wheelchair users and the absence of curb cuts and the presence of trash bins over the sidewalks among other hinders an economic independent mobility, second that the proposal based in the idea of modifying the wheelchairs to make it able to overcome certain obstacles helps people in wheelchair to improve their independent living and by reducing the costs of modification for the employer could improve their chances of finding work.

Keywords: anthropometrics, job placement, stand up wheelchair, user centered design

Procedia PDF Downloads 523
87 Development of Immersive Virtual Reality System for Planning of Cargo Loading Operations

Authors: Eugene Y. C. Wong, Daniel Y. W. Mo, Cosmo T. Y. Ng, Jessica K. Y. Chan, Leith K. Y. Chan, Henry Y. K. Lau

Abstract:

The real-time planning visualisation, precise allocation and loading optimisation in air cargo load planning operations are increasingly important as more considerations are needed on dangerous cargo loading, locations of lithium batteries, weight declaration and limited aircraft capacity. The planning of the unit load devices (ULD) can often be carried out only in a limited number of hours before flight departure. A dynamic air cargo load planning system is proposed with the optimisation of cargo load plan and visualisation of planning results in virtual reality systems. The system aims to optimise the cargo load planning and visualise the simulated loading planning decision on air cargo terminal operations. Adopting simulation tools, Cave Automatic Virtual Environment (CAVE) and virtual reality technologies, the results of planning with reference to weight and balance, Unit Load Device (ULD) dimensions, gateway, cargo nature and aircraft capacity are optimised and presented. The virtual reality system facilities planning, operations, education and training. Staff in terminals are usually trained in a traditional push-approach demonstration with enormous manual paperwork. With the support of newly customized immersive visualization environment, users can master the complex air cargo load planning techniques in a problem based training with the instant result being immersively visualised. The virtual reality system is developed with three-dimensional (3D) projectors, screens, workstations, truss system, 3D glasses, and demonstration platform and software. The content will be focused on the cargo planning and loading operations in an air cargo terminal. The system can assist decision-making process during cargo load planning in the complex operations of air cargo terminal operations. The processes of cargo loading, cargo build-up, security screening, and system monitoring can be further visualised. Scenarios are designed to support and demonstrate the daily operations of the air cargo terminal, including dangerous goods, pets and animals, and some special cargos.

Keywords: air cargo load planning, optimisation, virtual reality, weight and balance, unit load device

Procedia PDF Downloads 317
86 DEEPMOTILE: Motility Analysis of Human Spermatozoa Using Deep Learning in Sri Lankan Population

Authors: Chamika Chiran Perera, Dananjaya Perera, Chirath Dasanayake, Banuka Athuraliya

Abstract:

Male infertility is a major problem in the world, and it is a neglected and sensitive health issue in Sri Lanka. It can be determined by analyzing human semen samples. Sperm motility is one of many factors that can evaluate male’s fertility potential. In Sri Lanka, this analysis is performed manually. Manual methods are time consuming and depend on the person, but they are reliable and it can depend on the expert. Machine learning and deep learning technologies are currently being investigated to automate the spermatozoa motility analysis, and these methods are unreliable. These automatic methods tend to produce false positive results and false detection. Current automatic methods support different techniques, and some of them are very expensive. Due to the geographical variance in spermatozoa characteristics, current automatic methods are not reliable for motility analysis in Sri Lanka. The suggested system, DeepMotile, is to explore a method to analyze motility of human spermatozoa automatically and present it to the andrology laboratories to overcome current issues. DeepMotile is a novel deep learning method for analyzing spermatozoa motility parameters in the Sri Lankan population. To implement the current approach, Sri Lanka patient data were collected anonymously as a dataset, and glass slides were used as a low-cost technique to analyze semen samples. Current problem was identified as microscopic object detection and tackling the problem. YOLOv5 was customized and used as the object detector, and it achieved 94 % mAP (mean average precision), 86% Precision, and 90% Recall with the gathered dataset. StrongSORT was used as the object tracker, and it was validated with andrology experts due to the unavailability of annotated ground truth data. Furthermore, this research has identified many potential ways for further investigation, and andrology experts can use this system to analyze motility parameters with realistic accuracy.

Keywords: computer vision, deep learning, convolutional neural networks, multi-target tracking, microscopic object detection and tracking, male infertility detection, motility analysis of human spermatozoa

Procedia PDF Downloads 69
85 Assessing the Impact of Antiretroviral Mediated Drug-Drug Interactions on Piperaquine Antimalarial Treatment in Pregnant Women Using Physiologically Based Pharmacokinetic Modelling

Authors: Olusola Omolola Olafuyi, Michael Coleman, Raj Kumar Singh Badhan

Abstract:

Introduction: Malaria in pregnancy has morbidity and mortality implication on both mother and unborn child. Piperaquine (PQ) based antimalarial treatment is emerging as a choice antimalarial for pregnant women in the face of resistance to current antimalarial treatment recommendation in pregnancy. Physiological and biochemical changes in pregnant women may affect the pharmacokinetics of the antimalarial drug in these. In malaria endemic regions other infectious diseases like HIV/AIDs are prevalent. Pregnant women who are co-infected with malaria and HIV/AID are at even more greater risk of death not only due to complications of the diseases but also due to drug-drug interactions (DDIs) between antimalarials (AMT) and antiretroviral (ARVs). In this study, physiologically based pharmacokinetic (PBPK) modelling was used to investigate the effect of physiological and biochemical changes on the impact of ARV mediated DDIs in pregnant women in three countries. Method: A PBPK model for PQ was developed on SimCYP® using published physicochemical and pharmacokinetic data of PQ from literature, this was validated in three customized population groups from Thailand, Sudan and Papua New Guinea with clinical data. Validation of PQ model was also done in presence of interaction with efavirenz (pre-validated on SimCYP®). Different albumin levels and pregnancy stages was simulated in the presence of interaction with standard doses of efavirenz and ritonavir. PQ day 7 concentration of 30ng/ml was used as the efficacy endpoint for PQ treatment.. Results: The median day 7 concentration of PQ remained virtually consistent throughout pregnancy and were satisfactory across the three population groups ranging from 26-34.1ng/ml; this implied the efficacy of PQ throughout pregnancy. DDI interaction with ritonavir and efavirenz resulted in modest effect on the day 7 concentrations of PQ with AUCratio ranging from 0.56-0.8 and 1.64-1.79 for efavirenz and ritonavir respectively over 10-40 gestational weeks, however, a reduction in human serum albumin level reflective of severe malaria resulted in significantly reduced the number of subjects attaining the PQ day 7 concentration in the presence of both DDIs. The model demonstrated that the DDI between PQ and ARV in pregnant women with different malaria severities can alter the pharmacokinetic of PQ.

Keywords: antiretroviral, malaria, piperaquine, pregnancy, physiologically-based pharmacokinetics

Procedia PDF Downloads 156
84 End-Users Tools to Empower and Raise Awareness of Behavioural Change towards Energy Efficiency

Authors: G. Calleja-Rodriguez, N. Jimenez-Redondo, J. J. Peralta Escalante

Abstract:

This research work aims at developing a solution to take advantage of the potential energy saving related to occupants behaviour estimated in between 5-30 % according to existing studies. For that purpose, the following methodology has been followed: 1) literature review and gap analysis, 2) define concept and functional requirements, 3) evaluation and feedback by experts. As result, the concept for a tool-box that implements continuous behavior change interventions named as engagement methods and based on increasing energy literacy, increasing energy visibility, using bonus system, etc. has been defined. These engagement methods are deployed through a set of ICT tools: Building Automation and Control System (BACS) add-ons services installed in buildings and Users Apps installed in smartphones, smart-TVs or dashboards. The tool-box called eTEACHER identifies energy conservation measures (ECM) based on energy behavioral change through a what-if analysis that collects information about the building and its users (comfort feedback, behavior, etc.) and carry out cost-effective calculations to provide outputs such us efficient control settings of building systems. This information is processed and showed in an attractive way as tailored advice to the energy end-users. Therefore, eTEACHER goal is to change the behavior of building´s energy users towards energy efficiency, comfort and better health conditions by deploying customized ICT-based interventions taking into account building typology (schools, residential, offices, health care centres, etc.), users profile (occupants, owners, facility managers, employers, etc.) as well as cultural and demographic factors. One of the main findings of this work is the common failure when technological interventions on behavioural change are done to not consult, train and support users regarding technological changes leading to poor performance in practices. As conclusion, a strong need to carry out social studies to identify relevant behavioural issues and to identify effective pro-evironmental behavioral change strategies has been identified.

Keywords: energy saving, behavioral bhange, building users, engagement methods, energy conservation measures

Procedia PDF Downloads 143
83 Exploring the Potential of Modular Housing Designs for the Emergency Housing Need in Türkiye after the February Earthquake in 2023

Authors: Hailemikael Negussie, Sebla Arın Ensarioğlu

Abstract:

In February 2023 Southeastern Türkiye and Northwestern Syria were hit by two consecutive earthquakes with high magnitude leaving thousands dead and thousands more homeless. The housing crisis in the affected areas has resulted in the need for a fast and qualified solution. There are a number of solutions, one of which is the use of modular designs to rebuild the cities that have been affected. Modular designs are prefabricated building components that can be quickly and efficiently assembled on-site, making them ideal to build structures with faster speed and higher quality. These structures are flexible, adaptable, and can be customized to meet the specific needs of the inhabitants, in addition to being more energy-efficient and sustainable. The prefabricated nature also assures that the quality of the products can be easily controlled. The reason for the collapse of most of the buildings during the earthquakes was found out to be the lack of quality during the construction stage. Using modular designs allows a higher control over the quality of the construction materials being used. The use of modular designs for a project of this scale presents some challenges, including the high upfront cost to design and manufacture components. However, if implemented correctly, modular designs can offer an effective and efficient solution to the urgent housing needs. The aim of this paper is to explore the potential of modular housing for mid- and long-term earthquake-resistant housing needs in the affected disaster zones after the earthquakes of February 2023. In the scope of this paper the adaptability of modular, prefabricated housing designs for the post-disaster environment, the advantages and disadvantages of this system will be examined. Elements such as; the current conditions of the region where the destruction happened, climatic data, topographic factors will be examined. Additionally, the paper will examine; examples of similar local and international modular post-earthquake housing projects. The region is projected to enter a rapid reconstruction phase in the following periods. Therefore, this paper will present a proposal for a system that can be used to produce safe and healthy urbanization policies without causing new aggrievements while meeting the housing needs of the people in the affected regions.

Keywords: post-disaster housing, earthquake-resistant design, modular design, housing, Türkiye

Procedia PDF Downloads 55
82 Digital Immunity System for Healthcare Data Security

Authors: Nihar Bheda

Abstract:

Protecting digital assets such as networks, systems, and data from advanced cyber threats is the aim of Digital Immunity Systems (DIS), which are a subset of cybersecurity. With features like continuous monitoring, coordinated reactions, and long-term adaptation, DIS seeks to mimic biological immunity. This minimizes downtime by automatically identifying and eliminating threats. Traditional security measures, such as firewalls and antivirus software, are insufficient for enterprises, such as healthcare providers, given the rapid evolution of cyber threats. The number of medical record breaches that have occurred in recent years is proof that attackers are finding healthcare data to be an increasingly valuable target. However, obstacles to enhancing security include outdated systems, financial limitations, and a lack of knowledge. DIS is an advancement in cyber defenses designed specifically for healthcare settings. Protection akin to an "immune system" is produced by core capabilities such as anomaly detection, access controls, and policy enforcement. Coordination of responses across IT infrastructure to contain attacks is made possible by automation and orchestration. Massive amounts of data are analyzed by AI and machine learning to find new threats. After an incident, self-healing enables services to resume quickly. The implementation of DIS is consistent with the healthcare industry's urgent requirement for resilient data security in light of evolving risks and strict guidelines. With resilient systems, it can help organizations lower business risk, minimize the effects of breaches, and preserve patient care continuity. DIS will be essential for protecting a variety of environments, including cloud computing and the Internet of medical devices, as healthcare providers quickly adopt new technologies. DIS lowers traditional security overhead for IT departments and offers automated protection, even though it requires an initial investment. In the near future, DIS may prove to be essential for small clinics, blood banks, imaging centers, large hospitals, and other healthcare organizations. Cyber resilience can become attainable for the whole healthcare ecosystem with customized DIS implementations.

Keywords: digital immunity system, cybersecurity, healthcare data, emerging technology

Procedia PDF Downloads 35
81 A Web-Based Systems Immunology Toolkit Allowing the Visualization and Comparative Analysis of Publically Available Collective Data to Decipher Immune Regulation in Early Life

Authors: Mahbuba Rahman, Sabri Boughorbel, Scott Presnell, Charlie Quinn, Darawan Rinchai, Damien Chaussabel, Nico Marr

Abstract:

Collections of large-scale datasets made available in public repositories can be used to identify and fill gaps in biomedical knowledge. But first, these data need to be made readily accessible to researchers for analysis and interpretation. Here a collection of transcriptome datasets was made available to investigate the functional programming of human hematopoietic cells in early life. Thirty two datasets were retrieved from the NCBI Gene Expression Omnibus (GEO) and loaded in a custom, interactive web application called the Gene Expression browser (GXB), designed for visualization and query of integrated large-scale data. Multiple sample groupings and gene rank lists were created based on the study design and variables in each dataset. Web links to customized graphical views can be generated by users and subsequently be used to graphically present data in manuscripts for publication. The GXB tool also enables browsing of a single gene across datasets, which can provide information on the role of a given molecule across biological systems. The dataset collection is available online. As a proof-of-principle, one of the datasets (GSE25087) was re-analyzed to identify genes that are differentially expressed by regulatory T cells in early life. Re-analysis of this dataset and a cross-study comparison using multiple other datasets in the above mentioned collection revealed that PMCH, a gene encoding a precursor of melanin-concentrating hormone (MCH), a cyclic neuropeptide, is highly expressed in a variety of other hematopoietic cell types, including neonatal erythroid cells as well as plasmacytoid dendritic cells upon viral infection. Our findings suggest an as yet unrecognized role of MCH in immune regulation, thereby highlighting the unique potential of the curated dataset collection and systems biology approach to generate new hypotheses which can be tested in future mechanistic studies.

Keywords: early-life, GEO datasets, PMCH, interactive query, systems biology

Procedia PDF Downloads 268
80 Mobile Phone Text Reminders and Voice Call Follow-ups Improve Attendance for Community Retail Pharmacy Refills; Learnings from Lango Sub-region in Northern Uganda

Authors: Jonathan Ogwal, Louis H. Kamulegeya, John M. Bwanika, Davis Musinguzi

Abstract:

Introduction: Community retail Pharmacy drug distribution points (CRPDDP) were implemented in the Lango sub-region as part of the Ministry of Health’s response to improving access and adherence to antiretroviral treatment (ART). Clients received their ART refills from nearby local pharmacies; as such, the need for continuous engagement through mobile phone appointment reminders and health messages. We share learnings from the implementation of mobile text reminders and voice call follow-ups among ART clients attending the CRPDDP program in northern Uganda. Methods: A retrospective data review of electronic medical records from four pharmacies allocated for CRPDDP in the Lira and Apac districts of the Lango sub-region in Northern Uganda was done from February to August 2022. The process involved collecting phone contacts of eligible clients from the health facility appointment register and uploading them onto a messaging platform customized by Rapid-pro, an open-source software. Client information, including code name, phone number, next appointment date, and the allocated pharmacy for ART refill, was collected and kept confidential. Contacts received appointment reminder messages and other messages on positive living as an ART client. Routine voice call follow-ups were done to ascertain the picking of ART from the refill pharmacy. Findings: In total, 1,354 clients were reached from the four allocated pharmacies found in urban centers. 972 clients received short message service (SMS) appointment reminders, and 382 were followed up through voice calls. The majority (75%) of the clients returned for refills on the appointed date, 20% returned within four days after the appointment date, and the remaining 5% needed follow-up where they reported that they were not in the district by the appointment date due to other engagements. Conclusion: The use of mobile text reminders and voice call follow-ups improves the attendance of community retail pharmacy refills.

Keywords: antiretroviral treatment, community retail drug distribution points, mobile text reminders, voice call follow-up

Procedia PDF Downloads 77
79 Customer Acquisition through Time-Aware Marketing Campaign Analysis in Banking Industry

Authors: Harneet Walia, Morteza Zihayat

Abstract:

Customer acquisition has become one of the critical issues of any business in the 21st century; having a healthy customer base is the essential asset of the bank business. Term deposits act as a major source of cheap funds for the banks to invest and benefit from interest rate arbitrage. To attract customers, the marketing campaigns at most financial institutions consist of multiple outbound telephonic calls with more than one contact to a customer which is a very time-consuming process. Therefore, customized direct marketing has become more critical than ever for attracting new clients. As customer acquisition is becoming more difficult to archive, having an intelligent and redefined list is necessary to sell a product smartly. Our aim of this research is to increase the effectiveness of campaigns by predicting customers who will most likely subscribe to the fixed deposit and suggest the most suitable month to reach out to customers. We design a Time Aware Upsell Prediction Framework (TAUPF) using two different approaches, with an aim to find the best approach and technique to build the prediction model. TAUPF is implemented using Upsell Prediction Approach (UPA) and Clustered Upsell Prediction Approach (CUPA). We also address the data imbalance problem by examining and comparing different methods of sampling (Up-sampling and down-sampling). Our results have shown building such a model is quite feasible and profitable for the financial institutions. The Time Aware Upsell Prediction Framework (TAUPF) can be easily used in any industry such as telecom, automobile, tourism, etc. where the TAUPF (Clustered Upsell Prediction Approach (CUPA) or Upsell Prediction Approach (UPA)) holds valid. In our case, CUPA books more reliable. As proven in our research, one of the most important challenges is to define measures which have enough predictive power as the subscription to a fixed deposit depends on highly ambiguous situations and cannot be easily isolated. While we have shown the practicality of time-aware upsell prediction model where financial institutions can benefit from contacting the customers at the specified month, further research needs to be done to understand the specific time of the day. In addition, a further empirical/pilot study on real live customer needs to be conducted to prove the effectiveness of the model in the real world.

Keywords: customer acquisition, predictive analysis, targeted marketing, time-aware analysis

Procedia PDF Downloads 95
78 Students' Performance, Perception and Attitude towards Interactive Online Modules to Improve Undergraduate Quantitative Skills in Biological Science

Authors: C. Suphioglu , V. Simbag, J. Markham, C. Coady, S. Belward, G. Di Trapani, P. Chunduri, J. Chuck, Y. Hodgson, L. Lluka, L. Poladian, D. Watters

Abstract:

Advances in science have made quantitative skills (QS) an essential graduate outcome for undergraduate science programs in Australia and other parts of the world. However, many students entering into degrees in Australian universities either lack these skills or have little confidence in their ability to apply them in their biological science units. It has been previously reported that integration of quantitative skills into life science programs appears to have a positive effect on student attitudes towards the importance of mathematics and statistics in biological sciences. It has also been noted that there is deficiency in QS resources available and applicable to undergraduate science students in Australia. MathBench (http://mathbench.umd.edu) is a series of online modules involving quantitative biology scenarios developed by the University of Maryland. Through collaboration with Australian universities, a project was funded by the Australian government through its Office for Learning and Teaching (OLT) to develop customized MathBench biology modules to promote the quantitative skills of undergraduate biology students in Australia. This presentation will focus on the assessment of changes in performance, perception and attitude of students in a third year Cellular Physiology unit after use of interactive online cellular diffusion modules modified for the Australian context. The modules have been designed to integrate QS into the biological science curriculum using familiar scenarios and informal language and providing students with the opportunity to review solutions to diffusion QS-related problems with interactive graphics. This paper will discuss results of pre and post MathBench quizzes composed of general and module specific questions that assessed change in student QS after MathBench; and pre and post surveys, administered before and after using MathBench modules to evaluate the students’ change in perception towards the influence of the modules, their attitude towards QS and on the development of their confidence in completing the inquiry-based activity as well as changes to their appreciation of the relevance of mathematics to cellular processes. Results will be compared to changes reported by Thompson et al., (2010) at the University of Maryland and implications for further integration of interactive online activities in the curriculum will be explored and discussed.

Keywords: quantitative skills, MathBench, maths in biology

Procedia PDF Downloads 353
77 The Analyzer: Clustering Based System for Improving Business Productivity by Analyzing User Profiles to Enhance Human Computer Interaction

Authors: Dona Shaini Abhilasha Nanayakkara, Kurugamage Jude Pravinda Gregory Perera

Abstract:

E-commerce platforms have revolutionized the shopping experience, offering convenient ways for consumers to make purchases. To improve interactions with customers and optimize marketing strategies, it is essential for businesses to understand user behavior, preferences, and needs on these platforms. This paper focuses on recommending businesses to customize interactions with users based on their behavioral patterns, leveraging data-driven analysis and machine learning techniques. Businesses can improve engagement and boost the adoption of e-commerce platforms by aligning behavioral patterns with user goals of usability and satisfaction. We propose TheAnalyzer, a clustering-based system designed to enhance business productivity by analyzing user-profiles and improving human-computer interaction. The Analyzer seamlessly integrates with business applications, collecting relevant data points based on users' natural interactions without additional burdens such as questionnaires or surveys. It defines five key user analytics as features for its dataset, which are easily captured through users' interactions with e-commerce platforms. This research presents a study demonstrating the successful distinction of users into specific groups based on the five key analytics considered by TheAnalyzer. With the assistance of domain experts, customized business rules can be attached to each group, enabling The Analyzer to influence business applications and provide an enhanced personalized user experience. The outcomes are evaluated quantitatively and qualitatively, demonstrating that utilizing TheAnalyzer’s capabilities can optimize business outcomes, enhance customer satisfaction, and drive sustainable growth. The findings of this research contribute to the advancement of personalized interactions in e-commerce platforms. By leveraging user behavioral patterns and analyzing both new and existing users, businesses can effectively tailor their interactions to improve customer satisfaction, loyalty and ultimately drive sales.

Keywords: data clustering, data standardization, dimensionality reduction, human computer interaction, user profiling

Procedia PDF Downloads 36
76 Automated, Objective Assessment of Pilot Performance in Simulated Environment

Authors: Maciej Zasuwa, Grzegorz Ptasinski, Antoni Kopyt

Abstract:

Nowadays flight simulators offer tremendous possibilities for safe and cost-effective pilot training, by utilization of powerful, computational tools. Due to technology outpacing methodology, vast majority of training related work is done by human instructors. It makes assessment not efficient, and vulnerable to instructors’ subjectivity. The research presents an Objective Assessment Tool (gOAT) developed at the Warsaw University of Technology, and tested on SW-4 helicopter flight simulator. The tool uses database of the predefined manoeuvres, defined and integrated to the virtual environment. These were implemented, basing on Aeronautical Design Standard Performance Specification Handling Qualities Requirements for Military Rotorcraft (ADS-33), with predefined Mission-Task-Elements (MTEs). The core element of the gOAT enhanced algorithm that provides instructor a new set of information. In details, a set of objective flight parameters fused with report about psychophysical state of the pilot. While the pilot performs the task, the gOAT system automatically calculates performance using the embedded algorithms, data registered by the simulator software (position, orientation, velocity, etc.), as well as measurements of physiological changes of pilot’s psychophysiological state (temperature, sweating, heart rate). Complete set of measurements is presented on-line to instructor’s station and shown in dedicated graphical interface. The presented tool is based on open source solutions, and flexible for editing. Additional manoeuvres can be easily added using guide developed by authors, and MTEs can be changed by instructor even during an exercise. Algorithm and measurements used allow not only to implement basic stress level measurements, but also to reduce instructor’s workload significantly. Tool developed can be used for training purpose, as well as periodical checks of the aircrew. Flexibility and ease of modifications allow the further development to be wide ranged, and the tool to be customized. Depending on simulation purpose, gOAT can be adjusted to support simulator of aircraft, helicopter, or unmanned aerial vehicle (UAV).

Keywords: automated assessment, flight simulator, human factors, pilot training

Procedia PDF Downloads 124
75 Evaluation of Main Factors Affecting the Choice of a Freight Forwarder: A Sri Lankan Exporter’s Perspective

Authors: Ishani Maheshika

Abstract:

The intermediary role performed by freight forwarders in exportation has become significant in fulfilling businesses’ supply chain needs in this dynamic world. Since the success of exporter’s business is at present, highly reliant on supply chain optimization, cost efficiency, profitability, consistent service and responsiveness, the decision of selecting the most beneficial freight forwarder has become crucial for exporters. Although there are similar foreign researches, prior researches covering Sri Lankan setting are not in existence. Moreover, results vary with time, nature of industry and business environment factors. Therefore, a study from the perspective of Sri Lankan exporters was identified as a requisite to be researched. In order to identify and prioritize key factors which have affected the exporter’s decision in selecting freight forwarders in Sri Lankan context, Sri Lankan export industry was stratified into 22 sectors based on commodity using stratified sampling technique. One exporter from each sector was then selected using judgmental sampling to have a sample of 22. Factors which were identified through a pilot survey, was organized under 6 main criteria. A questionnaire was basically developed as pairwise comparisons using 9-point semantic differential scale and comparisons were done within main criteria and subcriteria. After a pre-testing, interviews and e-mail questionnaire survey were conducted. Data were analyzed using Analytic Hierarchy Process to determine priority vectors of criteria. Customer service was found to be the most important main criterion for Sri Lankan exporters. It was followed by reliability and operational efficiency respectively. The criterion of the least importance is company background and reputation. Whereas small sized exporters pay more attention to rate, reliability is the major concern among medium and large scale exporters. Irrespective of seniority of the exporter, reliability is given the prominence. Responsiveness is the most important sub criterion among Sri Lankan exporters. Consistency of judgments with respect to main criteria was verified through consistency ratio, which was less than 10%. Being more competitive, freight forwarders should come up with customized marketing strategies based on each target group’s requirements and expectations in offering services to retain existing exporters and attract new exporters.

Keywords: analytic hierarchy process, freight forwarders, main criteria, Sri Lankan exporters, subcriteria

Procedia PDF Downloads 380
74 Edge Enhancement Visual Methodology for Fat Amount and Distribution Assessment in Dry-Cured Ham Slices

Authors: Silvia Grassi, Stefano Schiavon, Ernestina Casiraghi, Cristina Alamprese

Abstract:

Dry-cured ham is an uncooked meat product particularly appreciated for its peculiar sensory traits among which lipid component plays a key role in defining quality and, consequently, consumers’ acceptability. Usually, fat content and distribution are chemically determined by expensive, time-consuming, and destructive analyses. Moreover, different sensory techniques are applied to assess product conformity to desired standards. In this context, visual systems are getting a foothold in the meat market envisioning more reliable and time-saving assessment of food quality traits. The present work aims at developing a simple but systematic and objective visual methodology to assess the fat amount of dry-cured ham slices, in terms of total, intermuscular and intramuscular fractions. To the aim, 160 slices from 80 PDO dry-cured hams were evaluated by digital image analysis and Soxhlet extraction. RGB images were captured by a flatbed scanner, converted in grey-scale images, and segmented based on intensity histograms as well as on a multi-stage algorithm aimed at edge enhancement. The latter was performed applying the Canny algorithm, which consists of image noise reduction, calculation of the intensity gradient for each image, spurious response removal, actual thresholding on corrected images, and confirmation of strong edge boundaries. The approach allowed for the automatic calculation of total, intermuscular and intramuscular fat fractions as percentages of the total slice area. Linear regression models were run to estimate the relationships between the image analysis results and the chemical data, thus allowing for the prediction of the total, intermuscular and intramuscular fat content by the dry-cured ham images. The goodness of fit of the obtained models was confirmed in terms of coefficient of determination (R²), hypothesis testing and pattern of residuals. Good regression models have been found being 0.73, 0.82, and 0.73 the R2 values for the total fat, the sum of intermuscular and intramuscular fat and the intermuscular fraction, respectively. In conclusion, the edge enhancement visual procedure brought to a good fat segmentation making the simple visual approach for the quantification of the different fat fractions in dry-cured ham slices sufficiently simple, accurate and precise. The presented image analysis approach steers towards the development of instruments that can overcome destructive, tedious and time-consuming chemical determinations. As future perspectives, the results of the proposed image analysis methodology will be compared with those of sensory tests in order to develop a fast grading method of dry-cured hams based on fat distribution. Therefore, the system will be able not only to predict the actual fat content but it will also reflect the visual appearance of samples as perceived by consumers.

Keywords: dry-cured ham, edge detection algorithm, fat content, image analysis

Procedia PDF Downloads 150
73 Innovations in the Implementation of Preventive Strategies and Measuring Their Effectiveness Towards the Prevention of Harmful Incidents to People with Mental Disabilities who Receive Home and Community Based Services

Authors: Carlos V. Gonzalez

Abstract:

Background: Providers of in-home and community based services strive for the elimination of preventable harm to the people under their care as well as to the employees who support them. Traditional models of safety and protection from harm have assumed that the absence of incidents of harm is a good indicator of safe practices. However, this model creates an illusion of safety that is easily shaken by sudden and inadvertent harmful events. As an alternative, we have developed and implemented an evidence-based resilient model of safety known as C.O.P.E. (Caring, Observing, Predicting and Evaluating). Within this model, safety is not defined by the absence of harmful incidents, but by the presence of continuous monitoring, anticipation, learning, and rapid response to events that may lead to harm. Objective: The objective was to evaluate the effectiveness of the C.O.P.E. model for the reduction of harm to individuals with mental disabilities who receive home and community based services. Methods: Over the course of 2 years we counted the number of incidents of harm and near misses. We trained employees on strategies to eliminate incidents before they fully escalated. We trained employees to track different levels of patient status within a scale from 0 to 10. Additionally, we provided direct support professionals and supervisors with customized smart phone applications to track and notify the team of changes in that status every 30 minutes. Finally, the information that we collected was saved in a private computer network that analyzes and graphs the outcome of each incident. Result and conclusions: The use of the COPE model resulted in: A reduction in incidents of harm. A reduction the use of restraints and other physical interventions. An increase in Direct Support Professional’s ability to detect and respond to health problems. Improvement in employee alertness by decreasing sleeping on duty. Improvement in caring and positive interaction between Direct Support Professionals and the person who is supported. Developing a method to globally measure and assess the effectiveness of prevention from harm plans. Future applications of the COPE model for the reduction of harm to people who receive home and community based services are discussed.

Keywords: harm, patients, resilience, safety, mental illness, disability

Procedia PDF Downloads 424
72 Artificial Intelligence Models for Detecting Spatiotemporal Crop Water Stress in Automating Irrigation Scheduling: A Review

Authors: Elham Koohi, Silvio Jose Gumiere, Hossein Bonakdari, Saeid Homayouni

Abstract:

Water used in agricultural crops can be managed by irrigation scheduling based on soil moisture levels and plant water stress thresholds. Automated irrigation scheduling limits crop physiological damage and yield reduction. Knowledge of crop water stress monitoring approaches can be effective in optimizing the use of agricultural water. Understanding the physiological mechanisms of crop responding and adapting to water deficit ensures sustainable agricultural management and food supply. This aim could be achieved by analyzing and diagnosing crop characteristics and their interlinkage with the surrounding environment. Assessments of plant functional types (e.g., leaf area and structure, tree height, rate of evapotranspiration, rate of photosynthesis), controlling changes, and irrigated areas mapping. Calculating thresholds of soil water content parameters, crop water use efficiency, and Nitrogen status make irrigation scheduling decisions more accurate by preventing water limitations between irrigations. Combining Remote Sensing (RS), the Internet of Things (IoT), Artificial Intelligence (AI), and Machine Learning Algorithms (MLAs) can improve measurement accuracies and automate irrigation scheduling. This paper is a review structured by surveying about 100 recent research studies to analyze varied approaches in terms of providing high spatial and temporal resolution mapping, sensor-based Variable Rate Application (VRA) mapping, the relation between spectral and thermal reflectance and different features of crop and soil. The other objective is to assess RS indices formed by choosing specific reflectance bands and identifying the correct spectral band to optimize classification techniques and analyze Proximal Optical Sensors (POSs) to control changes. The innovation of this paper can be defined as categorizing evaluation methodologies of precision irrigation (applying the right practice, at the right place, at the right time, with the right quantity) controlled by soil moisture levels and sensitiveness of crops to water stress, into pre-processing, processing (retrieval algorithms), and post-processing parts. Then, the main idea of this research is to analyze the error reasons and/or values in employing different approaches in three proposed parts reported by recent studies. Additionally, as an overview conclusion tried to decompose different approaches to optimizing indices, calibration methods for the sensors, thresholding and prediction models prone to errors, and improvements in classification accuracy for mapping changes.

Keywords: agricultural crops, crop water stress detection, irrigation scheduling, precision agriculture, remote sensing

Procedia PDF Downloads 41
71 A Smart Sensor Network Approach Using Affordable River Water Level Sensors

Authors: Dian Zhang, Brendan Heery, Maria O’Neill, Ciprian Briciu-Burghina, Noel E. O’Connor, Fiona Regan

Abstract:

Recent developments in sensors, wireless data communication and the cloud computing have brought the sensor web to a whole new generation. The introduction of the concept of ‘Internet of Thing (IoT)’ has brought the sensor research into a new level, which involves the developing of long lasting, low cost, environment friendly and smart sensors; new wireless data communication technologies; big data analytics algorithms and cloud based solutions that are tailored to large scale smart sensor network. The next generation of smart sensor network consists of several layers: physical layer, where all the smart sensors resident and data pre-processes occur, either on the sensor itself or field gateway; data transmission layer, where data and instructions exchanges happen; the data process layer, where meaningful information is extracted and organized from the pre-process data stream. There are many definitions of smart sensor, however, to summarize all these definitions, a smart sensor must be Intelligent and Adaptable. In future large scale sensor network, collected data are far too large for traditional applications to send, store or process. The sensor unit must be intelligent that pre-processes collected data locally on board (this process may occur on field gateway depends on the sensor network structure). In this case study, three smart sensing methods, corresponding to simple thresholding, statistical model and machine learning based MoPBAS method, are introduced and their strength and weakness are discussed as an introduction to the smart sensing concept. Data fusion, the integration of data and knowledge from multiple sources, are key components of the next generation smart sensor network. For example, in the water level monitoring system, weather forecast can be extracted from external sources and if a heavy rainfall is expected, the server can send instructions to the sensor notes to, for instance, increase the sampling rate or switch on the sleeping mode vice versa. In this paper, we describe the deployment of 11 affordable water level sensors in the Dublin catchment. The objective of this paper is to use the deployed river level sensor network at the Dodder catchment in Dublin, Ireland as a case study to give a vision of the next generation of a smart sensor network for flood monitoring to assist agencies in making decisions about deploying resources in the case of a severe flood event. Some of the deployed sensors are located alongside traditional water level sensors for validation purposes. Using the 11 deployed river level sensors in a network as a case study, a vision of the next generation of smart sensor network is proposed. Each key component of the smart sensor network is discussed, which hopefully inspires the researchers who are working in the sensor research domain.

Keywords: smart sensing, internet of things, water level sensor, flooding

Procedia PDF Downloads 350
70 Handling, Exporting and Archiving Automated Mineralogy Data Using TESCAN TIMA

Authors: Marek Dosbaba

Abstract:

Within the mining sector, SEM-based Automated Mineralogy (AM) has been the standard application for quickly and efficiently handling mineral processing tasks. Over the last decade, the trend has been to analyze larger numbers of samples, often with a higher level of detail. This has necessitated a shift from interactive sample analysis performed by an operator using a SEM, to an increased reliance on offline processing to analyze and report the data. In response to this trend, TESCAN TIMA Mineral Analyzer is designed to quickly create a virtual copy of the studied samples, thereby preserving all the necessary information. Depending on the selected data acquisition mode, TESCAN TIMA can perform hyperspectral mapping and save an X-ray spectrum for each pixel or segment, respectively. This approach allows the user to browse through elemental distribution maps of all elements detectable by means of energy dispersive spectroscopy. Re-evaluation of the existing data for the presence of previously unconsidered elements is possible without the need to repeat the analysis. Additional tiers of data such as a secondary electron or cathodoluminescence images can also be recorded. To take full advantage of these information-rich datasets, TIMA utilizes a new archiving tool introduced by TESCAN. The dataset size can be reduced for long-term storage and all information can be recovered on-demand in case of renewed interest. TESCAN TIMA is optimized for network storage of its datasets because of the larger data storage capacity of servers compared to local drives, which also allows multiple users to access the data remotely. This goes hand in hand with the support of remote control for the entire data acquisition process. TESCAN also brings a newly extended open-source data format that allows other applications to extract, process and report AM data. This offers the ability to link TIMA data to large databases feeding plant performance dashboards or geometallurgical models. The traditional tabular particle-by-particle or grain-by-grain export process is preserved and can be customized with scripts to include user-defined particle/grain properties.

Keywords: Tescan, electron microscopy, mineralogy, SEM, automated mineralogy, database, TESCAN TIMA, open format, archiving, big data

Procedia PDF Downloads 83
69 Carbon Based Wearable Patch Devices for Real-Time Electrocardiography Monitoring

Authors: Hachul Jung, Ahee Kim, Sanghoon Lee, Dahye Kwon, Songwoo Yoon, Jinhee Moon

Abstract:

We fabricated a wearable patch device including novel patch type flexible dry electrode based on carbon nanofibers (CNFs) and silicone-based elastomer (MED 6215) for real-time ECG monitoring. There are many methods to make flexible conductive polymer by mixing metal or carbon-based nanoparticles. In this study, CNFs are selected for conductive nanoparticles because carbon nanotubes (CNTs) are difficult to disperse uniformly in elastomer compare with CNFs and silver nanowires are relatively high cost and easily oxidized in the air. Wearable patch is composed of 2 parts that dry electrode parts for recording bio signal and sticky patch parts for mounting on the skin. Dry electrode parts were made by vortexer and baking in prepared mold. To optimize electrical performance and diffusion degree of uniformity, we developed unique mixing and baking process. Secondly, sticky patch parts were made by patterning and detaching from smooth surface substrate after spin-coating soft skin adhesive. In this process, attachable and detachable strengths of sticky patch are measured and optimized for them, using a monitoring system. Assembled patch is flexible, stretchable, easily skin mountable and connectable directly with the system. To evaluate the performance of electrical characteristics and ECG (Electrocardiography) recording, wearable patch was tested by changing concentrations of CNFs and thickness of the dry electrode. In these results, the CNF concentration and thickness of dry electrodes were important variables to obtain high-quality ECG signals without incidental distractions. Cytotoxicity test is conducted to prove biocompatibility, and long-term wearing test showed no skin reactions such as itching or erythema. To minimize noises from motion artifacts and line noise, we make the customized wireless, light-weight data acquisition system. Measured ECG Signals from this system are stable and successfully monitored simultaneously. To sum up, we could fully utilize fabricated wearable patch devices for real-time ECG monitoring easily.

Keywords: carbon nanofibers, ECG monitoring, flexible dry electrode, wearable patch

Procedia PDF Downloads 156
68 Unlocking Health Insights: Studying Data for Better Care

Authors: Valentina Marutyan

Abstract:

Healthcare data mining is a rapidly developing field at the intersection of technology and medicine that has the potential to change our understanding and approach to providing healthcare. Healthcare and data mining is the process of examining huge amounts of data to extract useful information that can be applied in order to improve patient care, treatment effectiveness, and overall healthcare delivery. This field looks for patterns, trends, and correlations in a variety of healthcare datasets, such as electronic health records (EHRs), medical imaging, patient demographics, and treatment histories. To accomplish this, it uses advanced analytical approaches. Predictive analysis using historical patient data is a major area of interest in healthcare data mining. This enables doctors to get involved early to prevent problems or improve results for patients. It also assists in early disease detection and customized treatment planning for every person. Doctors can customize a patient's care by looking at their medical history, genetic profile, current and previous therapies. In this way, treatments can be more effective and have fewer negative consequences. Moreover, helping patients, it improves the efficiency of hospitals. It helps them determine the number of beds or doctors they require in regard to the number of patients they expect. In this project are used models like logistic regression, random forests, and neural networks for predicting diseases and analyzing medical images. Patients were helped by algorithms such as k-means, and connections between treatments and patient responses were identified by association rule mining. Time series techniques helped in resource management by predicting patient admissions. These methods improved healthcare decision-making and personalized treatment. Also, healthcare data mining must deal with difficulties such as bad data quality, privacy challenges, managing large and complicated datasets, ensuring the reliability of models, managing biases, limited data sharing, and regulatory compliance. Finally, secret code of data mining in healthcare helps medical professionals and hospitals make better decisions, treat patients more efficiently, and work more efficiently. It ultimately comes down to using data to improve treatment, make better choices, and simplify hospital operations for all patients.

Keywords: data mining, healthcare, big data, large amounts of data

Procedia PDF Downloads 32
67 The Evaporation Study of 1-ethyl-3-methylimidazolium chloride

Authors: Kirill D. Semavin, Norbert S. Chilingarov, Eugene.V. Skokan

Abstract:

The ionic liquids (ILs) based on imidazolium cation are well known nowadays. The changing anions and substituents in imidazolium ring may lead to different physical and chemical properties of ILs. It is important that such ILs with halogen as anion are characterized by a low thermal stability. The data about thermal stability of 1-ethyl-3-methylimidazolium chloride are ambiguous. In the works of last years, thermal stability of this IL was investigated by thermogravimetric analysis and obtained results are contradictory. Moreover, in the last study, it was shown that the observed temperature of the beginning of decomposition significantly depends on the experimental conditions, for example, the heating rate of the sample. The vapor pressure of this IL is not presented at the literature. In this study, the vapor pressure of 1-ethyl-3-methylimidazolium chloride was obtained by Knudsen effusion mass-spectrometry (KEMS). The samples of [ЕMIm]Cl (purity > 98%) were supplied by Sigma–Aldrich and were additionally dried at dynamic vacuum (T = 60 0C). Preliminary procedures with Il were derived into glove box. The evaporation studies of [ЕMIm]Cl were carried out by KEMS with using original research equipment based on commercial MI1201 magnetic mass spectrometer. The stainless steel effusion cell had an effective evaporation/effusion area ratio of more than 6000. The cell temperature, measured by a Pt/Pt−Rh (10%) thermocouple, was controlled by a Termodat 128K5 device with an accuracy of ±1 K. In first step of this study, the optimal temperature of experiment and heating rate of samples were customized: 449 K and 5 K/min, respectively. In these conditions the sample is decomposed, but the experimental measurements of the vapor pressures are possible. The thermodynamic activity of [ЕMIm]Cl is close to 1 and products of decomposition don’t affect it at firstly 50 hours of experiment. Therefore, it lets to determine the saturated vapor pressure of IL. The electronic ionization mass-spectra shows that the decomposition of [ЕMIm]Cl proceeds with two ways. Nonetheless, the MALDI mass spectra of the starting sample and residue in the cell were similar. It means that the main decomposition products are gaseous under experimental conditions. This result allows us to obtain information about the kinetics of [ЕMIm]Cl decomposition. Thus, the original KEMS-based procedure made it possible to determine the IL vapor pressure under decomposition conditions. Also, the loss of sample mass due to the evaporation was obtained.

Keywords: ionic liquids, Knudsen effusion mass spectrometry, thermal stability, vapor pressure

Procedia PDF Downloads 159
66 Optimizing Residential Housing Renovation Strategies at Territorial Scale: A Data Driven Approach and Insights from the French Context

Authors: Rit M., Girard R., Villot J., Thorel M.

Abstract:

In a scenario of extensive residential housing renovation, stakeholders need models that support decision-making through a deep understanding of the existing building stock and accurate energy demand simulations. To address this need, we have modified an optimization model using open data that enables the study of renovation strategies at both territorial and national scales. This approach provides (1) a definition of a strategy to simplify decision trees from theoretical combinations, (2) input to decision makers on real-world renovation constraints, (3) more reliable identification of energy-saving measures (changes in technology or behaviour), and (4) discrepancies between currently planned and actually achieved strategies. The main contribution of the studies described in this document is the geographic scale: all residential buildings in the areas of interest were modeled and simulated using national data (geometries and attributes). These buildings were then renovated, when necessary, in accordance with the environmental objectives, taking into account the constraints applicable to each territory (number of renovations per year) or at the national level (renovation of thermal deficiencies (Energy Performance Certificates F&G)). This differs from traditional approaches that focus only on a few buildings or archetypes. This model can also be used to analyze the evolution of a building stock as a whole, as it can take into account both the construction of new buildings and their demolition or sale. Using specific case studies of French territories, this paper highlights a significant discrepancy between the strategies currently advocated by decision-makers and those proposed by our optimization model. This discrepancy is particularly evident in critical metrics such as the relationship between the number of renovations per year and achievable climate targets or the financial support currently available to households and the remaining costs. In addition, users are free to seek optimizations for their building stock across a range of different metrics (e.g., financial, energy, environmental, or life cycle analysis). These results are a clear call to re-evaluate existing renovation strategies and take a more nuanced and customized approach. As the climate crisis moves inexorably forward, harnessing the potential of advanced technologies and data-driven methodologies is imperative.

Keywords: residential housing renovation, MILP, energy demand simulations, data-driven methodology

Procedia PDF Downloads 43
65 Variables, Annotation, and Metadata Schemas for Early Modern Greek

Authors: Eleni Karantzola, Athanasios Karasimos, Vasiliki Makri, Ioanna Skouvara

Abstract:

Historical linguistics unveils the historical depth of languages and traces variation and change by analyzing linguistic variables over time. This field of linguistics usually deals with a closed data set that can only be expanded by the (re)discovery of previously unknown manuscripts or editions. In some cases, it is possible to use (almost) the entire closed corpus of a language for research, as is the case with the Thesaurus Linguae Graecae digital library for Ancient Greek, which contains most of the extant ancient Greek literature. However, concerning ‘dynamic’ periods when the production and circulation of texts in printed as well as manuscript form have not been fully mapped, representative samples and corpora of texts are needed. Such material and tools are utterly lacking for Early Modern Greek (16th-18th c.). In this study, the principles of the creation of EMoGReC, a pilot representative corpus of Early Modern Greek (16th-18th c.) are presented. Its design follows the fundamental principles of historical corpora. The selection of texts aims to create a representative and balanced corpus that gives insight into diachronic, diatopic and diaphasic variation. The pilot sample includes data derived from fully machine-readable vernacular texts, which belong to 4-5 different textual genres and come from different geographical areas. We develop a hierarchical linguistic annotation scheme, further customized to fit the characteristics of our text corpus. Regarding variables and their variants, we use as a point of departure the bundle of twenty-four features (or categories of features) for prose demotic texts of the 16th c. Tags are introduced bearing the variants [+old/archaic] or [+novel/vernacular]. On the other hand, further phenomena that are underway (cf. The Cambridge Grammar of Medieval and Early Modern Greek) are selected for tagging. The annotated texts are enriched with metalinguistic and sociolinguistic metadata to provide a testbed for the development of the first comprehensive set of tools for the Greek language of that period. Based on a relational management system with interconnection of data, annotations, and their metadata, the EMoGReC database aspires to join a state-of-the-art technological ecosystem for the research of observed language variation and change using advanced computational approaches.

Keywords: early modern Greek, variation and change, representative corpus, diachronic variables.

Procedia PDF Downloads 31
64 The Study of Thai Millennial Attitude toward End-of-Life Planning, Opportunity of Service Design Development

Authors: Mawong R., Bussracumpakorn C.

Abstract:

Millions of young people around the world have been affected by COVID-19 to their psychological and social effects. Millennials’ stresses have been shaped by a few global issues, including climate change, political instability, and financial crisis. In particular, the spread of COVID-19 has become laying psychological and socioeconomic scars on them. When end-of-life planning turns into more widely discussed, the stigma and taboos around this issue are greatly lessened. End-of-life planning is defined as a future life plan, such as financial, legacy, funeral, and memorial planning. This plan would help millennials to discover the value and meaning of life. This study explores the attitudes of Thai Millennials toward end-of-life planning as a new normal awareness of life in order to initiate an innovative service concept to fit with their value and meaning. The study conducts an in-depth interview with 12 potential participants who have awareness or action on the plan. The framework of the customer journey map is used to analyze the responses to examine trigger points, barriers, beliefs, and expectations. The findings pointed to a service concept that is suggested for a new end-of-life planning service that is suited to Thai Millennials in 4 different groups, which are 1. Social -Conscious as a socially aware who to donate time and riches to make the world and society a better place, their end-of-life planning value is inspired by the social impact of giving something or some action that they will be able to do after life or during life which provides a variety of choice based on their preference to give to society, 2. Life Fulfillment who make a life goal for themselves and attempt to achieve it before the time comes to their value will be to inspire life value with a customized plan and provide guidance to suggest, 3. Prevention of the After-Death Effect who want to plan to avoid the effects of their death as patriarch, head of the family, and anchor of someone, so they want to have a plan that brings confidence and feel relief while they are still alive and they want to find some reliable service that they can leave the death will or asset, and 4. No Guilty Planning who plan for when they wish to be worry-free as a self-responsible they want to have the plan which is easy to understand and easy to access. The overall finding of the study is to understand the new service concept of end-of-life planning which to improve knowledge of significant life worth rather than death planning, encouraging people to reassess their lives in a positive way, leading to higher self-esteem and intrinsic motivation for this generation in this time of global crisis.

Keywords: design management, end-of-life planning, millennial generation, service design solution

Procedia PDF Downloads 163
63 Exploring the Correlation between Body Constitution of an Individual as Per Ayurveda and Gut Microbiome in Healthy, Multi Ethnic Urban Population in Bangalore, India

Authors: Shalini TV, Gangadharan GG, Sriranjini S Jaideep, ASN Seshasayee, Awadhesh Pandit

Abstract:

Introduction: Prakriti (body-mind constitution of an individual) is a conventional, customized and unique understanding of which is essential for the personalized medicine described in Ayurveda, Indian System of Medicine. Based on the Doshas( functional, bio humoral unit in the body), individuals are categorized into three major Prakriti- Vata, Pitta, and Kapha. The human gut microbiome hosts plenty of highly diverse and metabolically active microorganisms, mainly dominated by the bacteria, which are known to influence the physiology of an individual. Few researches have shown the correlation between the Prakriti and the biochemical parameters. In this study, an attempt was made to explore any correlation between the Prakriti (phenotype of an individual) with the Genetic makeup of the gut microbiome in healthy individuals. Materials and methods: 270 multi-ethnic, healthy volunteers of both sex with the age group between 18 to 40 years, with no history of antibiotics in the last 6 months were recruited into three groups of Vata, Pitta, and Kapha. The Prakriti of the individual was determined using Ayusoft, a software designed by CDAC, Pune, India. The volunteers were subjected to initial screening for the assessment of their height, weight, Body Mass Index, Vital signs and Blood investigations to ensure they are healthy. The stool and saliva samples of the recruited volunteers were collected as per the standard operating procedure developed, and the bacterial DNA was isolated using Qiagen kits. The extracted DNA was subjected to 16s rRNA sequencing using the Illumina kits. The sequencing libraries are targeting the variable V3 and V4 regions of the 16s rRNA gene. Paired sequencing was done on the MiSeq system and data were analyzed using the CLC Genomics workbench 11. Results: The 16s rRNA sequencing of the V3 and V4 regions showed a diverse pattern in both the oral and stool microbial DNA. The study did not reveal any specific pattern of bacterial flora amongst the Prakriti. All the p-values were more than the effective alpha values for all OTUs in both the buccal cavity and stool samples. Therefore, there was no observed significant enrichment of an OTU in the patient samples from either the buccal cavity or stool samples. Conclusion: In healthy volunteers of multi-ethnicity, due to the influence of the various factors, the correlation between the Prakriti and the gut microbiome was not seen.

Keywords: gut microbiome, ayurveda Prakriti, sequencing, multi-ethnic urban population

Procedia PDF Downloads 106
62 TARF: Web Toolkit for Annotating RNA-Related Genomic Features

Authors: Jialin Ma, Jia Meng

Abstract:

Genomic features, the genome-based coordinates, are commonly used for the representation of biological features such as genes, RNA transcripts and transcription factor binding sites. For the analysis of RNA-related genomic features, such as RNA modification sites, a common task is to correlate these features with transcript components (5'UTR, CDS, 3'UTR) to explore their distribution characteristics in terms of transcriptomic coordinates, e.g., to examine whether a specific type of biological feature is enriched near transcription start sites. Existing approaches for performing these tasks involve the manipulation of a gene database, conversion from genome-based coordinate to transcript-based coordinate, and visualization methods that are capable of showing RNA transcript components and distribution of the features. These steps are complicated and time consuming, and this is especially true for researchers who are not familiar with relevant tools. To overcome this obstacle, we develop a dedicated web app TARF, which represents web toolkit for annotating RNA-related genomic features. TARF web tool intends to provide a web-based way to easily annotate and visualize RNA-related genomic features. Once a user has uploaded the features with BED format and specified a built-in transcript database or uploaded a customized gene database with GTF format, the tool could fulfill its three main functions. First, it adds annotation on gene and RNA transcript components. For every features provided by the user, the overlapping with RNA transcript components are identified, and the information is combined in one table which is available for copy and download. Summary statistics about ambiguous belongings are also carried out. Second, the tool provides a convenient visualization method of the features on single gene/transcript level. For the selected gene, the tool shows the features with gene model on genome-based view, and also maps the features to transcript-based coordinate and show the distribution against one single spliced RNA transcript. Third, a global transcriptomic view of the genomic features is generated utilizing the Guitar R/Bioconductor package. The distribution of features on RNA transcripts are normalized with respect to RNA transcript landmarks and the enrichment of the features on different RNA transcript components is demonstrated. We tested the newly developed TARF toolkit with 3 different types of genomics features related to chromatin H3K4me3, RNA N6-methyladenosine (m6A) and RNA 5-methylcytosine (m5C), which are obtained from ChIP-Seq, MeRIP-Seq and RNA BS-Seq data, respectively. TARF successfully revealed their respective distribution characteristics, i.e. H3K4me3, m6A and m5C are enriched near transcription starting sites, stop codons and 5’UTRs, respectively. Overall, TARF is a useful web toolkit for annotation and visualization of RNA-related genomic features, and should help simplify the analysis of various RNA-related genomic features, especially those related RNA modifications.

Keywords: RNA-related genomic features, annotation, visualization, web server

Procedia PDF Downloads 181
61 Coupling of Microfluidic Droplet Systems with ESI-MS Detection for Reaction Optimization

Authors: Julia R. Beulig, Stefan Ohla, Detlev Belder

Abstract:

In contrast to off-line analytical methods, lab-on-a-chip technology delivers direct information about the observed reaction. Therefore, microfluidic devices make an important scientific contribution, e.g. in the field of synthetic chemistry. Herein, the rapid generation of analytical data can be applied for the optimization of chemical reactions. These microfluidic devices enable a fast change of reaction conditions as well as a resource saving method of operation. In the presented work, we focus on the investigation of multiphase regimes, more specifically on a biphasic microfluidic droplet systems. Here, every single droplet is a reaction container with customized conditions. The biggest challenge is the rapid qualitative and quantitative readout of information as most detection techniques for droplet systems are non-specific, time-consuming or too slow. An exception is the electrospray mass spectrometry (ESI-MS). The combination of a reaction screening platform with a rapid and specific detection method is an important step in droplet-based microfluidics. In this work, we present a novel approach for synthesis optimization on the nanoliter scale with direct ESI-MS detection. The development of a droplet-based microfluidic device, which enables the modification of different parameters while simultaneously monitoring the effect on the reaction within a single run, is shown. By common soft- and photolithographic techniques a polydimethylsiloxane (PDMS) microfluidic chip with different functionalities is developed. As an interface for the MS detection, we use a steel capillary for ESI and improve the spray stability with a Teflon siphon tubing, which is inserted underneath the steel capillary. By optimizing the flow rates, it is possible to screen parameters of various reactions, this is exemplarity shown by a Domino Knoevenagel Hetero-Diels-Alder reaction. Different starting materials, catalyst concentrations and solvent compositions are investigated. Due to the high repetition rate of the droplet production, each set of reaction condition is examined hundreds of times. As a result, of the investigation, we receive possible reagents, the ideal water-methanol ratio of the solvent and the most effective catalyst concentration. The developed system can help to determine important information about the optimal parameters of a reaction within a short time. With this novel tool, we make an important step on the field of combining droplet-based microfluidics with organic reaction screening.

Keywords: droplet, mass spectrometry, microfluidics, organic reaction, screening

Procedia PDF Downloads 269
60 Additive Manufacturing of Microstructured Optical Waveguides Using Two-Photon Polymerization

Authors: Leonnel Mhuka

Abstract:

Background: The field of photonics has witnessed substantial growth, with an increasing demand for miniaturized and high-performance optical components. Microstructured optical waveguides have gained significant attention due to their ability to confine and manipulate light at the subwavelength scale. Conventional fabrication methods, however, face limitations in achieving intricate and customizable waveguide structures. Two-photon polymerization (TPP) emerges as a promising additive manufacturing technique, enabling the fabrication of complex 3D microstructures with submicron resolution. Objectives: This experiment aimed to utilize two-photon polymerization to fabricate microstructured optical waveguides with precise control over geometry and dimensions. The objective was to demonstrate the feasibility of TPP as an additive manufacturing method for producing functional waveguide devices with enhanced performance. Methods: A femtosecond laser system operating at a wavelength of 800 nm was employed for two-photon polymerization. A custom-designed CAD model of the microstructured waveguide was converted into G-code, which guided the laser focus through a photosensitive polymer material. The waveguide structures were fabricated using a layer-by-layer approach, with each layer formed by localized polymerization induced by non-linear absorption of the laser light. Characterization of the fabricated waveguides included optical microscopy, scanning electron microscopy, and optical transmission measurements. The optical properties, such as mode confinement and propagation losses, were evaluated to assess the performance of the additive manufactured waveguides. Conclusion: The experiment successfully demonstrated the additive manufacturing of microstructured optical waveguides using two-photon polymerization. Optical microscopy and scanning electron microscopy revealed the intricate 3D structures with submicron resolution. The measured optical transmission indicated efficient light propagation through the fabricated waveguides. The waveguides exhibited well-defined mode confinement and relatively low propagation losses, showcasing the potential of TPP-based additive manufacturing for photonics applications. The experiment highlighted the advantages of TPP in achieving high-resolution, customized, and functional microstructured optical waveguides. Conclusion: his experiment substantiates the viability of two-photon polymerization as an innovative additive manufacturing technique for producing complex microstructured optical waveguides. The successful fabrication and characterization of these waveguides open doors to further advancements in the field of photonics, enabling the development of high-performance integrated optical devices for various applications

Keywords: Additive Manufacturing, Microstructured Optical Waveguides, Two-Photon Polymerization, Photonics Applications

Procedia PDF Downloads 67