Search results for: patent analytics
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 532

Search results for: patent analytics

172 From Linear to Circular Model: An Artificial Intelligence-Powered Approach in Fosso Imperatore

Authors: Carlotta D’Alessandro, Giuseppe Ioppolo, Katarzyna Szopik-Depczyńska

Abstract:

— The growing scarcity of resources and the mounting pressures of climate change, water pollution, and chemical contamination have prompted societies, governments, and businesses to seek ways to minimize their environmental impact. To combat climate change, and foster sustainability, Industrial Symbiosis (IS) offers a powerful approach, facilitating the shift toward a circular economic model. IS has gained prominence in the European Union's policy framework as crucial enabler of resource efficiency and circular economy practices. The essence of IS lies in the collaborative sharing of resources such as energy, material by-products, waste, and water, thanks to geographic proximity. It can be exemplified by eco-industrial parks (EIPs), which are natural environments for boosting cooperation and resource sharing between businesses. EIPs are characterized by group of businesses situated in proximity, connected by a network of both cooperative and competitive interactions. They represent a sustainable industrial model aimed at reducing resource use, waste, and environmental impact while fostering economic and social wellbeing. IS, combined with Artificial Intelligence (AI)-driven technologies, can further optimize resource sharing and efficiency within EIPs. This research, supported by the “CE_IPs” project, aims to analyze the potential for IS and AI, in advancing circularity and sustainability at Fosso Imperatore. The Fosso Imperatore Industrial Park in Nocera Inferiore, Italy, specializes in agriculture and the industrial transformation of agricultural products, particularly tomatoes, tobacco, and textile fibers. This unique industrial cluster, centered around tomato cultivation and processing, also includes mechanical engineering enterprises and agricultural packaging firms. To stimulate the shift from a traditional to a circular economic model, an AI-powered Local Development Plan (LDP) is developed for Fosso Imperatore. It can leverage data analytics, predictive modeling, and stakeholder engagement to optimize resource utilization, reduce waste, and promote sustainable industrial practices. A comprehensive SWOT analysis of the AI-powered LDP revealed several key factors influencing its potential success and challenges. Among the notable strengths and opportunities arising from AI implementation are reduced processing times, fewer human errors, and increased revenue generation. Furthermore, predictive analytics minimize downtime, bolster productivity, and elevate quality while mitigating workplace hazards. However, the integration of AI also presents potential weaknesses and threats, including significant financial investment, since implementing and maintaining AI systems can be costly. The widespread adoption of AI could lead to job losses in certain sectors. Lastly, AI systems are susceptible to cyberattacks, posing risks to data security and operational continuity. Moreover, an Analytic Hierarchy Process (AHP) analysis was employed to yield a prioritized ranking of the outlined AI-driven LDP practices based on the stakeholder input, ensuring a more comprehensive and representative understanding of their relative significance for achieving sustainability in Fosso Imperatore Industrial Park. While this study provides valuable insights into the potential of AIpowered LDP at the Fosso Imperatore, it is important to note that the findings may not be directly applicable to all industrial parks, particularly those with different sizes, geographic locations, or industry compositions. Additional study is necessary to scrutinize the generalizability of these results and to identify best practices for implementing AI-driven LDP in diverse contexts.

Keywords: artificial intelligence, climate change, Fosso Imperatore, industrial park, industrial symbiosis

Procedia PDF Downloads 19
171 Duality of Leagility and Governance: A New Normal Demand Network Management Paradigm under Pandemic

Authors: Jacky Hau

Abstract:

The prevalence of emerging technologies disrupts various industries as well as consumer behavior. Data collection has been in the fingertip and inherited through enabled Internet-of-things (IOT) devices. Big data analytics (BDA) becomes possible and allows real-time demand network management (DNM) through leagile supply chain. To enhance further on its resilience and predictability, governance is going to be examined to promote supply chain transparency and trust in an efficient manner. Leagility combines lean thinking and agile techniques in supply chain management. It aims at reducing costs and waste, as well as maintaining responsiveness to any volatile consumer demand by means of adjusting the decoupling point where the product flow changes from push to pull. Leagility would only be successful when collaborative planning, forecasting, and replenishment (CPFR) process or alike is in place throughout the supply chain business entities. Governance and procurement of the supply chain, however, is crucial and challenging for the execution of CPFR as every entity has to walk-the-talk generously for the sake of overall benefits of supply chain performance, not to mention the complexity of exercising the polices at both of within across various supply chain business entities on account of organizational behavior and mutual trust. Empirical survey results showed that the effective timespan on demand forecasting had been drastically shortening in the magnitude of months to weeks planning horizon, thus agility shall come first and preferably following by lean approach in a timely manner.

Keywords: governance, leagility, procure-to-pay, source-to-contract

Procedia PDF Downloads 108
170 Comprehensive Study of Data Science

Authors: Asifa Amara, Prachi Singh, Kanishka, Debargho Pathak, Akshat Kumar, Jayakumar Eravelly

Abstract:

Today's generation is totally dependent on technology that uses data as its fuel. The present study is all about innovations and developments in data science and gives an idea about how efficiently to use the data provided. This study will help to understand the core concepts of data science. The concept of artificial intelligence was introduced by Alan Turing in which the main principle was to create an artificial system that can run independently of human-given programs and can function with the help of analyzing data to understand the requirements of the users. Data science comprises business understanding, analyzing data, ethical concerns, understanding programming languages, various fields and sources of data, skills, etc. The usage of data science has evolved over the years. In this review article, we have covered a part of data science, i.e., machine learning. Machine learning uses data science for its work. Machines learn through their experience, which helps them to do any work more efficiently. This article includes a comparative study image between human understanding and machine understanding, advantages, applications, and real-time examples of machine learning. Data science is an important game changer in the life of human beings. Since the advent of data science, we have found its benefits and how it leads to a better understanding of people, and how it cherishes individual needs. It has improved business strategies, services provided by them, forecasting, the ability to attend sustainable developments, etc. This study also focuses on a better understanding of data science which will help us to create a better world.

Keywords: data science, machine learning, data analytics, artificial intelligence

Procedia PDF Downloads 78
169 From Ride-Hailing App to Diversified and Sustainable Platform Business Model

Authors: Ridwan Dewayanto Rusli

Abstract:

We show how prisoner's dilemma-type competition problems can be mitigated through rapid platform diversification and ecosystem expansion. We analyze a ride-hailing company in Southeast Asia, Gojek, whose network grew to more than 170 million users comprising consumers, partner drivers, merchants, and complementors within a few years and has already achieved higher contribution margins than ride-hailing peers Uber and Lyft. Its ecosystem integrates ride-hailing, food delivery and logistics, merchant solutions, e-commerce, marketplace and advertising, payments, and fintech offerings. The company continues growing its network of complementors and App developers, expanding content and gaining critical mass in consumer data analytics and advertising. We compare the company's growth and diversification trajectory with those of its main international rivals and peers. The company's rapid growth and future potential are analyzed using Cusumano's (2012) Staying Power and Six Principles, Hax and Wilde's (2003) and Hax's (2010) The Delta Model as well as Santos' (2016) home-market advantages frameworks. The recently announced multi-billion-dollar merger with one of Southeast Asia's largest e-commerce majors lends additional support to the above arguments.

Keywords: ride-hailing, prisoner's dilemma, platform and ecosystem strategy, digital applications, diversification, home market advantages, e-commerce

Procedia PDF Downloads 91
168 Mining User-Generated Contents to Detect Service Failures with Topic Model

Authors: Kyung Bae Park, Sung Ho Ha

Abstract:

Online user-generated contents (UGC) significantly change the way customers behave (e.g., shop, travel), and a pressing need to handle the overwhelmingly plethora amount of various UGC is one of the paramount issues for management. However, a current approach (e.g., sentiment analysis) is often ineffective for leveraging textual information to detect the problems or issues that a certain management suffers from. In this paper, we employ text mining of Latent Dirichlet Allocation (LDA) on a popular online review site dedicated to complaint from users. We find that the employed LDA efficiently detects customer complaints, and a further inspection with the visualization technique is effective to categorize the problems or issues. As such, management can identify the issues at stake and prioritize them accordingly in a timely manner given the limited amount of resources. The findings provide managerial insights into how analytics on social media can help maintain and improve their reputation management. Our interdisciplinary approach also highlights several insights by applying machine learning techniques in marketing research domain. On a broader technical note, this paper illustrates the details of how to implement LDA in R program from a beginning (data collection in R) to an end (LDA analysis in R) since the instruction is still largely undocumented. In this regard, it will help lower the boundary for interdisciplinary researcher to conduct related research.

Keywords: latent dirichlet allocation, R program, text mining, topic model, user generated contents, visualization

Procedia PDF Downloads 185
167 Programmable Microfluidic Device Based on Stimuli Responsive Hydrogels

Authors: Martin Elstner

Abstract:

Processing of information by means of handling chemicals is a ubiquitous phenomenon in nature. Technical implementations of chemical information processing lack of low integration densities compared to electronic devices. Stimuli responsive hydrogels are promising candidates for materials with information processing capabilities. These hydrogels are sensitive toward chemical stimuli like metal ions or amino acids. The binding of an analyte molecule induces conformational changes inside the polymer network and subsequently the water content and volume of the hydrogel varies. This volume change can control material flows, and concurrently information flows, in microfluidic devices. The combination of this technology with powerful chemical logic gates yields in a platform for highly integrated chemical circuits. The manufacturing process of such devices is very challenging and rapid prototyping is a key technology used in the study. 3D printing allows generating three-dimensional defined structures of high complexity in a single and fast process step. This thermoplastic master is molded into PDMS and the master is removed by dissolution in an organic solvent. A variety of hydrogel materials is prepared by dispenser printing of pre-polymer solutions. By a variation of functional groups or cross-linking units, the functionality of the hole circuit can be programmed. Finally, applications in the field of bio-molecular analytics were demonstrated with an autonomously operating microfluidic chip.

Keywords: bioanalytics, hydrogels, information processing, microvalve

Procedia PDF Downloads 304
166 An Evaluation of Existing Models to Smart Cities Development Around the World

Authors: Aqsa Mehmood, Muhammad Ali Tahir, Hafiz Syed Hamid Arshad, Salman Atif, Ejaz Hussain, Gavin McArdle, Michela Bertolotto

Abstract:

The evolution of smart cities in recent years has been developing dramatically. As urbanization increases, the demand for big data analytics and digital technology-based solutions for cities has also increased. Many cities around the world have now planned to focus on smart cities. To obtain a systematic overview of smart city models, we carried out a bibliometric analysis in the context of seven regions of the world to understand the main dimensions that characterize smart cities. This paper analyses articles published between 2017 and 2021 that were captured from Web of Science and Scopus. Specifically, we investigated publication trends to highlight the research gaps and current developments in smart cities research. Our survey provides helpful insights into the geographical distribution of smart city publications with respect to regions of the world and explores the current key topics relevant to smart cities and the co-occurrences of keywords used in these publications. A systematic literature review and keyword analysis were performed. The results have focused on identifying future directions in smart city development, including smart citizens, ISO standards, Open Geospatial Consortium and the sustainability factor of smart cities. This article will assist researchers and urban planners in understanding the latest trends in research and highlight the aspects which need further attention.

Keywords: smart cities, sustainability, regions, urban development, VOS viewer, research trends

Procedia PDF Downloads 108
165 A Data Driven Methodological Approach to Economic Pre-Evaluation of Reuse Projects of Ancient Urban Centers

Authors: Pietro D'Ambrosio, Roberta D'Ambrosio

Abstract:

The upgrading of the architectural and urban heritage of the urban historic centers almost always involves the planning for the reuse and refunctionalization of the structures. Such interventions have complexities linked to the need to take into account the urban and social context in which the structure and its intrinsic characteristics such as historical and artistic value are inserted. To these, of course, we have to add the need to make a preliminary estimate of recovery costs and more generally to assess the economic and financial sustainability of the whole project of re-socialization. Particular difficulties are encountered during the pre-assessment of costs since it is often impossible to perform analytical surveys and structural tests for both structural conditions and obvious cost and time constraints. The methodology proposed in this work, based on a multidisciplinary and data-driven approach, is aimed at obtaining, at very low cost, reasonably priced economic evaluations of the interventions to be carried out. In addition, the specific features of the approach used, derived from the predictive analysis techniques typically applied in complex IT domains (big data analytics), allow to obtain as a result indirectly the evaluation process of a shared database that can be used on a generalized basis to estimate such other projects. This makes the methodology particularly indicated in those cases where it is expected to intervene massively across entire areas of historical city centers. The methodology has been partially tested during a study aimed at assessing the feasibility of a project for the reuse of the monumental complex of San Massimo, located in the historic center of Salerno, and is being further investigated.

Keywords: evaluation, methodology, restoration, reuse

Procedia PDF Downloads 181
164 A Corpus-Based Approach to Understanding Market Access in Fisheries and Aquaculture: A Systematic Literature Review

Authors: Cheryl Marie Cordeiro

Abstract:

Although fisheries and aquaculture studies might seem marginal to international business (IB) studies in general, fisheries and aquaculture IB (FAIB) management is currently facing increasing pressure to meet global demand and consumption for fish in the next coming decades. In part address to this challenge, the purpose of this systematic review of literature (SLR) study is to investigate the use of the term ‘market access’ in its context of use in the generic literature and business sector discourse, in comparison to the more specific literature and discourse in fisheries, aquaculture and seafood. This SLR aims to uncover the knowledge/interest gaps between the academic subject discourses and business sector practices. Corpus driven in methodology and using a triangulation method of three different text analysis software including AntConc, VOSviewer and Web of Science (WoS) analytics, the SLR results indicate a gap in conceptual knowledge and business practices in how ‘market access’ is conceived and used in the context of the pharmaceutical healthcare industry and FAIB research and practice. While it is acknowledged that the product orientation of different business sectors might differ, this SLR study works with the assumption that both business sectors are global in orientation. These business sectors are complex in their operations from product to market. This SLR suggests a conceptual model in understanding the challenges, the potential barriers as well as avenues for solutions to developing market access for FAIB.

Keywords: market access, fisheries and aquaculture, international business, systematic literature review

Procedia PDF Downloads 142
163 Forecasting Future Society to Explore Promising Security Technologies

Authors: Jeonghwan Jeon, Mintak Han, Youngjun Kim

Abstract:

Due to the rapid development of information and communication technology (ICT), a substantial transformation is currently happening in the society. As the range of intelligent technologies and services is continuously expanding, ‘things’ are becoming capable of communicating one another and even with people. However, such “Internet of Things” has the technical weakness so that a great amount of such information transferred in real-time may be widely exposed to the threat of security. User’s personal data are a typical example which is faced with a serious security threat. The threats of security will be diversified and arose more frequently because next generation of unfamiliar technology develops. Moreover, as the society is becoming increasingly complex, security vulnerability will be increased as well. In the existing literature, a considerable number of private and public reports that forecast future society have been published as a precedent step of the selection of future technology and the establishment of strategies for competitiveness. Although there are previous studies that forecast security technology, they have focused only on technical issues and overlooked the interrelationships between security technology and social factors are. Therefore, investigations of security threats in the future and security technology that is able to protect people from various threats are required. In response, this study aims to derive potential security threats associated with the development of technology and to explore the security technology that can protect against them. To do this, first of all, private and public reports that forecast future and online documents from technology-related communities are collected. By analyzing the data, future issues are extracted and categorized in terms of STEEP (Society, Technology, Economy, Environment, and Politics), as well as security. Second, the components of potential security threats are developed based on classified future issues. Then, points that the security threats may occur –for example, mobile payment system based on a finger scan technology– are identified. Lastly, alternatives that prevent potential security threats are proposed by matching security threats with points and investigating related security technologies from patent data. Proposed approach can identify the ICT-related latent security menaces and provide the guidelines in the ‘problem – alternative’ form by linking the threat point with security technologies.

Keywords: future society, information and communication technology, security technology, technology forecasting

Procedia PDF Downloads 462
162 Laparoscopic Uterovaginal Anastomosis in Cervicovaginal Agenesis

Authors: Anamika Choudhary, Neha Qurrat Ain

Abstract:

Background: Congenital agenesis of uterine cervix is a rare anomaly often associated with partial or complete agenesis of vagina. Here is a case report of a 14 year old girl who presented with primary amenorrhea and cyclical abdominal pain since last one year with suprapubic mass palpable. On examination complete absence of a vagina was found, and ultrasound along with magnetic resonance imaging (MRI) suggested cervicovaginal agenesis associated with cryptomenorrhea, which resulted in hematometra and b/l hematosalpinx with pelvic endometriosis. After proper counseling regarding anastomosis failure and the need for future laprotomy or hysterectomy, the patient planned for laparoscopic uterovaginal anastomosis with modified McIndoe vaginoplasty with split skin graft. Case Summary: Chief complaint: The 14 year old girl presented with primary amenorrhea and cyclical abdominal pain. Diagnosis:On history, examination and investigations we made differential diagnoses of cervicovaginal agenesis, cervicovaginal atresia. Post operatively, we concluded it’s a cervicovaginal agenesis. Intervention: Laparoscopic uterovaginal anastomosis was done, and neovagina was created using split skin graft from the thigh and silicone stent. The graft was kept patent, and restenosis was prevented using a dental mould as vaginal dilator. Outcome: Postoperatively 1 year follow-up has been done. We have observed successful uterovaginal anastomosis and good uptake of graft. We also observed the resumption of normal menstrual bleeding. Currently, there has been no restenosis, abnormal vaginal discharge and decreased dysmenorrhea. Conclusion: Laparoscopic-assisted uterovaginal anastomosis can be the treatment of choice in patients with cervical agenesis and atresia instead of hysterectomy, thereby preserving the reproductive function. This conservative approach has better outcomes, as stated in the procedure below. The procedure is successful insofar as the resumption of menstrual function. However, long-term reproductive outcomes, progression of endometriosis, functioning of fallopian tubes, and sexual life in these girls will require further follow-up.

Keywords: cervicovaginal agenesis, uterovaginal anastomosis, dental mould, silicon stent

Procedia PDF Downloads 15
161 A Machine Learning Approach for Performance Prediction Based on User Behavioral Factors in E-Learning Environments

Authors: Naduni Ranasinghe

Abstract:

E-learning environments are getting more popular than any other due to the impact of COVID19. Even though e-learning is one of the best solutions for the teaching-learning process in the academic process, it’s not without major challenges. Nowadays, machine learning approaches are utilized in the analysis of how behavioral factors lead to better adoption and how they related to better performance of the students in eLearning environments. During the pandemic, we realized the academic process in the eLearning approach had a major issue, especially for the performance of the students. Therefore, an approach that investigates student behaviors in eLearning environments using a data-intensive machine learning approach is appreciated. A hybrid approach was used to understand how each previously told variables are related to the other. A more quantitative approach was used referred to literature to understand the weights of each factor for adoption and in terms of performance. The data set was collected from previously done research to help the training and testing process in ML. Special attention was made to incorporating different dimensionality of the data to understand the dependency levels of each. Five independent variables out of twelve variables were chosen based on their impact on the dependent variable, and by considering the descriptive statistics, out of three models developed (Random Forest classifier, SVM, and Decision tree classifier), random forest Classifier (Accuracy – 0.8542) gave the highest value for accuracy. Overall, this work met its goals of improving student performance by identifying students who are at-risk and dropout, emphasizing the necessity of using both static and dynamic data.

Keywords: academic performance prediction, e learning, learning analytics, machine learning, predictive model

Procedia PDF Downloads 150
160 Big Data in Construction Project Management: The Colombian Northeast Case

Authors: Sergio Zabala-Vargas, Miguel Jiménez-Barrera, Luz VArgas-Sánchez

Abstract:

In recent years, information related to project management in organizations has been increasing exponentially. Performance data, management statistics, indicator results have forced the collection, analysis, traceability, and dissemination of project managers to be essential. In this sense, there are current trends to facilitate efficient decision-making in emerging technology projects, such as: Machine Learning, Data Analytics, Data Mining, and Big Data. The latter is the most interesting in this project. This research is part of the thematic line Construction methods and project management. Many authors present the relevance that the use of emerging technologies, such as Big Data, has taken in recent years in project management in the construction sector. The main focus is the optimization of time, scope, budget, and in general mitigating risks. This research was developed in the northeastern region of Colombia-South America. The first phase was aimed at diagnosing the use of emerging technologies (Big-Data) in the construction sector. In Colombia, the construction sector represents more than 50% of the productive system, and more than 2 million people participate in this economic segment. The quantitative approach was used. A survey was applied to a sample of 91 companies in the construction sector. Preliminary results indicate that the use of Big Data and other emerging technologies is very low and also that there is interest in modernizing project management. There is evidence of a correlation between the interest in using new data management technologies and the incorporation of Building Information Modeling BIM. The next phase of the research will allow the generation of guidelines and strategies for the incorporation of technological tools in the construction sector in Colombia.

Keywords: big data, building information modeling, tecnology, project manamegent

Procedia PDF Downloads 124
159 Sexual Health And Male Fertility: Improving Sperm Health With Focus On Technology

Authors: Diana Peninger

Abstract:

Over 10% of couples in the U.S. have infertility problems, with roughly 40% traceable to the male partner. Yet, little attention has been given to improving men’s contribution to the conception process. One solution that is showing promise in increasing conception rates for IVF and other assisted reproductive technology treatments is a first-of-its-kind semen collection that has been engineered to mitigate sperm damage caused by traditional collection methods. Patients are able to collect semen at home and deliver to clinics within 48 hours for use in fertility analysis and treatment, with less stress and improved specimen viability. This abstract will share these findings along with expert insight and tips to help attendees understand the key role sperm collection plays in addressing and treating reproductive issues, while helping to improve patient outcomes and success. Our research was to determine if male reproductive outcomes can be increased by improving sperm specimen health with a focus on technology. We utilized a redesigned semen collection cup (patented as the Device for Improved Semen Collection/DISC—U.S. Patent 6864046 – known commercially as a ProteX) that met a series of physiological parameters. Previous research demonstrated significant improvement in semen perimeters (motility forward, progression, viability, and longevity) and overall sperm biochemistry when the DISC is used for collection. Animal studies have also shown dramatic increases in pregnancy rates. Our current study compares samples collected in the DISC, next-generation DISC (DISCng), and a standard specimen cup (SSC), dry, with the 1 mL measured amount of media and media in excess ( 5mL). Both human and animal testing will be included. With sperm counts declining at alarming rates due to environmental, lifestyle, and other health factors, accurate evaluations of sperm health are critical to understanding reproductive health, origins, and treatments of infertility. An increase in the health of the sperm as measured by extensive semen parameter analysis and improved semen parameters stable for 48 hours, expanding the processing time from 1 hour to 48 hours were also demonstrated.

Keywords: reprodutive, sperm, male, infertility

Procedia PDF Downloads 127
158 Critically Analyzing the Application of Big Data for Smart Transportation: A Case Study of Mumbai

Authors: Tanuj Joshi

Abstract:

Smart transportation is fast emerging as a solution to modern cities’ approach mobility issues, delayed emergency response rate and high congestion on streets. Present day scenario with Google Maps, Waze, Yelp etc. demonstrates how information and communications technologies controls the intelligent transportation system. This intangible and invisible infrastructure is largely guided by the big data analytics. On the other side, the exponential increase in Indian urban population has intensified the demand for better services and infrastructure to satisfy the transportation needs of its citizens. No doubt, India’s huge internet usage is looked as an important resource to guide to achieve this. However, with a projected number of over 40 billion objects connected to the Internet by 2025, the need for systems to handle massive volume of data (big data) also arises. This research paper attempts to identify the ways of exploiting the big data variables which will aid commuters on Indian tracks. This study explores real life inputs by conducting survey and interviews to identify which gaps need to be targeted to better satisfy the customers. Several experts at Mumbai Metropolitan Region Development Authority (MMRDA), Mumbai Metro and Brihanmumbai Electric Supply and Transport (BEST) were interviewed regarding the Information Technology (IT) systems currently in use. The interviews give relevant insights and requirements into the workings of public transportation systems whereas the survey investigates the macro situation.

Keywords: smart transportation, mobility issue, Mumbai transportation, big data, data analysis

Procedia PDF Downloads 172
157 Enhancing the Pricing Expertise of an Online Distribution Channel

Authors: Luis N. Pereira, Marco P. Carrasco

Abstract:

Dynamic pricing is a revenue management strategy in which hotel suppliers define, over time, flexible and different prices for their services for different potential customers, considering the profile of e-consumers and the demand and market supply. This means that the fundamentals of dynamic pricing are based on economic theory (price elasticity of demand) and market segmentation. This study aims to define a dynamic pricing strategy and a contextualized offer to the e-consumers profile in order to improve the number of reservations of an online distribution channel. Segmentation methods (hierarchical and non-hierarchical) were used to identify and validate an optimal number of market segments. A profile of the market segments was studied, considering the characteristics of the e-consumers and the probability of reservation a room. In addition, the price elasticity of demand was estimated for each segment using econometric models. Finally, predictive models were used to define rules for classifying new e-consumers into pre-defined segments. The empirical study illustrates how it is possible to improve the intelligence of an online distribution channel system through an optimal dynamic pricing strategy and a contextualized offer to the profile of each new e-consumer. A database of 11 million e-consumers of an online distribution channel was used in this study. The results suggest that an appropriate policy of market segmentation in using of online reservation systems is benefit for the service suppliers because it brings high probability of reservation and generates more profit than fixed pricing.

Keywords: dynamic pricing, e-consumers segmentation, online reservation systems, predictive analytics

Procedia PDF Downloads 233
156 Epidemiology of Congenital Heart Defects in Kazakhstan: Data from Unified National Electronic Healthcare System 2014-2020

Authors: Dmitriy Syssoyev, Aslan Seitkamzin, Natalya Lim, Kamilla Mussina, Abduzhappar Gaipov, Dimitri Poddighe, Dinara Galiyeva

Abstract:

Background: Data on the epidemiology of congenital heart defects (CHD) in Kazakhstan is scarce. Therefore, the aim of this study was to describe the incidence, prevalence and all-cause mortality of patients with CHD in Kazakhstan, using national large-scale registry data from the Unified National Electronic Healthcare System (UNEHS) for the period of 2014-2020. Methods: In this retrospective cohort study, the included data pertained to all patients diagnosed with CHD in Kazakhstan and registered in UNEHS between January 2014 and December 2020. CHD was defined based on International Classification of Diseases 10th Revision (ICD-10) codes Q20-Q26. Incidence, prevalence, and all-cause mortality rates were calculated per 100,000 population. Survival analysis was performed using Cox proportional hazards regression modeling and the Kaplan-Meier method. Results: In total, 66,512 patients were identified. Among them, 59,534 (89.5%) were diagnosed with a single CHD, while 6,978 (10.5%) had more than two CHDs. The median age at diagnosis was 0.08 years (interquartile range (IQR) 0.01 – 0.66) for people with multiple CHD types and 0.39 years (IQR 0.04 – 8.38) for those with a single CHD type. The most common CHD types were atrial septal defect (ASD) and ventricular septal defect (VSD), accounting for 25.8% and 21.2% of single CHD cases, respectively. The most common multiple types of CHD were ASD with VSD (23.4%), ASD with patent ductus arteriosus (PDA) (19.5%), and VSD with PDA (17.7%). The incidence rate of CHD decreased from 64.6 to 47.1 cases per 100,000 population among men and from 68.7 to 42.4 among women. The prevalence rose from 66.1 to 334.1 cases per 100,000 population among men and from 70.8 to 328.7 among women. Mortality rates showed a slight increase from 3.5 to 4.7 deaths per 100,000 in men and from 2.9 to 3.7 in women. Median follow-up was 5.21 years (IQR 2.47 – 11.69). Male sex (HR 1.60, 95% CI 1.45 - 1.77), having multiple CHDs (HR 2.45, 95% CI 2.01 - 2.97), and living in a rural area (HR 1.32, 95% CI 1.19 - 1.47) were associated with a higher risk of all-cause mortality. Conclusion: The incidence of CHD in Kazakhstan has shown a moderate decrease between 2014 and 2020, while prevalence and mortality have increased. Male sex, multiple CHD types, and rural residence were significantly associated with a higher risk of all-cause mortality.

Keywords: congenital heart defects (CHD), epidemiology, incidence, Kazakhstan, mortality, prevalence

Procedia PDF Downloads 91
155 Predicting the Next Offensive Play Types will be Implemented to Maximize the Defense’s Chances of Success in the National Football League

Authors: Chris Schoborg, Morgan C. Wang

Abstract:

In the realm of the National Football League (NFL), substantial dedication of time and effort is invested by both players and coaches in meticulously analyzing the game footage of their opponents. The primary aim is to anticipate the actions of the opposing team. Defensive players and coaches are especially focused on deciphering their adversaries' intentions to effectively counter their strategies. Acquiring insights into the specific play type and its intended direction on the field would confer a significant competitive advantage. This study establishes pre-snap information as the cornerstone for predicting both the play type (e.g., deep pass, short pass, or run) and its spatial trajectory (right, left, or center). The dataset for this research spans the regular NFL season data for all 32 teams from 2013 to 2022. This dataset is acquired using the nflreadr package, which conveniently extracts play-by-play data from NFL games and imports it into the R environment as structured datasets. In this study, we employ a recently developed machine learning algorithm, XGBoost. The final predictive model achieves an impressive lift of 2.61. This signifies that the presented model is 2.61 times more effective than random guessing—a significant improvement. Such a model has the potential to markedly enhance defensive coaches' ability to formulate game plans and adequately prepare their players, thus mitigating the opposing offense's yardage and point gains.

Keywords: lift, NFL, sports analytics, XGBoost

Procedia PDF Downloads 53
154 Radio Frequency Identification Device Based Emergency Department Critical Care Billing: A Framework for Actionable Intelligence

Authors: Shivaram P. Arunachalam, Mustafa Y. Sir, Andy Boggust, David M. Nestler, Thomas R. Hellmich, Kalyan S. Pasupathy

Abstract:

Emergency departments (EDs) provide urgent care to patients throughout the day in a complex and chaotic environment. Real-time location systems (RTLS) are increasingly being utilized in healthcare settings, and have shown to improve safety, reduce cost, and increase patient satisfaction. Radio Frequency Identification Device (RFID) data in an ED has been shown to compute variables such as patient-provider contact time, which is associated with patient outcomes such as 30-day hospitalization. These variables can provide avenues for improving ED operational efficiency. A major challenge with ED financial operations is under-coding of critical care services due to physicians’ difficulty reporting accurate times for critical care provided under Current Procedural Terminology (CPT) codes 99291 and 99292. In this work, the authors propose a framework to optimize ED critical care billing using RFID data. RFID estimated physician-patient contact times could accurately quantify direct critical care services which will help model a data-driven approach for ED critical care billing. This paper will describe the framework and provide insights into opportunities to prevent under coding as well as over coding to avoid insurance audits. Future work will focus on data analytics to demonstrate the feasibility of the framework described.

Keywords: critical care billing, CPT codes, emergency department, RFID

Procedia PDF Downloads 130
153 Effectiveness of Research Promotion Organizations in Higher Education and Research (ESR)

Authors: Jonas Sanon

Abstract:

The valorization of research is becoming a transversal instrument linking different sectors (academic, public and industrial). The practice of valorization seems to impact innovation techniques within companies where, there is often the implementation of industrial conventions of training through research (CIFRE), continuous training programs for employees, collaborations and partnerships around joint research and R&D laboratories focused on the needs of companies to improve or develop more efficient innovations. Furthermore, many public initiatives to support innovation and technology transfer have been developed at the international, European and national levels, with significant budget allocations. Thus, in the context of this work, we tried to analyze the way in which research transfer structures are evaluated within the Saclay ecosystem. In fact, the University-Paris-Saclay is one of the best French universities; it is made up of 10 university components, more than 275 laboratories and is in partnership with the largest French research centers This work mainly focused on how evaluations affected research transfer structures, how evaluations were conducted, and what the managers of research transfer structures thought about assessments. Thus, with the aid of the conducted interviews, it appears that the evaluations do not have a significant impact on the qualitative aspect of research and innovation, but is rather present a directive aspect to allow the structures to benefit or not from the financial resources to develop certain research work, sometimes directed and influenced by the market, some researchers might try to accentuate their research and experimentation work on themes that are not necessarily their areas of interest, but just to comply with the calls for proposed thematic projects. The field studies also outline the primary indicators used to assess the effectiveness of valorization structures as "the number of start-ups generated, the license agreements signed, the structure's patent portfolio, and the innovations of items developed from public research.". Finally, after mapping the actors, it became clear that the ecosystem of the University of Paris-Saclay benefits from a richness allowing it to better value its research in relation to the three categories of actors it has (internal, external and transversal), united and linked by a relationship of proximity of sharing and endowed with a real opportunity to innovate openly.

Keywords: research valorization, technology transfer, innovation, evaluation, impacts and performances, innovation policy

Procedia PDF Downloads 71
152 Commercial Automobile Insurance: A Practical Approach of the Generalized Additive Model

Authors: Nicolas Plamondon, Stuart Atkinson, Shuzi Zhou

Abstract:

The insurance industry is usually not the first topic one has in mind when thinking about applications of data science. However, the use of data science in the finance and insurance industry is growing quickly for several reasons, including an abundance of reliable customer data, ferocious competition requiring more accurate pricing, etc. Among the top use cases of data science, we find pricing optimization, customer segmentation, customer risk assessment, fraud detection, marketing, and triage analytics. The objective of this paper is to present an application of the generalized additive model (GAM) on a commercial automobile insurance product: an individually rated commercial automobile. These are vehicles used for commercial purposes, but for which there is not enough volume to apply pricing to several vehicles at the same time. The GAM model was selected as an improvement over GLM for its ease of use and its wide range of applications. The model was trained using the largest split of the data to determine model parameters. The remaining part of the data was used as testing data to verify the quality of the modeling activity. We used the Gini coefficient to evaluate the performance of the model. For long-term monitoring, commonly used metrics such as RMSE and MAE will be used. Another topic of interest in the insurance industry is to process of producing the model. We will discuss at a high level the interactions between the different teams with an insurance company that needs to work together to produce a model and then monitor the performance of the model over time. Moreover, we will discuss the regulations in place in the insurance industry. Finally, we will discuss the maintenance of the model and the fact that new data does not come constantly and that some metrics can take a long time to become meaningful.

Keywords: insurance, data science, modeling, monitoring, regulation, processes

Procedia PDF Downloads 74
151 Open Source Knowledge Management Approach to Manage and Disseminate Distributed Content in a Global Enterprise

Authors: Rahul Thakur, Onkar Chandel

Abstract:

Red Hat is the world leader in providing open source software and solutions. A global enterprise, like Red Hat, has unique issues of connecting employees with content because of distributed offices, multiple teams spread across geographies, multiple languages, and different cultures. Employees, of a global company, create content that is distributed across departments, teams, regions, and countries. This makes finding the best content difficult since owners keep iterating on the existing content. When employees are unable to find the content, they end up creating it once again and in the process duplicating existing material and effort. Also, employees may not find the relevant content and spend time reviewing obsolete duplicate, or irrelevant content. On an average, a person spends 15 minutes/day in failed searches that might result in missed business opportunities, employee frustration, and substandard deliverables. Red Hat Knowledge Management Office (KMO) applied 'open source strategy' to solve the above problems. Under the Open Source Strategy, decisions are taken collectively. The strategy aims at accomplishing common goals with the help of communities. The objectives of this initiative were to save employees' time, get them authentic content, improve their content search experience, avoid duplicate content creation, provide context based search, improve analytics, improve content management workflows, automate content classification, and automate content upload. This session will describe open source strategy, its applicability in content management, challenges, recommended solutions, and outcome.

Keywords: content classification, content management, knowledge management, open source

Procedia PDF Downloads 209
150 Q-Map: Clinical Concept Mining from Clinical Documents

Authors: Sheikh Shams Azam, Manoj Raju, Venkatesh Pagidimarri, Vamsi Kasivajjala

Abstract:

Over the past decade, there has been a steep rise in the data-driven analysis in major areas of medicine, such as clinical decision support system, survival analysis, patient similarity analysis, image analytics etc. Most of the data in the field are well-structured and available in numerical or categorical formats which can be used for experiments directly. But on the opposite end of the spectrum, there exists a wide expanse of data that is intractable for direct analysis owing to its unstructured nature which can be found in the form of discharge summaries, clinical notes, procedural notes which are in human written narrative format and neither have any relational model nor any standard grammatical structure. An important step in the utilization of these texts for such studies is to transform and process the data to retrieve structured information from the haystack of irrelevant data using information retrieval and data mining techniques. To address this problem, the authors present Q-Map in this paper, which is a simple yet robust system that can sift through massive datasets with unregulated formats to retrieve structured information aggressively and efficiently. It is backed by an effective mining technique which is based on a string matching algorithm that is indexed on curated knowledge sources, that is both fast and configurable. The authors also briefly examine its comparative performance with MetaMap, one of the most reputed tools for medical concepts retrieval and present the advantages the former displays over the latter.

Keywords: information retrieval, unified medical language system, syntax based analysis, natural language processing, medical informatics

Procedia PDF Downloads 131
149 Predictive Analytics Algorithms: Mitigating Elementary School Drop Out Rates

Authors: Bongs Lainjo

Abstract:

Educational institutions and authorities that are mandated to run education systems in various countries need to implement a curriculum that considers the possibility and existence of elementary school dropouts. This research focuses on elementary school dropout rates and the ability to replicate various predictive models carried out globally on selected Elementary Schools. The study was carried out by comparing the classical case studies in Africa, North America, South America, Asia and Europe. Some of the reasons put forward for children dropping out include the notion of being successful in life without necessarily going through the education process. Such mentality is coupled with a tough curriculum that does not take care of all students. The system has completely led to poor school attendance - truancy which continuously leads to dropouts. In this study, the focus is on developing a model that can systematically be implemented by school administrations to prevent possible dropout scenarios. At the elementary level, especially the lower grades, a child's perception of education can be easily changed so that they focus on the better future that their parents desire. To deal effectively with the elementary school dropout problem, strategies that are put in place need to be studied and predictive models are installed in every educational system with a view to helping prevent an imminent school dropout just before it happens. In a competency-based curriculum that most advanced nations are trying to implement, the education systems have wholesome ideas of learning that reduce the rate of dropout.

Keywords: elementary school, predictive models, machine learning, risk factors, data mining, classifiers, dropout rates, education system, competency-based curriculum

Procedia PDF Downloads 170
148 Is Sodium Channel Nav1.7 an Ideal Therapeutically Analgesic Target? A Systematic Review

Authors: Yutong Wan, John N. Wood

Abstract:

Introduction: SCN9A encoded Nav1.7 is an ideal therapeutic target with minimal side effects for the pharmaceutical industry because SCN9A variants can cause both human gains of function pain-related mutations and loss of function pain-free mutations. This study reviews the clinical effectiveness of existing Nav1.7 inhibitors, which theoretically should be powerful analgesics. Methods: A systematic review is conducted on the effectiveness of current Nav1.7 blockers undergoing clinical trials. Studies were mainly extracted from PubMed, U.S. National Library of Medicine Clinical Trials, World Health Organization International Clinical Trials Registry, ISRCTN registry platform, and Integrated Research Approval System by NHS. Only studies with full text available and those conducted using double-blinded, placebo controlled, and randomised designs and reporting at least one analgesic measurement were included. Results: Overall, 61 trials were screened, and eight studies covering PF 05089771 (Pfizer), TV 45070 (Teva & Xenon), and BIIB074 (Biogen) met the inclusion criteria. Most studies were excluded because results were not published. All three compounds demonstrated insignificant analgesic effects, and the comparison between PF 05089771 and pregabalin/ibuprofen showed that PF 05089771 was a much weaker analgesic. All three drug candidates only have mild side effects, indicating the potentials for further investigation of Nav1.7 antagonists. Discussion: The failure of current Nav1.7 small molecule inhibitors might attribute to ignorance of the key role of endogenous systems in Nav1.7 null mutants, the lack of selectivity and blocking potency, and central impermeability. The synergistic combination of analgesic drugs, a recent UCL patent, combining a small dose of Nav1.7 blockers and opioids or enkephalinase inhibitors dramatically enhanced the analgesic effects. Conclusion: The current clinical testing Nav1.7 blockers are generally disappointing. However, the newer generation of Nav1.7 targeting analgesics has overcome the major constraints of its predecessors.

Keywords: chronic pain, Nav1.7 blockers, SCN9A, systematic review

Procedia PDF Downloads 125
147 Cybersecurity Challenges and Solutions in ICT Management at the Federal Polytechnic, Ado-Ekiti: A Quantitative Study

Authors: Innocent Uzougbo Onwuegbuzie, Siene Elizabeth Eke

Abstract:

This study investigates cybersecurity challenges and solutions in managing Information and Communication Technology (ICT) at the Federal Polytechnic, Ado-Ekiti, South-West Nigeria. The rapid evolution of ICT has revolutionized organizational operations and impacted various sectors, including education, healthcare, and finance. While ICT advancements facilitate seamless communication, complex data analytics, and strategic decision-making, they also introduce significant cybersecurity risks such as data breaches, ransomware, and other malicious attacks. These threats jeopardize the confidentiality, integrity, and availability of information systems, necessitating robust cybersecurity measures. The primary aim of this research is to identify prevalent cybersecurity challenges in ICT management, evaluate their impact on the institution's operations, and assess the effectiveness of current cybersecurity solutions. Adopting a quantitative research approach, data was collected through surveys and structured questionnaires from students, staff, and IT professionals at the Federal Polytechnic, Ado-Ekiti. The findings underscore the critical need for continuous investment in cybersecurity technologies, employee and student training, and regulatory compliance to mitigate evolving cyber threats. This research contributes to bridging the knowledge gap in cybersecurity management and provides valuable insights into effective strategies and technologies for safeguarding ICT systems in educational institutions. The study's objectives are to enhance the security posture of the Federal Polytechnic, Ado-Ekiti, in an increasingly digital world by identifying and addressing the cybersecurity challenges faced by its ICT management.

Keywords: cybersecurity challenges, cyber threat mitigation, federal polytechnic Ado-Ekiti, ICT management

Procedia PDF Downloads 31
146 Photocatalytic Eco-Active Ceramic Slabs to Abate Air Pollution under LED Light

Authors: Claudia L. Bianchi, Giuseppina Cerrato, Federico Galli, Federica Minozzi, Valentino Capucci

Abstract:

At the beginning of the industrial productions, porcelain gres tiles were considered as just a technical material, aesthetically not very beautiful. Today thanks to new industrial production methods, both properties, and beauty of these materials completely fit the market requests. In particular, the possibility to prepare slabs of large sizes is the new frontier of building materials. Beside these noteworthy architectural features, new surface properties have been introduced in the last generation of these materials. In particular, deposition of TiO₂ transforms the traditional ceramic into a photocatalytic eco-active material able to reduce polluting molecules present in air and water, to eliminate bacteria and to reduce the surface dirt thanks to the self-cleaning property. The problem of photocatalytic materials resides in the fact that it is necessary a UV light source to activate the oxidation processes on the surface of the material, processes that are turned off inexorably when the material is illuminated by LED lights and, even more so, when we are in darkness. First, it was necessary a thorough study change the existing plants to deposit the photocatalyst very evenly and this has been done thanks to the advent of digital printing and the development of an ink custom-made that stabilizes the powdered TiO₂ in its formulation. In addition, the commercial TiO₂, which is used for the traditional photocatalytic coating, has been doped with metals in order to activate it even in the visible region and thus in the presence of sunlight or LED. Thanks to this active coating, ceramic slabs are able to purify air eliminating odors and VOCs, and also can be cleaned with very soft detergents due to the self-cleaning properties given by the TiO₂ present at the ceramic surface. Moreover, the presence of dopant metals (patent WO2016157155) also allows the material to work as well as antibacterial in the dark, by eliminating one of the negative features of photocatalytic building materials that have so far limited its use on a large scale. Considering that we are constantly in contact with bacteria, some of which are dangerous for health. Active tiles are 99,99% efficient on all bacteria, from the most common such as Escherichia coli to the most dangerous such as Staphilococcus aureus Methicillin-resistant (MRSA). DIGITALIFE project LIFE13 ENV/IT/000140 – award for best project of October 2017.

Keywords: Ag-doped microsized TiO₂, eco-active ceramic, photocatalysis, digital coating

Procedia PDF Downloads 224
145 Towards an Equitable Proprietary Regime: Property Rights Over Human Genes as a Case Study

Authors: Aileen Editha

Abstract:

The legal recognition of property rights over human genes is a divisive topic to which there is no resolution. As a frequently discussed topic, scholars and practitioners often highlight the inadequacies of a proprietary regime. However, little has been said in regard to the nature of human genetic materials (HGMs). This paper proposes approaching the issue of property over HGMs from an alternative perspective that looks at the personal and social value and valuation of HGMs. This paper will highlight how the unique and unresolved status of HGMs is incompatible with the main tenets of property and, consequently, contributes to legal ambiguity and uncertainty in the regulation of property rights over human genes. HGMs are perceived as part of nature and a free-for-all while also being within an individual’s private sphere. Additionally, it is also considered to occupy a unique “not-private-nor-public” status. This limbo-like position clashes with property’s fundamental characteristic that relies heavily on a clear public/private dichotomy. Moreover, as property is intrinsically linked to the legal recognition of one’s personhood, this irresolution benefits some while disadvantages others. In particular, it demands the publicization of once-private genes for the “common good” but subsequently encourages privatization (through labor) of these now-public genes. This results in the gain of some (already privileged) individuals while enabling the disenfranchisement of members of minority groups, such as Indigenous communities. This paper will discuss real and intellectual property rights over human genes, such as the right to income or patent rights, in Canada and the US. This paper advocates for a sui generis approach to governing rights and interests over human genes that would not rely on having a strict public/private dichotomy. Not only would this improve legal certainty and clarity, but it would also alleviate—or, at the very least, minimize—the role that the current law plays in further entrenching existing systemic inequalities. Despite the specificity of this topic, this paper argues that there are broader lessons to be learned. This issue is an insightful case study on the interconnection of various principles in law, society, and property, and what must be done when discordance between one or more of those principles has detrimental societal outcomes. Ultimately, it must be remembered that property is an adaptable and malleable instrument that can be developed to ensure it contributes to equity and flourishing.

Keywords: property rights, human genetic materials, critical legal scholarship, systemic inequalities

Procedia PDF Downloads 78
144 Learning Dynamic Representations of Nodes in Temporally Variant Graphs

Authors: Sandra Mitrovic, Gaurav Singh

Abstract:

In many industries, including telecommunications, churn prediction has been a topic of active research. A lot of attention has been drawn on devising the most informative features, and this area of research has gained even more focus with spread of (social) network analytics. The call detail records (CDRs) have been used to construct customer networks and extract potentially useful features. However, to the best of our knowledge, no studies including network features have yet proposed a generic way of representing network information. Instead, ad-hoc and dataset dependent solutions have been suggested. In this work, we build upon a recently presented method (node2vec) to obtain representations for nodes in observed network. The proposed approach is generic and applicable to any network and domain. Unlike node2vec, which assumes a static network, we consider a dynamic and time-evolving network. To account for this, we propose an approach that constructs the feature representation of each node by generating its node2vec representations at different timestamps, concatenating them and finally compressing using an auto-encoder-like method in order to retain reasonably long and informative feature vectors. We test the proposed method on churn prediction task in telco domain. To predict churners at timestamp ts+1, we construct training and testing datasets consisting of feature vectors from time intervals [t1, ts-1] and [t2, ts] respectively, and use traditional supervised classification models like SVM and Logistic Regression. Observed results show the effectiveness of proposed approach as compared to ad-hoc feature selection based approaches and static node2vec.

Keywords: churn prediction, dynamic networks, node2vec, auto-encoders

Procedia PDF Downloads 312
143 Kantian Epistemology in Examination of the Axiomatic Principles of Economics: The Synthetic a Priori in the Economic Structure of Society

Authors: Mirza Adil Ahmad Mughal

Abstract:

Transcendental analytics, in the critique of pure reason, combines space and time as conditions of the possibility of the phenomenon from the transcendental aesthetic with the pure magnitude-intuition notion. The property of continuity as a qualitative result of the additive magnitude brings the possibility of connecting with experience, even though only as a potential because of the a priori necessity from assumption, as syntheticity of the a priori task of a scientific method of philosophy given by Kant, which precludes the application of categories to something not empirically reducible to the content of such a category's corresponding and possible object. This continuity as the qualitative result of a priori constructed notion of magnitude lies as a fundamental assumption and property of, what in Microeconomic theory is called as, 'choice rules' which combine the potentially-empirical and practical budget-price pairs with preference relations. This latter result is the purest qualitative side of the choice rules', otherwise autonomously, quantitative nature. The theoretical, barring the empirical, nature of this qualitative result is a synthetic a priori truth, which, if at all, it should be, if the axiomatic structure of the economic theory is held to be correct. It has a potentially verifiable content as its possible object in the form of quantitative price-budget pairs. Yet, the object that serves the respective Kantian category is qualitative itself, which is utility. This article explores the validity of Kantian qualifications for this application of 'categories' to the economic structure of society.

Keywords: categories of understanding, continuity, convexity, psyche, revealed preferences, synthetic a priori

Procedia PDF Downloads 95