Search results for: transient data
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 24684

Search results for: transient data

24294 Study on Impact of Road Loads on Full Vehicle Squeak and Rattle Performance

Authors: R. Praveen, B. R. Chandan Ravi, M. Harikrishna

Abstract:

Squeak and rattle noises are the most annoying transient vehicle noises produced due to different terrain conditions. Interpretation and prohibition of squeak and rattle noises are the dominant aspects of a vehicle refinement. This paper describes the computer-aided engineering (CAE) approach to evaluating the full vehicle squeak and rattle performance with the measured road surface profile as enforced excitation at the tire patch points. The E-Line methodology has been used to predict the relative displacement at the interface points and the risk areas were identified. Squeak and rattle performance has been evaluated at different speeds and at different road conditions to understand the vehicle characteristics. The competence of the process in predicting the risk and root cause of the problems showcased us a pleasing conformity between the physical testing and CAE simulation results.

Keywords: e-line, enforced excitation, full vehicle, squeak and rattle, road excitation

Procedia PDF Downloads 122
24293 A Review on Intelligent Systems for Geoscience

Authors: R Palson Kennedy, P.Kiran Sai

Abstract:

This article introduces machine learning (ML) researchers to the hurdles that geoscience problems present, as well as the opportunities for improvement in both ML and geosciences. This article presents a review from the data life cycle perspective to meet that need. Numerous facets of geosciences present unique difficulties for the study of intelligent systems. Geosciences data is notoriously difficult to analyze since it is frequently unpredictable, intermittent, sparse, multi-resolution, and multi-scale. The first half addresses data science’s essential concepts and theoretical underpinnings, while the second section contains key themes and sharing experiences from current publications focused on each stage of the data life cycle. Finally, themes such as open science, smart data, and team science are considered.

Keywords: Data science, intelligent system, machine learning, big data, data life cycle, recent development, geo science

Procedia PDF Downloads 117
24292 Acupoint Injection of High Concentration of Glucose Attenuates Mice Chronic Pain and Depression Comorbidity

Authors: Chanya Inprasit, Yi-Wen Lin

Abstract:

Inflammation causes changes of peripheral and central nervous system properties, affecting both neuronal and non-neuronal cells, resulting in inflammatory pain. Acupoint injection (AI) was developed in the 1950s and has been widely used for relieving pain. It is an acupoint-stimulating technique that utilizes anatomically based meridians derived from Chinese medicine theory. AI has been accepted as an effective treatment and is thought to display superior results when compared to traditional acupuncture methods. However, the mechanism of AI needs to be ratified by more scientific evidence in order to support the theory and its therapeutic development. In this study, we explored the effect of AI on the comorbidity of chronic pain and depression. Mice hindpaw was injected by complete Freund’s adjuvant (CFA) to induce the condition of chronic pain. Measurements of mechanical and thermal hyperalgesia and depression-like behavior were analyzed. The results indicated a positive tendency to AI treatment. The comorbidity of chronic pain and depression was investigated with relation to transient receptor potential V1 (TRPV1) mechanism through the use of TRPV1 gene deletion. The expression of nociceptors such as voltage-gated sodium channels (Navs) or TRPV1, was significantly down-regulated by AI. The expression of inflammation-activated molecules: astrocytic marker glial fibrillary acidic protein (GFAP), the microglial marker Iba-1, S100B, and related kinases, were reversed by AI in both the peripheral and central nervous system. Taken together, these data provided a detailed molecular mechanism of AI-induced analgesia and anti-inflammatory properties. This finding may be utilized for clinical practice to treat chronic pain and depression comorbidity.

Keywords: inflammatory pain, acupoint injection, TRPV1, GFAP, S100B

Procedia PDF Downloads 127
24291 Study of Heat Transfer in the Absorber Plates of a Flat-Plate Solar Collector Using Dual-Phase-Lag Model

Authors: Yu-Ching Yang, Haw-Long Lee, Win-Jin Chang

Abstract:

The present work numerically analyzes the transient heat transfer in the absorber plates of a flat-plate solar collector based on the dual-phase-lag (DPL) heat conduction model. An efficient numerical scheme involving the hybrid application of the Laplace transform and control volume methods is used to solve the linear hyperbolic heat conduction equation. This work also examines the effect of different medium parameters on the behavior of heat transfer. Results show that, while the heat-flux phase lag induces thermal waves in the medium, the temperature-gradient phase lag smoothens the thermal waves by promoting non-Fourier diffusion-like conduction into the medium.

Keywords: absorber plates, dual-phase-lag, non-Fourier, solar collector

Procedia PDF Downloads 364
24290 Data Quality as a Pillar of Data-Driven Organizations: Exploring the Benefits of Data Mesh

Authors: Marc Bachelet, Abhijit Kumar Chatterjee, José Manuel Avila

Abstract:

Data quality is a key component of any data-driven organization. Without data quality, organizations cannot effectively make data-driven decisions, which often leads to poor business performance. Therefore, it is important for an organization to ensure that the data they use is of high quality. This is where the concept of data mesh comes in. Data mesh is an organizational and architectural decentralized approach to data management that can help organizations improve the quality of data. The concept of data mesh was first introduced in 2020. Its purpose is to decentralize data ownership, making it easier for domain experts to manage the data. This can help organizations improve data quality by reducing the reliance on centralized data teams and allowing domain experts to take charge of their data. This paper intends to discuss how a set of elements, including data mesh, are tools capable of increasing data quality. One of the key benefits of data mesh is improved metadata management. In a traditional data architecture, metadata management is typically centralized, which can lead to data silos and poor data quality. With data mesh, metadata is managed in a decentralized manner, ensuring accurate and up-to-date metadata, thereby improving data quality. Another benefit of data mesh is the clarification of roles and responsibilities. In a traditional data architecture, data teams are responsible for managing all aspects of data, which can lead to confusion and ambiguity in responsibilities. With data mesh, domain experts are responsible for managing their own data, which can help provide clarity in roles and responsibilities and improve data quality. Additionally, data mesh can also contribute to a new form of organization that is more agile and adaptable. By decentralizing data ownership, organizations can respond more quickly to changes in their business environment, which in turn can help improve overall performance by allowing better insights into business as an effect of better reports and visualization tools. Monitoring and analytics are also important aspects of data quality. With data mesh, monitoring, and analytics are decentralized, allowing domain experts to monitor and analyze their own data. This will help in identifying and addressing data quality problems in quick time, leading to improved data quality. Data culture is another major aspect of data quality. With data mesh, domain experts are encouraged to take ownership of their data, which can help create a data-driven culture within the organization. This can lead to improved data quality and better business outcomes. Finally, the paper explores the contribution of AI in the coming years. AI can help enhance data quality by automating many data-related tasks, like data cleaning and data validation. By integrating AI into data mesh, organizations can further enhance the quality of their data. The concepts mentioned above are illustrated by AEKIDEN experience feedback. AEKIDEN is an international data-driven consultancy that has successfully implemented a data mesh approach. By sharing their experience, AEKIDEN can help other organizations understand the benefits and challenges of implementing data mesh and improving data quality.

Keywords: data culture, data-driven organization, data mesh, data quality for business success

Procedia PDF Downloads 109
24289 Longitudinal Static and Dynamic Stability of a Typical Reentry Body in Subsonic Conditions Using Computational Fluid Dynamics

Authors: M. Jathaveda, Joben Leons, G. Vidya

Abstract:

Reentry from orbit is a critical phase in the entry trajectory. For a non-propulsive ballistic entry, static and dynamic stability play an important role in the trajectory, especially for the safe deployment of parachutes, typically at subsonic Mach numbers. Static stability of flight vehicles are being estimated through CFD techniques routinely. Advances in CFD software as well as computational facilities have enabled the estimation of the dynamic stability derivatives also through CFD techniques. Longitudinal static and dynamic stability of a typical reentry body for subsonic Mach number of 0.6 is predicted using commercial software CFD++ and presented here. Steady state simulations are carried out for α = 2° on an unstructured grid using SST k-ω model. Transient simulation using forced oscillation method is used to compute pitch damping derivatives.

Keywords: stability, typical reentry body, subsonic, static and dynamic

Procedia PDF Downloads 85
24288 Big Data Analysis with RHadoop

Authors: Ji Eun Shin, Byung Ho Jung, Dong Hoon Lim

Abstract:

It is almost impossible to store or analyze big data increasing exponentially with traditional technologies. Hadoop is a new technology to make that possible. R programming language is by far the most popular statistical tool for big data analysis based on distributed processing with Hadoop technology. With RHadoop that integrates R and Hadoop environment, we implemented parallel multiple regression analysis with different sizes of actual data. Experimental results showed our RHadoop system was much faster as the number of data nodes increases. We also compared the performance of our RHadoop with lm function and big lm packages available on big memory. The results showed that our RHadoop was faster than other packages owing to paralleling processing with increasing the number of map tasks as the size of data increases.

Keywords: big data, Hadoop, parallel regression analysis, R, RHadoop

Procedia PDF Downloads 412
24287 A Mutually Exclusive Task Generation Method Based on Data Augmentation

Authors: Haojie Wang, Xun Li, Rui Yin

Abstract:

In order to solve the memorization overfitting in the meta-learning MAML algorithm, a method of generating mutually exclusive tasks based on data augmentation is proposed. This method generates a mutex task by corresponding one feature of the data to multiple labels, so that the generated mutex task is inconsistent with the data distribution in the initial dataset. Because generating mutex tasks for all data will produce a large number of invalid data and, in the worst case, lead to exponential growth of computation, this paper also proposes a key data extraction method, that only extracts part of the data to generate the mutex task. The experiments show that the method of generating mutually exclusive tasks can effectively solve the memorization overfitting in the meta-learning MAML algorithm.

Keywords: data augmentation, mutex task generation, meta-learning, text classification.

Procedia PDF Downloads 72
24286 Efficient Positioning of Data Aggregation Point for Wireless Sensor Network

Authors: Sifat Rahman Ahona, Rifat Tasnim, Naima Hassan

Abstract:

Data aggregation is a helpful technique for reducing the data communication overhead in wireless sensor network. One of the important tasks of data aggregation is positioning of the aggregator points. There are a lot of works done on data aggregation. But, efficient positioning of the aggregators points is not focused so much. In this paper, authors are focusing on the positioning or the placement of the aggregation points in wireless sensor network. Authors proposed an algorithm to select the aggregators positions for a scenario where aggregator nodes are more powerful than sensor nodes.

Keywords: aggregation point, data communication, data aggregation, wireless sensor network

Procedia PDF Downloads 132
24285 Spatial Econometric Approaches for Count Data: An Overview and New Directions

Authors: Paula Simões, Isabel Natário

Abstract:

This paper reviews a number of theoretical aspects for implementing an explicit spatial perspective in econometrics for modelling non-continuous data, in general, and count data, in particular. It provides an overview of the several spatial econometric approaches that are available to model data that are collected with reference to location in space, from the classical spatial econometrics approaches to the recent developments on spatial econometrics to model count data, in a Bayesian hierarchical setting. Considerable attention is paid to the inferential framework, necessary for structural consistent spatial econometric count models, incorporating spatial lag autocorrelation, to the corresponding estimation and testing procedures for different assumptions, to the constrains and implications embedded in the various specifications in the literature. This review combines insights from the classical spatial econometrics literature as well as from hierarchical modeling and analysis of spatial data, in order to look for new possible directions on the processing of count data, in a spatial hierarchical Bayesian econometric context.

Keywords: spatial data analysis, spatial econometrics, Bayesian hierarchical models, count data

Procedia PDF Downloads 563
24284 A NoSQL Based Approach for Real-Time Managing of Robotics's Data

Authors: Gueidi Afef, Gharsellaoui Hamza, Ben Ahmed Samir

Abstract:

This paper deals with the secret of the continual progression data that new data management solutions have been emerged: The NoSQL databases. They crossed several areas like personalization, profile management, big data in real-time, content management, catalog, view of customers, mobile applications, internet of things, digital communication and fraud detection. Nowadays, these database management systems are increasing. These systems store data very well and with the trend of big data, a new challenge’s store demands new structures and methods for managing enterprise data. The new intelligent machine in the e-learning sector, thrives on more data, so smart machines can learn more and faster. The robotics are our use case to focus on our test. The implementation of NoSQL for Robotics wrestle all the data they acquire into usable form because with the ordinary type of robotics; we are facing very big limits to manage and find the exact information in real-time. Our original proposed approach was demonstrated by experimental studies and running example used as a use case.

Keywords: NoSQL databases, database management systems, robotics, big data

Procedia PDF Downloads 326
24283 Fuzzy Optimization Multi-Objective Clustering Ensemble Model for Multi-Source Data Analysis

Authors: C. B. Le, V. N. Pham

Abstract:

In modern data analysis, multi-source data appears more and more in real applications. Multi-source data clustering has emerged as a important issue in the data mining and machine learning community. Different data sources provide information about different data. Therefore, multi-source data linking is essential to improve clustering performance. However, in practice multi-source data is often heterogeneous, uncertain, and large. This issue is considered a major challenge from multi-source data. Ensemble is a versatile machine learning model in which learning techniques can work in parallel, with big data. Clustering ensemble has been shown to outperform any standard clustering algorithm in terms of accuracy and robustness. However, most of the traditional clustering ensemble approaches are based on single-objective function and single-source data. This paper proposes a new clustering ensemble method for multi-source data analysis. The fuzzy optimized multi-objective clustering ensemble method is called FOMOCE. Firstly, a clustering ensemble mathematical model based on the structure of multi-objective clustering function, multi-source data, and dark knowledge is introduced. Then, rules for extracting dark knowledge from the input data, clustering algorithms, and base clusterings are designed and applied. Finally, a clustering ensemble algorithm is proposed for multi-source data analysis. The experiments were performed on the standard sample data set. The experimental results demonstrate the superior performance of the FOMOCE method compared to the existing clustering ensemble methods and multi-source clustering methods.

Keywords: clustering ensemble, multi-source, multi-objective, fuzzy clustering

Procedia PDF Downloads 156
24282 Modeling Activity Pattern Using XGBoost for Mining Smart Card Data

Authors: Eui-Jin Kim, Hasik Lee, Su-Jin Park, Dong-Kyu Kim

Abstract:

Smart-card data are expected to provide information on activity pattern as an alternative to conventional person trip surveys. The focus of this study is to propose a method for training the person trip surveys to supplement the smart-card data that does not contain the purpose of each trip. We selected only available features from smart card data such as spatiotemporal information on the trip and geographic information system (GIS) data near the stations to train the survey data. XGboost, which is state-of-the-art tree-based ensemble classifier, was used to train data from multiple sources. This classifier uses a more regularized model formalization to control the over-fitting and show very fast execution time with well-performance. The validation results showed that proposed method efficiently estimated the trip purpose. GIS data of station and duration of stay at the destination were significant features in modeling trip purpose.

Keywords: activity pattern, data fusion, smart-card, XGboost

Procedia PDF Downloads 219
24281 Recombination Center Levels in Gold and Platinum Doped N-type Silicon for High-Speed Thyristor

Authors: Nam Chol Yu, GyongIl Chu, HoJong Ri

Abstract:

Using DLTS (Deep-level transient spectroscopy) measurement techniques, we determined the dominant recombination center levels (defects of both A and B) in gold and platinum doped n-type silicon. Also, the injection and temperature dependence of the Shockley-Read-Hall (SRH) carrier lifetime was studied under low-level injection and high-level injection. Here measurements show that the dominant level under low-level injection located at EC-0.25 eV (A) correlated to the Pt+G1 and the dominant level under high-level injection located at EC-0.54 eV (B) correlated to the Au+G4. Finally, A and B are the same dominant levels for controlling the lifetime in gold-platinum doped n-silicon.

Keywords: recombination center level, lifetime, carrier lifetime control, Gold, Platinum, Silicon

Procedia PDF Downloads 41
24280 Value Chain Network: A Social Network Analysis of the Value Chain Actors of Recycled Polymer Products in Lagos Metropolis, Nigeria

Authors: Olamide Shittu, Olayinka Akanle

Abstract:

Value Chain Analysis is a common method of examining the stages involved in the production of a product, mostly agricultural produce, from the input to the consumption stage including the actors involved in each stage. However, the Functional Institutional Analysis is the most common method in literature employed to analyze the value chain of products. Apart from studying the relatively neglected phenomenon of recycled polymer products in Lagos Metropolis, this paper adopted the use of social network analysis to attempt a grounded theory of the nature of social network that exists among the value chain actors of the subject matter. The study adopted a grounded theory approach by conducting in-depth interviews, administering questionnaires and conducting observations among the identified value chain actors of recycled polymer products in Lagos Metropolis, Nigeria. The thematic analysis of the collected data gave the researchers the needed background to formulate a truly representative network of the social relationships among the value chain actors of recycled polymer products in Lagos Metropolis. The paper introduced concepts such as Transient and Perennial Social Ties to explain the observed social relations among the actors. Some actors have more social capital than others as a result of the structural holes that exist in their triad network. Households and resource recoverers are at disadvantaged position in the network as they have high constraints in their relationships with other actors. The study attempted to provide a new perspective in the study of the environmental value chain by analyzing the network of actors to bring about policy action points and improve recycling in Nigeria. Government and social entrepreneurs can exploit the structural holes that exist in the network for the socio-economic and sustainable development of the state.

Keywords: recycled polymer products, social network analysis, social ties, value chain analysis

Procedia PDF Downloads 382
24279 A Mutually Exclusive Task Generation Method Based on Data Augmentation

Authors: Haojie Wang, Xun Li, Rui Yin

Abstract:

In order to solve the memorization overfitting in the model-agnostic meta-learning MAML algorithm, a method of generating mutually exclusive tasks based on data augmentation is proposed. This method generates a mutex task by corresponding one feature of the data to multiple labels so that the generated mutex task is inconsistent with the data distribution in the initial dataset. Because generating mutex tasks for all data will produce a large number of invalid data and, in the worst case, lead to an exponential growth of computation, this paper also proposes a key data extraction method that only extract part of the data to generate the mutex task. The experiments show that the method of generating mutually exclusive tasks can effectively solve the memorization overfitting in the meta-learning MAML algorithm.

Keywords: mutex task generation, data augmentation, meta-learning, text classification.

Procedia PDF Downloads 113
24278 Dependence of Shaft Stiffness on the Crack Location

Authors: H. M. Mobarak, Helen Wu, Chunhui Yang

Abstract:

In this study, an analytical model is developed to study crack breathing behavior under the effect of crack location and unbalance force. Crack breathing behavior is determined using effectual bending angle by studying the transient change in closed area of the crack. The status of the crack of a balanced shaft is symmetrical about shaft rotational angle and the duration of each crack status remains unchanged. The global stiffness of the balanced shaft is independent of crack location. Different crack breathing behavior for the unbalanced shaft has been observed. The influence of crack location on the unbalanced shaft stiffness can be divided into three regions. When the crack is located between 0.3L and 0.8335L, where L is the total length of the shaft, the unbalanced shaft is less stiff and when located outside this region it is stiffer than the balanced shaft. It was also found that unbalanced shaft stiffness has a maximum value with a crack at 0.1946L, a minimum value at 0.8053L and same value as balanced shaft at 0.3L and 0.8335L.

Keywords: cracked shaft, crack location, shaft stiffness, unbalanced force

Procedia PDF Downloads 286
24277 Revolutionizing Traditional Farming Using Big Data/Cloud Computing: A Review on Vertical Farming

Authors: Milind Chaudhari, Suhail Balasinor

Abstract:

Due to massive deforestation and an ever-increasing population, the organic content of the soil is depleting at a much faster rate. Due to this, there is a big chance that the entire food production in the world will drop by 40% in the next two decades. Vertical farming can help in aiding food production by leveraging big data and cloud computing to ensure plants are grown naturally by providing the optimum nutrients sunlight by analyzing millions of data points. This paper outlines the most important parameters in vertical farming and how a combination of big data and AI helps in calculating and analyzing these millions of data points. Finally, the paper outlines how different organizations are controlling the indoor environment by leveraging big data in enhancing food quantity and quality.

Keywords: big data, IoT, vertical farming, indoor farming

Procedia PDF Downloads 148
24276 Data Challenges Facing Implementation of Road Safety Management Systems in Egypt

Authors: A. Anis, W. Bekheet, A. El Hakim

Abstract:

Implementing a Road Safety Management System (SMS) in a crowded developing country such as Egypt is a necessity. Beginning a sustainable SMS requires a comprehensive reliable data system for all information pertinent to road crashes. In this paper, a survey for the available data in Egypt and validating it for using in an SMS in Egypt. The research provides some missing data, and refer to the unavailable data in Egypt, looking forward to the contribution of the scientific society, the authorities, and the public in solving the problem of missing or unreliable crash data. The required data for implementing an SMS in Egypt are divided into three categories; the first is available data such as fatality and injury rates and it is proven in this research that it may be inconsistent and unreliable, the second category of data is not available, but it may be estimated, an example of estimating vehicle cost is available in this research, the third is not available and can be measured case by case such as the functional and geometric properties of a facility. Some inquiries are provided in this research for the scientific society, such as how to improve the links among stakeholders of road safety in order to obtain a consistent, non-biased, and reliable data system.

Keywords: road safety management system, road crash, road fatality, road injury

Procedia PDF Downloads 97
24275 Big Data-Driven Smart Policing: Big Data-Based Patrol Car Dispatching in Abu Dhabi, UAE

Authors: Oualid Walid Ben Ali

Abstract:

Big Data has become one of the buzzwords today. The recent explosion of digital data has led the organization, either private or public, to a new era towards a more efficient decision making. At some point, business decided to use that concept in order to learn what make their clients tick with phrases like ‘sales funnel’ analysis, ‘actionable insights’, and ‘positive business impact’. So, it stands to reason that Big Data was viewed through green (read: money) colored lenses. Somewhere along the line, however someone realized that collecting and processing data doesn’t have to be for business purpose only, but also could be used for other purposes to assist law enforcement or to improve policing or in road safety. This paper presents briefly, how Big Data have been used in the fields of policing order to improve the decision making process in the daily operation of the police. As example, we present a big-data driven system which is sued to accurately dispatch the patrol cars in a geographic environment. The system is also used to allocate, in real-time, the nearest patrol car to the location of an incident. This system has been implemented and applied in the Emirate of Abu Dhabi in the UAE.

Keywords: big data, big data analytics, patrol car allocation, dispatching, GIS, intelligent, Abu Dhabi, police, UAE

Procedia PDF Downloads 465
24274 Microstructural Evolution of an Interface Region in a Nickel-Based Superalloy Joint Produced by Direct Energy Deposition

Authors: Matthew Ferguson, Tatyana Konkova, Ioannis Violatos

Abstract:

Microstructure analysis of additively manufactured (AM) materials is an important step in understanding the interrelationship between mechanical properties and materials performance. Literature on the effect of laser-based AM process parameters on the microstructure in the substrate-deposit interface is limited. The interface region, the adjoining area of substrate and deposit, is characterized by the presence of the fusion zone (FZ) and heat-affected zone (HAZ), experiencing rapid thermal gyrations resulting in thermal-induced transformations. Inconel 718 was utilized as work material for both the substrate and deposit. Three blocks of Inconel 718 material were deposited by Direct Energy Deposition (DED) using three different laser powers, 550W, 750W and 950W, respectively. A coupled thermo-mechanical transient approach was utilized to correlate temperature history to the evolution of microstructure. The thermal history of the deposition process was monitored with the thermocouples installed inside the substrate material. The interface region of the blocks was analyzed with Optical Microscopy (OM) and Scanning Electron Microscopy (SEM), including the electron back-scattered diffraction (EBSD) technique. Laser power was found to influence the dissolution of intermetallic precipitated phases in the substrate and grain growth in the interface region. Microstructure and thermal history data were utilized to draw conclusive comparisons between the investigated process parameters.

Keywords: additive manufacturing, direct energy deposition, electron back-scattered diffraction, finite element analysis, inconel 718, microstructure, optical microscopy, scanning electron microscopy, substrate-deposit interface region

Procedia PDF Downloads 176
24273 Mining Multicity Urban Data for Sustainable Population Relocation

Authors: Xu Du, Aparna S. Varde

Abstract:

In this research, we propose to conduct diagnostic and predictive analysis about the key factors and consequences of urban population relocation. To achieve this goal, urban simulation models extract the urban development trends as land use change patterns from a variety of data sources. The results are treated as part of urban big data with other information such as population change and economic conditions. Multiple data mining methods are deployed on this data to analyze nonlinear relationships between parameters. The result determines the driving force of population relocation with respect to urban sprawl and urban sustainability and their related parameters. Experiments so far reveal that data mining methods discover useful knowledge from the multicity urban data. This work sets the stage for developing a comprehensive urban simulation model for catering to specific questions by targeted users. It contributes towards achieving sustainability as a whole.

Keywords: data mining, environmental modeling, sustainability, urban planning

Procedia PDF Downloads 274
24272 A Patient-Centered Approach to Clinical Trial Development: Real-World Evidence from a Canadian Medical Cannabis Clinic

Authors: Lucile Rapin, Cynthia El Hage, Rihab Gamaoun, Maria-Fernanda Arboleda, Erin Prosk

Abstract:

Introduction: Sante Cannabis (SC), a Canadian group of clinics dedicated to medical cannabis, based in Montreal and in the province of Quebec, has served more than 8000 patients seeking cannabis-based treatment over the past five years. As randomized clinical trials with natural medical cannabis are scarce, real-world evidence offers the opportunity to fill research gaps between scientific evidence and clinical practice. Data on the use of medical cannabis products from SC patients were prospectively collected, leading to a large real-world database on the use of medical cannabis. The aim of this study was to report information on the profiles of both patients and prescribed medical cannabis products at SC clinics, and to assess the safety of medical cannabis among Canadian patients. Methods: This is an observational retrospective study of 1342 adult patients who were authorized with medical cannabis products between October 2017 and September 2019. Information regarding demographic characteristics, therapeutic indications for medical cannabis use, patterns in dosing and dosage form of medical cannabis and adverse effects over one-year follow-up (initial and 4 follow-up (FUP) visits) were collected. Results: 59% of SC patients were female, with a mean age of 56.7 (SD= 15.6, range= (19-97)). Cannabis products were authorized mainly for patients with a diagnosis of chronic pain (68.8% of patients), cancer (6.7%), neurological disorders (5.6%), and mood disorders (5.4 %). At initial visit, a large majority (70%) of patients were authorized exclusively medical cannabis products, 27% were authorized a combination of pharmaceutical cannabinoids and medical cannabis and 3% were prescribed only pharmaceutical cannabinoids. This pattern was recurrent over the one-year follow-up. Overall, oil was the preferred formulation (average over visits 72.5%) followed by a combination of oil and dry (average 19%), other routes of administration accounted for less than 4%. Patients were predominantly prescribed products with a balanced THC:CBD ratio (59%-75% across visits). 28% of patients reported at least one adverse effect (AE) at the 3-month follow-up visit and 12% at the six-month FUP visit. 84.8% of total AEs were mild and transient. No serious AE was reported. Overall, the most common side effects reported were dizziness (11.95% of total AEs), drowsiness (11.4%), dry mouth (5.5%), nausea (4.8%), headaches (4.6%), cough (4.4%), anxiety (4.1%) and euphoria (3.5%). Other adverse effects accounted for less than 3% of total AE. Conclusion: Our results confirm that the primary area of clinical use for medical cannabis is in pain management. Patients in this cohort are largely utilizing plant-based cannabis oil products with a balanced ratio of THC:CBD. Reported adverse effects were mild and included dizziness and drowsiness. This real-world data confirms the tolerable safety profile of medical cannabis and suggests medical indications not yet validated in controlled clinical trials. Such data offers an important opportunity for the investigation of the long-term effects of cannabinoid exposure in real-life conditions. Real-world evidence can be used to direct clinical trial research efforts on specific indications and dosing patterns for product development.

Keywords: medical cannabis, safety, real-world data, Canada

Procedia PDF Downloads 106
24271 Model Order Reduction for Frequency Response and Effect of Order of Method for Matching Condition

Authors: Aref Ghafouri, Mohammad javad Mollakazemi, Farhad Asadi

Abstract:

In this paper, model order reduction method is used for approximation in linear and nonlinearity aspects in some experimental data. This method can be used for obtaining offline reduced model for approximation of experimental data and can produce and follow the data and order of system and also it can match to experimental data in some frequency ratios. In this study, the method is compared in different experimental data and influence of choosing of order of the model reduction for obtaining the best and sufficient matching condition for following the data is investigated in format of imaginary and reality part of the frequency response curve and finally the effect and important parameter of number of order reduction in nonlinear experimental data is explained further.

Keywords: frequency response, order of model reduction, frequency matching condition, nonlinear experimental data

Procedia PDF Downloads 375
24270 An Empirical Study of the Impacts of Big Data on Firm Performance

Authors: Thuan Nguyen

Abstract:

In the present time, data to a data-driven knowledge-based economy is the same as oil to the industrial age hundreds of years ago. Data is everywhere in vast volumes! Big data analytics is expected to help firms not only efficiently improve performance but also completely transform how they should run their business. However, employing the emergent technology successfully is not easy, and assessing the roles of big data in improving firm performance is even much harder. There was a lack of studies that have examined the impacts of big data analytics on organizational performance. This study aimed to fill the gap. The present study suggested using firms’ intellectual capital as a proxy for big data in evaluating its impact on organizational performance. The present study employed the Value Added Intellectual Coefficient method to measure firm intellectual capital, via its three main components: human capital efficiency, structural capital efficiency, and capital employed efficiency, and then used the structural equation modeling technique to model the data and test the models. The financial fundamental and market data of 100 randomly selected publicly listed firms were collected. The results of the tests showed that only human capital efficiency had a significant positive impact on firm profitability, which highlighted the prominent human role in the impact of big data technology.

Keywords: big data, big data analytics, intellectual capital, organizational performance, value added intellectual coefficient

Procedia PDF Downloads 216
24269 Automated Test Data Generation For some types of Algorithm

Authors: Hitesh Tahbildar

Abstract:

The cost of test data generation for a program is computationally very high. In general case, no algorithm to generate test data for all types of algorithms has been found. The cost of generating test data for different types of algorithm is different. Till date, people are emphasizing the need to generate test data for different types of programming constructs rather than different types of algorithms. The test data generation methods have been implemented to find heuristics for different types of algorithms. Some algorithms that includes divide and conquer, backtracking, greedy approach, dynamic programming to find the minimum cost of test data generation have been tested. Our experimental results say that some of these types of algorithm can be used as a necessary condition for selecting heuristics and programming constructs are sufficient condition for selecting our heuristics. Finally we recommend the different heuristics for test data generation to be selected for different types of algorithms.

Keywords: ongest path, saturation point, lmax, kL, kS

Procedia PDF Downloads 380
24268 Haemodynamics Study in Subject Specific Carotid Bifurcation Using FSI

Authors: S. M. Abdul Khader, Anurag Ayachit, Raghuvir Pai, K. A. Ahmed, V. R. K Rao, S. Ganesh Kamath

Abstract:

The numerical simulation has made tremendous advances in investigating the blood flow phenomenon through elastic arteries. Such study can be useful in demonstrating the disease progression and haemodynamics of cardiovascular diseases such as atherosclerosis. In the present study, patient specific case diagnosed with partially stenosed complete right ICA and normal left carotid bifurcation without any atherosclerotic plaque formation is considered. 3D patient specific carotid bifurcation model is generated based on CT scan data using MIMICS-4.0 and numerical analysis is performed using FSI solver in ANSYS-14.5. The blood flow is assumed to be incompressible, homogenous and Newtonian, while the artery wall is assumed to be linearly elastic. The two-way sequentially-coupled transient FSI analysis is performed using FSI solver for three pulse cycles. The haemodynamic parameters such as flow pattern, Wall Shear Stress, pressure contours and arterial wall deformation are studied at the bifurcation and critical zones such as stenosis. The variation in flow behavior is studied throughout the pulse cycle. Also, the simulation results reveals that there is a considerable increase in the flow behavior in stenosed carotid in contrast to the normal carotid bifurcation system. The investigation also demonstrates the disturbed flow pattern especially at the bifurcation and stenosed zone elevating the haemodynamics, particularly during peak systole and later part of the pulse cycle. The results obtained agree well with the clinical observation and demonstrates the potential of patient specific numerical studies in prognosis of disease progression and plaque rupture.

Keywords: fluid-structure interaction, arterial stenosis, wall shear stress, carotid artery bifurcation

Procedia PDF Downloads 554
24267 Hybrid Quasi-Steady Thermal Lattice Boltzmann Model for Studying the Behavior of Oil in Water Emulsions Used in Machining Tool Cooling and Lubrication

Authors: W. Hasan, H. Farhat, A. Alhilo, L. Tamimi

Abstract:

Oil in water (O/W) emulsions are utilized extensively for cooling and lubricating cutting tools during parts machining. A robust Lattice Boltzmann (LBM) thermal-surfactants model, which provides a useful platform for exploring complex emulsions’ characteristics under variety of flow conditions, is used here for the study of the fluid behavior during conventional tools cooling. The transient thermal capabilities of the model are employed for simulating the effects of the flow conditions of O/W emulsions on the cooling of cutting tools. The model results show that the temperature outcome is slightly affected by reversing the direction of upper plate (workpiece). On the other hand, an important increase in effective viscosity is seen which supports better lubrication during the work.

Keywords: hybrid lattice Boltzmann method, Gunstensen model, thermal, surfactant-covered droplet, Marangoni stress

Procedia PDF Downloads 281
24266 The Perspective on Data Collection Instruments for Younger Learners

Authors: Hatice Kübra Koç

Abstract:

For academia, collecting reliable and valid data is one of the most significant issues for researchers. However, it is not the same procedure for all different target groups; meanwhile, during data collection from teenagers, young adults, or adults, researchers can use common data collection tools such as questionnaires, interviews, and semi-structured interviews; yet, for young learners and very young ones, these reliable and valid data collection tools cannot be easily designed or applied by the researchers. In this study, firstly, common data collection tools are examined for ‘very young’ and ‘young learners’ participant groups since it is thought that the quality and efficiency of an academic study is mainly based on its valid and correct data collection and data analysis procedure. Secondly, two different data collection instruments for very young and young learners are stated as discussing the efficacy of them. Finally, a suggested data collection tool – a performance-based questionnaire- which is specifically developed for ‘very young’ and ‘young learners’ participant groups in the field of teaching English to young learners as a foreign language is presented in this current study. The designing procedure and suggested items/factors for the suggested data collection tool are accordingly revealed at the end of the study to help researchers have studied with young and very learners.

Keywords: data collection instruments, performance-based questionnaire, young learners, very young learners

Procedia PDF Downloads 58
24265 Perceptions on Community Media for Effective Acculturation in Nigerian Indigenous Languages

Authors: Chima Onwukwe

Abstract:

This study examined perceptions on the effectiveness, attendant challenges and remedies of community media for effective acculturation in Nigerian languages. The qualitative survey design was adopted with Focus Group Discussions (FGD) and Key Informant Interviews (KIIs) of 50 purposively chosen informants. It was perceived that community media could serve as veritable platform for effective acculturation in Nigerian languages since they would engender the setting of acculturation in Nigerian languages as national objective or goal. It was further held that the strengths of community media for acculturation were in being goal-defined, ensuring local content and diversification. The study identified that as palatable as the proposal for community media for effective acculturation in Nigerian languages is; it would be fraught with some set-backs or challenges that were very much surmountable. Perceptions pointed towards transient nature of community media and funding as challenges, as well as multi-based funding as one remedy. Immediate establishment of community media for the purpose of acculturation in Nigerian languages was recommended.

Keywords: perception, community media, acculturation, indigenous language

Procedia PDF Downloads 247