Search results for: leadership models
2972 Design and Optimization of Flow Field for Cavitation Reduction of Valve Sleeves
Authors: Kamal Upadhyay, Zhou Hua, Yu Rui
Abstract:
This paper aims to improve the streamline linked with the flow field and cavitation on the valve sleeve. We observed that local pressure fluctuation produces a low-pressure zone, central to the formation of vapor volume fraction within the valve chamber led to air-bubbles (or cavities). Thus, it allows simultaneously to a severe negative impact on the inner surface and lifespan of the valve sleeves. Cavitation reduction is a vitally important issue to pressure control valves. The optimization of the flow field is proposed in this paper to reduce the cavitation of valve sleeves. In this method, the inner wall of the valve sleeve is changed from a cylindrical surface to the conical surface, leading to the decline of the fluid flow velocity and the rise of the outlet pressure. Besides, the streamline is distributed inside the sleeve uniformly. Thus, the bubble generation is lessened. The fluid models are built and analysis of flow field distribution, pressure, vapor volume and velocity was carried out using computational fluid dynamics (CFD) and numerical technique. The results indicate that this structure can suppress the cavitation of valve sleeves effectively.Keywords: streamline, cavitation, optimization, computational fluid dynamics
Procedia PDF Downloads 1492971 Insight Into Database Forensics
Authors: Enas K., Fatimah A., Abeer A., Ghadah A.
Abstract:
Database forensics is a specialized field of digital forensics that investigates and analyzes database systems to recover and evaluate data, particularly in cases of cyberattacks and data breaches. The increasing significance of securing data confidentiality, integrity, and availability has emphasized the need for robust forensic models to preserve data integrity and maintain the chain of evidence. Organizations rely on Database Forensic Investigation (DBFI) to protect critical data, maintain trust, and support legal actions in the event of breaches. To address the complexities of relational and non-relational databases, structured forensic frameworks and tools have been developed. These include the Three-Tier Database Forensic Model (TT-DF) for comprehensive investigations, blockchain-backed logging systems for enhanced evidence reliability, and the FORC tool for mobile SQLite database forensics. Such advancements facilitate data recovery, identify unauthorized access, and reconstruct events for legal proceedings. Practical demonstrations of these tools and frameworks further illustrate their real-world applicability, advancing the effectiveness of database forensics in mitigating modern cybersecurity threats.Keywords: database forensics, cybersecurity, SQLite forensics, digital forensics
Procedia PDF Downloads 32970 Theoretical Analysis of Self-Starting Busemann Intake Family
Authors: N. Moradian, E. Timofeev, R. Tahir
Abstract:
In this work, startability of the Busemann intake family with weak/strong conical shock, as most efficient intakes, via overboard mass spillage method is theoretically analyzed. Masterix and Candifix codes are used to numerically simulate few models of this type of intake and verify the theoretical results. Portions of the intake corresponding to various flow capture angles are considered to have mass spillage in the starting process of this intake. This approach allows for overboard mass spillage via a V-shaped slot with the tip of V coinciding with the focal point of the Busemann flow. The theoretical results, achieved using two different theories, of self-started Busemann takes with weak/strong conical shock show that significant improve in intake startability using overboard spillage technique. The starting phenomena of Busemann intakes with weak conical shock and seven different capture angles are numerically simulated at freestream Mach number of 3 to find the minimum area ratios of self-started intakes. The numerical results confirm the theoretical ones achieved by authors.Keywords: Busemann intake, conical shock, overboard spillage, startability
Procedia PDF Downloads 2072969 Three Dimensional Flexible Dynamics of Continuous Cislunar Payloads Transfer System
Authors: Y. Yang, Dian Ming Xing, Qiu Hua Du
Abstract:
Based on the Motorized Momentum Exchange Tether (MMET), with the principle of momentum exchange, the three dimension flexible dynamics of continuous cislunar payloads transferring system (CCPTS) is built by Lagrange method and its numerical solution is solved by Mathematica software. In the derivation precession of potential energy, this paper uses the Tylor expansion method to simplify the Lagrange equation. Furthermore, the tension coming from the centripetal load is considered in the elastic potential energy. The comparison simulation results between the 3D rigid model and 3D flexible model of CCPTS shows that the tether flexibility has important influence on CCPTS’s orbital parameters (such as radius of CCPTS’s COM and the true anomaly) and the tether’s rotational movement, the relative deviation of radius and the true anomaly between the two dynamic models is about 0.00678% and 0.00259%, the relative deviation of the angle of tether-span and local gravity gradient is about 3.55%. Additionally, the external torque has an apparent influence on the tether’s axial vibration.Keywords: cislunar transfer, dynamics, momentum exchange, tether
Procedia PDF Downloads 2692968 Global Best Practice Paradox; the Failure of One Size Fits All Approach to Development a Case Study of Pakistan
Authors: Muhammad Naveed Iftikhar, Farah Khalid
Abstract:
Global best practices as ordained by international organizations comprise a broader top-down approach to development problems, without taking into account country-specific factors. The political economy of each country is extremely different and the failure of several attempts of international organizations to implement global best practice models in developing countries each with its unique set of variables, goes on to show that this is not the most efficient solution to development problems. This paper is a humble attempt at shedding light on some specific examples of failures of the global best practices. Pakistan has its unique set of problems and unless those are added to the broader equation of development, country-specific reform and growth will continue to pose a challenge to reform programs initiated by international organizations. The three case studies presented in this paper are just a few prominent examples of failure of the global best practice, top-down, universalistic approach to development as ordained by international organizations. Development and reform can only be achieved if local dynamics are given their due importance. The modus operandi of international organizations needs to be tailored according to each country’s unique politico-economic environment.Keywords: best practice, development, context
Procedia PDF Downloads 4742967 Universal Design Building Standard for India: A Critical Inquiry
Authors: Sushil Kumar Solanki, Rachna Khare
Abstract:
Universal Design is a concept of built environment creation, where all people are facilitated to the maximum extent possible without using any type of specialized design. However, accessible design is a design process in which the needs of people with disabilities are specifically considered. Building standards on accessibility contains scoping and technical requirements for accessibility to sites, facilities, building and elements by individual with disability. India is also following its prescriptive types of various building standards for the creation of physical environment for people with disabilities. These building standards are based on western models instead of research based standards to serve Indian needs. These standards lack contextual connect when reflects in its application in the urban and rural environment. This study focuses on critical and comparative study of various international building standards and codes, with existing Indian accessibility standards to understand problems and prospects of concept of Universal Design building standards for India. The result of this study is an analysis of existing state of Indian building standard pertaining to accessibility and future need of performance based Universal Design concept.Keywords: accessibility, building standard, built-environment, universal design
Procedia PDF Downloads 2962966 Intuitive Decision Making When Facing Risks
Authors: Katharina Fellnhofer
Abstract:
The more information and knowledge that technology provides, the more important are profoundly human skills like intuition, the skill of using nonconscious information. As our world becomes more complex, shaken by crises, and characterized by uncertainty, time pressure, ambiguity, and rapidly changing conditions, intuition is increasingly recognized as a key human asset. However, due to methodological limitations of sample size or time frame or a lack of real-world or cross-cultural scope, precisely how to measure intuition when facing risks on a nonconscious level remains unclear. In light of the measurement challenge related to intuition’s nonconscious nature, a technique is introduced to measure intuition via hidden images as nonconscious additional information to trigger intuition. This technique has been tested in a within-subject fully online design with 62,721 real-world investment decisions made by 657 subjects in Europe and the United States. Bayesian models highlight the technique’s potential to measure skill at using nonconscious information for conscious decision making. Over the long term, solving the mysteries of intuition and mastering its use could be of immense value in personal and organizational decision-making contexts.Keywords: cognition, intuition, investment decisions, methodology
Procedia PDF Downloads 882965 Integrating Deep Learning For Improved State Of Charge Estimation In Electric Bus
Authors: Ms. Hema Ramachandran, Dr. N. Vasudevan
Abstract:
Accurate estimation of the battery State of Charge (SOC) is essential for optimizing the range and performance of modern electric vehicles. This paper focuses on analysing historical driving data from electric buses, with an emphasis on feature extraction and data preprocessing of driving conditions. By selecting relevant parameters, a set of characteristic variables tailored to specific driving scenarios is established. A battery SOC prediction model based on a combination a bidirectional long short-term memory (LSTM) architecture and a standard fully connected neural network (FCNN) is then proposed, where various hyperparameters are identified and fine-tuned to enhance prediction accuracy. The results indicate that with optimized hyperparameters, the prediction achieves a Root Mean Square Error (RMSE) of 1.98% and a Mean Absolute Error (MAE) of 1.72%. This approach is expected to improve the efficiency of battery management systems and battery utilization in electric vehicles.Keywords: long short-term memory (lstm), battery health monitoring, data-driven models, battery charge-discharge cycles, adaptive soc estimation, voltage and current sensing
Procedia PDF Downloads 92964 Motor Controller Implementation Using Model Based Design
Authors: Cau Tran, Tu Nguyen, Tien Pham
Abstract:
Model-based design (MBD) is a mathematical and visual technique for addressing design issues in the fields of communications, signal processing, and complicated control systems. It is utilized in several automotive, aerospace, industrial, and motion control applications. Virtual models are at the center of the software development process with model based design. A method used in the creation of embedded software is model-based design. In this study, the LAT motor is modeled in a simulation environment, and the LAT motor control is designed with a cascade structure, a speed and current control loop, and a controller that is used in the next part. A PID structure serves as this controller. Based on techniques and motor parameters that match the design goals, the PID controller is created for the model using traditional design principles. The MBD approach will be used to build embedded software for motor control. The paper will be divided into three distinct sections. The first section will introduce the design process and the benefits and drawbacks of the MBD technique. The design of control software for LAT motors will be the main topic of the next section. The experiment's results are the subject of the last section.Keywords: model based design, limited angle torque, intellectual property core, hardware description language, controller area network, user datagram protocol
Procedia PDF Downloads 962963 Review and Classification of the Indicators and Trends Used in Bridge Performance Modeling
Authors: S. Rezaei, Z. Mirzaei, M. Khalighi, J. Bahrami
Abstract:
Bridges, as an essential part of road infrastructures, are affected by various deterioration mechanisms over time due to the changes in their performance. As changes in performance can have many negative impacts on society, it is essential to be able to evaluate and measure the performance of bridges throughout their life. This evaluation includes the development or the choice of the appropriate performance indicators, which, in turn, are measured based on the selection of appropriate models for the existing deterioration mechanism. The purpose of this article is a statistical study of indicators and deterioration mechanisms of bridges in order to discover further research capacities in bridges performance assessment. For this purpose, some of the most common indicators of bridge performance, including reliability, risk, vulnerability, robustness, and resilience, were selected. The researches performed on each index based on the desired deterioration mechanisms and hazards were comprehensively reviewed. In addition, the formulation of the indicators and their relationship with each other were studied. The research conducted on the mentioned indicators were classified from the point of view of deterministic or probabilistic method, the level of study (element level, object level, etc.), and the type of hazard and the deterioration mechanism of interest. For each of the indicators, a number of challenges and recommendations were presented according to the review of previous studies.Keywords: bridge, deterioration mechanism, lifecycle, performance indicator
Procedia PDF Downloads 1062962 Recurrent Patterns of Netspeak among Selected Nigerians on WhatsApp Platform: A Quest for Standardisation
Authors: Lily Chimuanya, Esther Ajiboye, Emmanuel Uba
Abstract:
One of the consequences of online communication is the birth of new orthography genres characterised by novel conventions of abbreviation and acronyms usually referred to as Netspeak. Netspeak, also known as internet slang, is a style of writing mainly used in online communication to limit the length of text characters and to save time. The aim of this study is to evaluate how second language users of the English language have internalised this new convention of writing; identify the recurrent patterns of Netspeak; and assess the consistency of the use of the identified patterns in relation to their meanings. The study is corpus-based, and data drawn from WhatsApp chart pages of selected groups of Nigerian English speakers show a large occurrence of inconsistencies in the patterns of Netspeak and their meanings. The study argues that rather than emphasise the negative impact of Netspeak on the communicative competence of second language users, studies should focus on suggesting models as yardsticks for standardising the usage of Netspeak and indeed all other emerging language conventions resulting from online communication. This stance stems from the inevitable global language transformation that is eminent with the coming of age of information technology.Keywords: abbreviation, acronyms, Netspeak, online communication, standardisation
Procedia PDF Downloads 3932961 Determination of Chemical and Adsorption Kinetics: An Investigation of a Petrochemical Wastewater Treatment Utilizing GAC
Authors: Leila Vafajoo, Feria Ghanaat, Alireza Mohmadi Kartalaei, Amin Ghalebi
Abstract:
Petrochemical industries are playing an important role in producing wastewaters. Nowadays different methods are employed to treat these materials. The goal of the present research was to reduce the COD of a petrochemical wastewater via adsorption technique using a commercial granular activated carbon (GAC) as adsorbent. In the current study, parameters of kinetic models as well as; adsorption isotherms were determined through utilizing the Langmuir and Freundlich isotherms. The key parameters of KL= 0.0009 and qm= 33.33 for the former and nf=0.5 and Kf= 0.000004 for the latter isotherms resulted. Moreover, a correlation coefficient of above 90% for both cases proved logical use of such isotherms. On the other hand, pseudo-first and -second order kinetics equations were implemented. These resulted in coefficients of k1=0.005 and qe=2018 as well as; K2=0.009 and qe=1250; respectively. In addition, obtaining the correlation coefficients of 0.94 and 0.68 for these 1st and 2nd order kinetics; respectively indicated advantageous use of the former model. Furthermore, a significant experimental reduction of the petrochemical wastewater COD revealed that, using GAC for the process undertaken was an efficient mean of treatment. Ultimately, the current investigation paved down the road for predicting the system’s behavior on industrial scale.Keywords: petrochemical wastewater, adsorption, granular activated carbon, equilibrium isotherm, kinetic model
Procedia PDF Downloads 3642960 Indoor Fingerprint Localization Using 5G NR Multi-SSB Beam Features with GAN-Based Interpolation
Authors: LiRen Kang, LingXia Li, KaiKai Liu, Yue Jin, ZengShan Tian
Abstract:
With the widespread adoption of 5G technology in the Internet of Things (IoT), indoor localization methods based on 5G signals have gradually become a research hotspot. However, traditional methods often perform poorly in multipath interference and signal attenuation environments. To address these challenges, this paper proposes an innovative fingerprint localization method that utilizes the multiple synchronization signal block (SSB) beam features of 5G signals combined with generative adversarial networks (GANs) for interpolation. Our method incorporates a ray tracing model as an auxiliary, integrating signal propagation models to enhance the interpolation process. We precisely extract the multiple SSB beam features from 5G signals; in the localization stage, deep learning neural networks (DNN) are used for localization. Field tests show that localization errors of less than 1.5 meters can be achieved within about 200 square meters of indoor environment. Our method represents a 56.7% improvement compared to traditional methods that use received signal strength (RSS) as a single feature.Keywords: 5G NR, fingerprint localization, generative adversarial networks, Internet of Things, indoor localization systems
Procedia PDF Downloads 82959 Compilation of Tall Building with Green Architecture Case Study: Babolsar City (North of Iran) at 2014-2015
Authors: Seyyed Hossein Alavi, Soudabeh Mehri Talarposhti
Abstract:
Quick development of urban population need for housing on the one hand and prevention of irregular urban extension for optimum usage of urban land, resolving problems of urban physiognomy, land using, and environmental issues and urban transport, on the other hand, proposed tall building as urban area extension requirement in developing and advanced countries. Beside the tall building, protection, and creation of green architecture is one the most important issues of today's architecture world. This research is about attending tall building with green architecture in Babolsar city 2015. For this, the issues that can make favorite conditions for green architecture has been discussed. The purpose of this discussion is skeleton extension and accessing interactions between architecture and related technologies. This discussion with using of qualitative research methods (Analytical Description) tried to studying designed performance models and also studying and analyzing the inside and foreign articles and books. Hope this research is useful in solving the existing problems in this issue.Keywords: tall building, green architecture, skeleton extension, Babolsar city
Procedia PDF Downloads 4322958 Measurement and Modelling of HIV Epidemic among High Risk Groups and Migrants in Two Districts of Maharashtra, India: An Application of Forecasting Software-Spectrum
Authors: Sukhvinder Kaur, Ashok Agarwal
Abstract:
Background: For the first time in 2009, India was able to generate estimates of HIV incidence (the number of new HIV infections per year). Analysis of epidemic projections helped in revealing that the number of new annual HIV infections in India had declined by more than 50% during the last decade (GOI Ministry of Health and Family Welfare, 2010). Then, National AIDS Control Organisation (NACO) planned to scale up its efforts in generating projections through epidemiological analysis and modelling by taking recent available sources of evidence such as HIV Sentinel Surveillance (HSS), India Census data and other critical data sets. Recently, NACO generated current round of HIV estimates-2012 through globally recommended tool “Spectrum Software” and came out with the estimates for adult HIV prevalence, annual new infections, number of people living with HIV, AIDS-related deaths and treatment needs. State level prevalence and incidence projections produced were used to project consequences of the epidemic in spectrum. In presence of HIV estimates generated at state level in India by NACO, USIAD funded PIPPSE project under the leadership of NACO undertook the estimations and projections to district level using same Spectrum software. In 2011, adult HIV prevalence in one of the high prevalent States, Maharashtra was 0.42% ahead of the national average of 0.27%. Considering the heterogeneity of HIV epidemic between districts, two districts of Maharashtra – Thane and Mumbai were selected to estimate and project the number of People-Living-with-HIV/AIDS (PLHIV), HIV-prevalence among adults and annual new HIV infections till 2017. Methodology: Inputs in spectrum included demographic data from Census of India since 1980 and sample registration system, programmatic data on ‘Alive and on ART (adult and children)’,‘Mother-Baby pairs under PPTCT’ and ‘High Risk Group (HRG)-size mapping estimates’, surveillance data from various rounds of HSS, National Family Health Survey–III, Integrated Biological and Behavioural Assessment and Behavioural Sentinel Surveillance. Major Findings: Assuming current programmatic interventions in these districts, an estimated decrease of 12% points in Thane and 31% points in Mumbai among new infections in HRGs and migrants is observed from 2011 by 2017. Conclusions: Project also validated decrease in HIV new infection among one of the high risk groups-FSWs using program cohort data since 2012 to 2016. Though there is a decrease in HIV prevalence and new infections in Thane and Mumbai, further decrease is possible if appropriate programme response, strategies and interventions are envisaged for specific target groups based on this evidence. Moreover, evidence need to be validated by other estimation/modelling techniques; and evidence can be generated for other districts of the state, where HIV prevalence is high and reliable data sources are available, to understand the epidemic within the local context.Keywords: HIV sentinel surveillance, high risk groups, projections, new infections
Procedia PDF Downloads 2132957 Predictive Analysis for Big Data: Extension of Classification and Regression Trees Algorithm
Authors: Ameur Abdelkader, Abed Bouarfa Hafida
Abstract:
Since its inception, predictive analysis has revolutionized the IT industry through its robustness and decision-making facilities. It involves the application of a set of data processing techniques and algorithms in order to create predictive models. Its principle is based on finding relationships between explanatory variables and the predicted variables. Past occurrences are exploited to predict and to derive the unknown outcome. With the advent of big data, many studies have suggested the use of predictive analytics in order to process and analyze big data. Nevertheless, they have been curbed by the limits of classical methods of predictive analysis in case of a large amount of data. In fact, because of their volumes, their nature (semi or unstructured) and their variety, it is impossible to analyze efficiently big data via classical methods of predictive analysis. The authors attribute this weakness to the fact that predictive analysis algorithms do not allow the parallelization and distribution of calculation. In this paper, we propose to extend the predictive analysis algorithm, Classification And Regression Trees (CART), in order to adapt it for big data analysis. The major changes of this algorithm are presented and then a version of the extended algorithm is defined in order to make it applicable for a huge quantity of data.Keywords: predictive analysis, big data, predictive analysis algorithms, CART algorithm
Procedia PDF Downloads 1422956 Ensemble Methods in Machine Learning: An Algorithmic Approach to Derive Distinctive Behaviors of Criminal Activity Applied to the Poaching Domain
Authors: Zachary Blanks, Solomon Sonya
Abstract:
Poaching presents a serious threat to endangered animal species, environment conservations, and human life. Additionally, some poaching activity has even been linked to supplying funds to support terrorist networks elsewhere around the world. Consequently, agencies dedicated to protecting wildlife habitats have a near intractable task of adequately patrolling an entire area (spanning several thousand kilometers) given limited resources, funds, and personnel at their disposal. Thus, agencies need predictive tools that are both high-performing and easily implementable by the user to help in learning how the significant features (e.g. animal population densities, topography, behavior patterns of the criminals within the area, etc) interact with each other in hopes of abating poaching. This research develops a classification model using machine learning algorithms to aid in forecasting future attacks that is both easy to train and performs well when compared to other models. In this research, we demonstrate how data imputation methods (specifically predictive mean matching, gradient boosting, and random forest multiple imputation) can be applied to analyze data and create significant predictions across a varied data set. Specifically, we apply these methods to improve the accuracy of adopted prediction models (Logistic Regression, Support Vector Machine, etc). Finally, we assess the performance of the model and the accuracy of our data imputation methods by learning on a real-world data set constituting four years of imputed data and testing on one year of non-imputed data. This paper provides three main contributions. First, we extend work done by the Teamcore and CREATE (Center for Risk and Economic Analysis of Terrorism Events) research group at the University of Southern California (USC) working in conjunction with the Department of Homeland Security to apply game theory and machine learning algorithms to develop more efficient ways of reducing poaching. This research introduces ensemble methods (Random Forests and Stochastic Gradient Boosting) and applies it to real-world poaching data gathered from the Ugandan rain forest park rangers. Next, we consider the effect of data imputation on both the performance of various algorithms and the general accuracy of the method itself when applied to a dependent variable where a large number of observations are missing. Third, we provide an alternate approach to predict the probability of observing poaching both by season and by month. The results from this research are very promising. We conclude that by using Stochastic Gradient Boosting to predict observations for non-commercial poaching by season, we are able to produce statistically equivalent results while being orders of magnitude faster in computation time and complexity. Additionally, when predicting potential poaching incidents by individual month vice entire seasons, boosting techniques produce a mean area under the curve increase of approximately 3% relative to previous prediction schedules by entire seasons.Keywords: ensemble methods, imputation, machine learning, random forests, statistical analysis, stochastic gradient boosting, wildlife protection
Procedia PDF Downloads 2942955 Applied Bayesian Regularized Artificial Neural Network for Up-Scaling Wind Speed Profile and Distribution
Authors: Aghbalou Nihad, Charki Abderafi, Saida Rahali, Reklaoui Kamal
Abstract:
Maximize the benefit from the wind energy potential is the most interest of the wind power stakeholders. As a result, the wind tower size is radically increasing. Nevertheless, choosing an appropriate wind turbine for a selected site require an accurate estimate of vertical wind profile. It is also imperative from cost and maintenance strategy point of view. Then, installing tall towers or even more expensive devices such as LIDAR or SODAR raises the costs of a wind power project. Various models were developed coming within this framework. However, they suffer from complexity, generalization and lacks accuracy. In this work, we aim to investigate the ability of neural network trained using the Bayesian Regularization technique to estimate wind speed profile up to height of 100 m based on knowledge of wind speed lower heights. Results show that the proposed approach can achieve satisfactory predictions and proof the suitability of the proposed method for generating wind speed profile and probability distributions based on knowledge of wind speed at lower heights.Keywords: bayesian regularization, neural network, wind shear, accuracy
Procedia PDF Downloads 5042954 A Study on Mesh Size Dependency on Bed Expansion Zone in a Three-Phase Fluidized Bed Reactor
Authors: Liliana Patricia Olivo Arias
Abstract:
The present study focused on the hydrodynamic study in a three-phase fluidized bed reactor and the influence of important aspects, such as volume fractions (Hold up), velocity magnitude of gas, liquid and solid phases (hydrogen, gasoil, and gamma alumina), interactions of phases, through of drag models with the k-epsilon turbulence model. For this purpose was employed a Euler-Euler model and also considers the system is constituted of three phases, gaseous, liquid and solid, characterized by its physical and thermal properties, the transport processes that are developed within the transient regime. The proposed model of the three-phase fluidized bed reactor was solved numerically using the ANSYS-Fluent software with different mesh refinements on bed expansion zone in order to observe the influence of the hydrodynamic parameters and convergence criteria. With this model and the numerical simulations obtained for its resolution, it was possible to predict the results of the volume fractions (Hold ups) and the velocity magnitude for an unsteady system from the initial and boundaries conditions were established.Keywords: three-phase fluidized bed system, CFD simulation, mesh dependency study, hydrodynamic study
Procedia PDF Downloads 1672953 Novel Recommender Systems Using Hybrid CF and Social Network Information
Authors: Kyoung-Jae Kim
Abstract:
Collaborative Filtering (CF) is a popular technique for the personalization in the E-commerce domain to reduce information overload. In general, CF provides recommending items list based on other similar users’ preferences from the user-item matrix and predicts the focal user’s preference for particular items by using them. Many recommender systems in real-world use CF techniques because it’s excellent accuracy and robustness. However, it has some limitations including sparsity problems and complex dimensionality in a user-item matrix. In addition, traditional CF does not consider the emotional interaction between users. In this study, we propose recommender systems using social network and singular value decomposition (SVD) to alleviate some limitations. The purpose of this study is to reduce the dimensionality of data set using SVD and to improve the performance of CF by using emotional information from social network data of the focal user. In this study, we test the usability of hybrid CF, SVD and social network information model using the real-world data. The experimental results show that the proposed model outperforms conventional CF models.Keywords: recommender systems, collaborative filtering, social network information, singular value decomposition
Procedia PDF Downloads 2942952 Critical Review of Oceanic and Geological Storage of Carbon Sequestration
Authors: Milad Nooshadi, Alessandro Manzardo
Abstract:
CO₂ emissions in the atmosphere continue to rise, mostly as a result of the combustion of fossil fuels. CO₂ injection into the oceans and geological formation as a process of physical carbon capture are two of the most promising emerging strategies for mitigating climate change and global warming. The purpose of this research is to evaluate the two mentioned methods of CO₂ sequestration and to assess information on previous and current advancements, limitations, and uncertainties associated with carbon sequestration in order to identify possible prospects for ensuring the timely implementation of the technology, such as determining how governments and companies can gain a better understanding of CO₂ storage in terms of which media have the most applicable capacity, which type of injection has the fewer environmental impact, and how much carbon sequestration and storage will cost. The behavior of several forms is characterized as a near field, a far field, and a see-floor in ocean storage, and three medias in geological formations as an oil and gas reservoir, a saline aquifer, and a coal bed. To determine the capacity of various forms of media, an analysis of some models and practical experiments are necessary. Additionally, as a major component of sequestration, the various injection methods into diverse media and their monitoring are associated with a variety of environmental impacts and financial consequences.Keywords: carbon sequestration, ocean storage, geologic storage, carbon transportation
Procedia PDF Downloads 1042951 Top Management Support as an Enabling Factor for Academic Innovation through Knowledge Sharing
Authors: Sawsan J. Al-husseini, Talib A. Dosa
Abstract:
Educational institutions are today facing increasing pressures due to economic, political and social upheaval. This is only exacerbated by the nature of education as an intangible good which relies upon the intellectual assets of the organisation, its staff. Top management support has been acknowledged as having a positive general influence on knowledge management and creativity. However, there is a lack of models linking top management support, knowledge sharing, and innovation within higher education institutions, in general within developing countries, and particularly in Iraq. This research sought to investigate the impact of top management support on innovation through the mediating role of knowledge sharing in Iraqi private HEIs. A quantitative approach was taken and 262 valid responses were collected to test the causal relationships between top management support, knowledge sharing, and innovation. Employing structural equation modelling with AMOS v.25, the research demonstrated that knowledge sharing plays a pivotal role in the relationship between top management support and innovation. The research has produced some guidelines for researchers as well as leaders, and provided evidence to support the use of knowledge sharing to increase innovation within the higher education environment in developing countries, particularly Iraq.Keywords: top management support, knowledge sharing, innovation, structural equation modelling
Procedia PDF Downloads 3272950 A Comparative Study of Generalized Autoregressive Conditional Heteroskedasticity (GARCH) and Extreme Value Theory (EVT) Model in Modeling Value-at-Risk (VaR)
Authors: Longqing Li
Abstract:
The paper addresses the inefficiency of the classical model in measuring the Value-at-Risk (VaR) using a normal distribution or a Student’s t distribution. Specifically, the paper focuses on the one day ahead Value-at-Risk (VaR) of major stock market’s daily returns in US, UK, China and Hong Kong in the most recent ten years under 95% confidence level. To improve the predictable power and search for the best performing model, the paper proposes using two leading alternatives, Extreme Value Theory (EVT) and a family of GARCH models, and compares the relative performance. The main contribution could be summarized in two aspects. First, the paper extends the GARCH family model by incorporating EGARCH and TGARCH to shed light on the difference between each in estimating one day ahead Value-at-Risk (VaR). Second, to account for the non-normality in the distribution of financial markets, the paper applies Generalized Error Distribution (GED), instead of the normal distribution, to govern the innovation term. A dynamic back-testing procedure is employed to assess the performance of each model, a family of GARCH and the conditional EVT. The conclusion is that Exponential GARCH yields the best estimate in out-of-sample one day ahead Value-at-Risk (VaR) forecasting. Moreover, the discrepancy of performance between the GARCH and the conditional EVT is indistinguishable.Keywords: Value-at-Risk, Extreme Value Theory, conditional EVT, backtesting
Procedia PDF Downloads 3242949 Market Segmentation and Conjoint Analysis for Apple Family Design
Authors: Abbas Al-Refaie, Nour Bata
Abstract:
A distributor of Apple products' experiences numerous difficulties in developing marketing strategies for new and existing mobile product entries that maximize customer satisfaction and the firm's profitability. This research, therefore, integrates market segmentation in platform-based product family design and conjoint analysis to identify iSystem combinations that increase customer satisfaction and business profits. First, the enhanced market segmentation grid is created. Then, the estimated demand model is formulated. Finally, the profit models are constructed then used to determine the ideal product family design that maximizes profit. Conjoint analysis is used to explore customer preferences with their satisfaction levels. A total of 200 surveys are collected about customer preferences. Then, simulation is used to determine the importance values for each attribute. Finally, sensitivity analysis is conducted to determine the product family design that maximizes both objectives. In conclusion, the results of this research shall provide great support to Apple distributors in determining the best marketing strategies that enhance their market share.Keywords: market segmentation, conjoint analysis, market strategies, optimization
Procedia PDF Downloads 3762948 Measurement Errors and Misclassifications in Covariates in Logistic Regression: Bayesian Adjustment of Main and Interaction Effects and the Sample Size Implications
Authors: Shahadut Hossain
Abstract:
Measurement errors in continuous covariates and/or misclassifications in categorical covariates are common in epidemiological studies. Regression analysis ignoring such mismeasurements seriously biases the estimated main and interaction effects of covariates on the outcome of interest. Thus, adjustments for such mismeasurements are necessary. In this research, we propose a Bayesian parametric framework for eliminating deleterious impacts of covariate mismeasurements in logistic regression. The proposed adjustment method is unified and thus can be applied to any generalized linear and non-linear regression models. Furthermore, adjustment for covariate mismeasurements requires validation data usually in the form of either gold standard measurements or replicates of the mismeasured covariates on a subset of the study population. Initial investigation shows that adequacy of such adjustment depends on the sizes of main and validation samples, especially when prevalences of the categorical covariates are low. Thus, we investigate the impact of main and validation sample sizes on the adjusted estimates, and provide a general guideline about these sample sizes based on simulation studies.Keywords: measurement errors, misclassification, mismeasurement, validation sample, Bayesian adjustment
Procedia PDF Downloads 4102947 The Adsorption of Perfluorooctanoic Acid on Coconut Shell Activated Carbons
Authors: Premrudee Kanchanapiya, Supachai Songngam, Thanapol Tantisattayakul
Abstract:
Perfluorooctanoic acid (PFOA) is one of per- and polyfluoroalkyl substances (PFAS) that have increasingly attracted concerns due to their global distribution in environment, persistence, high bioaccumulation, and toxicity. It is important to study the effective treatment to remove PFOA from contaminated water. The feasibility of using commercial coconut shell activated carbon produced in Thailand to remove PFOA from water was investigated with regard to their adsorption kinetics and isotherms of powder activated carbon (PAC-325) and granular activated carbon (GAC-20x50). Adsorption kinetic results show that the adsorbent size significantly affected the adsorption rate of PFOA, and GAC-20x50 required at least 100 h to achieve the equilibrium, much longer than 3 h for PAC-325. Two kinetic models were fitted to the experimental data, and the pseudo-second-order model well described the adsorption of PFOA on both PAC-325 and GAC-20x50. PAC-325 trended to adsorb PFOA faster than GAC-20x50, and testing with the shortest adsorption times (5 min) still yielded substantial PFOA removal (~80% for PAC-325). The adsorption isotherms show that the adsorption capacity of PAC-325 was 0.80 mmol/g, which is 83 % higher than that for GAC-20x50 (0.13 mmol/g), according to the Langmuir fitting.Keywords: perfluorooctanoic acid, PFOA, coconut shell activated carbons, adsorption, water treatment
Procedia PDF Downloads 1462946 The Hubs of Transformation Dictated by the Innovation Wave: Boston as a Case Study. Exploring How Design is Emerging as an Essential Feature in the Process of Laboratorisation of Cities
Authors: Luana Parisi, Sohrab Donyavi
Abstract:
Cities have become the nodes of global networks, standing at the intersection points of the flows of capital, goods, workers, businesses and travellers, making them the spots where innovation, progress and economic development occur. The primary challenge for them is to create the most fertile ecosystems for triggering innovation activities. Design emerges as an essential feature in this process of laboratorisation of cities. This paper aims at exploring the spatial hubs of transformation within the knowledge economy, providing an overview of the current models of innovation spaces, before focusing on the innovation district of one of the cities that are riding the innovation wave, namely, Boston, USA. Useful lessons will be drawn from the case study of the innovation district in Boston, allowing to define precious tools for policymakers, in the form of a range of factors that define the broad strategy able to implement the model successfully. A mixed methodology is implemented, including information from observations, exploratory interviews to key stakeholders and on-desk data.Keywords: Innovation District, innovation ecosystem, economic development, urban regeneration
Procedia PDF Downloads 1272945 A Comparison of Methods for Neural Network Aggregation
Authors: John Pomerat, Aviv Segev
Abstract:
Recently, deep learning has had many theoretical breakthroughs. For deep learning to be successful in the industry, however, there need to be practical algorithms capable of handling many real-world hiccups preventing the immediate application of a learning algorithm. Although AI promises to revolutionize the healthcare industry, getting access to patient data in order to train learning algorithms has not been easy. One proposed solution to this is data- sharing. In this paper, we propose an alternative protocol, based on multi-party computation, to train deep learning models while maintaining both the privacy and security of training data. We examine three methods of training neural networks in this way: Transfer learning, average ensemble learning, and series network learning. We compare these methods to the equivalent model obtained through data-sharing across two different experiments. Additionally, we address the security concerns of this protocol. While the motivating example is healthcare, our findings regarding multi-party computation of neural network training are purely theoretical and have use-cases outside the domain of healthcare.Keywords: neural network aggregation, multi-party computation, transfer learning, average ensemble learning
Procedia PDF Downloads 1632944 RANS Simulation of Viscous Flow around Hull of Multipurpose Amphibious Vehicle
Authors: M. Nakisa, A. Maimun, Yasser M. Ahmed, F. Behrouzi, A. Tarmizi
Abstract:
The practical application of the Computational Fluid Dynamics (CFD), for predicting the flow pattern around Multipurpose Amphibious Vehicle (MAV) hull has made much progress over the last decade. Today, several of the CFD tools play an important role in the land and water going vehicle hull form design. CFD has been used for analysis of MAV hull resistance, sea-keeping, maneuvering and investigating its variation when changing the hull form due to varying its parameters, which represents a very important task in the principal and final design stages. Resistance analysis based on CFD (Computational Fluid Dynamics) simulation has become a decisive factor in the development of new, economically efficient and environmentally friendly hull forms. Three-dimensional finite volume method (FVM) based on Reynolds Averaged Navier-Stokes equations (RANS) has been used to simulate incompressible flow around three types of MAV hull bow models in steady-state condition. Finally, the flow structure and streamlines, friction and pressure resistance and velocity contours of each type of hull bow will be compared and discussed.Keywords: RANS simulation, multipurpose amphibious vehicle, viscous flow structure, mechatronic
Procedia PDF Downloads 3132943 Cost Sensitive Feature Selection in Decision-Theoretic Rough Set Models for Customer Churn Prediction: The Case of Telecommunication Sector Customers
Authors: Emel Kızılkaya Aydogan, Mihrimah Ozmen, Yılmaz Delice
Abstract:
In recent days, there is a change and the ongoing development of the telecommunications sector in the global market. In this sector, churn analysis techniques are commonly used for analysing why some customers terminate their service subscriptions prematurely. In addition, customer churn is utmost significant in this sector since it causes to important business loss. Many companies make various researches in order to prevent losses while increasing customer loyalty. Although a large quantity of accumulated data is available in this sector, their usefulness is limited by data quality and relevance. In this paper, a cost-sensitive feature selection framework is developed aiming to obtain the feature reducts to predict customer churn. The framework is a cost based optional pre-processing stage to remove redundant features for churn management. In addition, this cost-based feature selection algorithm is applied in a telecommunication company in Turkey and the results obtained with this algorithm.Keywords: churn prediction, data mining, decision-theoretic rough set, feature selection
Procedia PDF Downloads 449