Search results for: adaptable business models
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 9328

Search results for: adaptable business models

5098 QCARNet: Networks for Quality-Adaptive Compression Artifact

Authors: Seung Ho Park, Young Su Moon, Nam Ik Cho

Abstract:

We propose a convolution neural network (CNN) for quality adaptive compression artifact reduction named QCARNet. The proposed method is different from the existing discriminative models that learn a specific model at a certain quality level. The method is composed of a quality estimation CNN (QECNN) and a compression artifact reduction CNN (CARCNN), which are two functionally separate CNNs. By connecting the QECNN and CARCNN, each CARCNN layer is able to adaptively reduce compression artifacts and preserve details depending on the estimated quality level map generated by the QECNN. We experimentally demonstrate that the proposed method achieves better performance compared to other state-of-the-art blind compression artifact reduction methods.

Keywords: compression artifact reduction, deblocking, image denoising, image restoration

Procedia PDF Downloads 121
5097 The Efficacy of Thymbra spicata Ethanolic Extract and its Main Component Carvacrol on In vitro Model of Metabolically-Associated Dysfunctions

Authors: Farah Diab, Mohamad Khalil, Francesca Storace, Francesca Baldini, Piero Portincasaa, Giulio Lupidi, Laura Vergani

Abstract:

Thymbra spicata is a thyme-like plant belonging to the Lamiaceae family that shows a global distribution, especially in the eastern Mediterranean region. Leaves of T. spicata contain large amounts of phenols such as phenolic acids (rosmarinic acid), phenolic monoterpenes (carvacrol), and flavonoids. In Lebanon, T. spicata is currently used as a culinary herb in salad and infusion, as well as for traditional medicinal purposes. Carvacrol (5-isopropyl-2-methyl phenol), the most abundant polyphenol in the organic extract and essential oils, has a great array of pharmacological properties. In fact, carvacrol is largely employed as a food additive and neutraceutical agent. Our aim is to investigate the beneficial effects of T. spicata ethanolic extract (TE) and its main component, carvacrol, using in vitro models of hepatic steatosis and endothelial dysfunction. As a further point, we focused on investigating if and how the binding of carvacrol to albumin, the physiological transporter for drugs in the blood, might be altered by the presence of high levels of fatty acids (FAs), thus impairing the carvacrol bio-distribution in vivo. For that reason, hepatic FaO cells treated with exogenous FAs such as oleate and palmitate mimic hepatosteatosis; endothelial HECV cells exposed to hydrogen peroxide are a model of endothelial dysfunction. In these models, we measured lipid accumulation, free radical production, lipoperoxidation, and nitric oxide release before and after treatment with carvacrol. The carvacrol binding to albumin with/without high levels of long-chain FAs was assessed by absorption and emission spectroscopies. Our findings show that both TE and carvacrol (i) counteracted lipid accumulation in hepatocytes by decreasing the intracellular and extracellular lipid contents in steatotic FaO cells; (ii) decreased oxidative stress in endothelial cells by significantly reducing lipoperoxidation and free radical production, as well as, attenuating the nitric oxide release; (ii) high levels of circulating FAs reduced the binding of carvacrol to albumin. The beneficial effects of TE and carvacrol on both hepatic and endothelial cells point to a nutraceutical potential. However, high levels of circulating FAs, such as those occurring in metabolic disorders, might hinder the carvacrol transport, bio-distribution, and pharmacodynamics.

Keywords: carvacrol, endothelial dysfunction, fatty acids, non-alcoholic fatty liver diseases, serum albumin

Procedia PDF Downloads 175
5096 AI-Enabled Smart Contracts for Reliable Traceability in the Industry 4.0

Authors: Harris Niavis, Dimitra Politaki

Abstract:

The manufacturing industry was collecting vast amounts of data for monitoring product quality thanks to the advances in the ICT sector and dedicated IoT infrastructure is deployed to track and trace the production line. However, industries have not yet managed to unleash the full potential of these data due to defective data collection methods and untrusted data storage and sharing. Blockchain is gaining increasing ground as a key technology enabler for Industry 4.0 and the smart manufacturing domain, as it enables the secure storage and exchange of data between stakeholders. On the other hand, AI techniques are more and more used to detect anomalies in batch and time-series data that enable the identification of unusual behaviors. The proposed scheme is based on smart contracts to enable automation and transparency in the data exchange, coupled with anomaly detection algorithms to enable reliable data ingestion in the system. Before sensor measurements are fed to the blockchain component and the smart contracts, the anomaly detection mechanism uniquely combines artificial intelligence models to effectively detect unusual values such as outliers and extreme deviations in data coming from them. Specifically, Autoregressive integrated moving average, Long short-term memory (LSTM) and Dense-based autoencoders, as well as Generative adversarial networks (GAN) models, are used to detect both point and collective anomalies. Towards the goal of preserving the privacy of industries' information, the smart contracts employ techniques to ensure that only anonymized pointers to the actual data are stored on the ledger while sensitive information remains off-chain. In the same spirit, blockchain technology guarantees the security of the data storage through strong cryptography as well as the integrity of the data through the decentralization of the network and the execution of the smart contracts by the majority of the blockchain network actors. The blockchain component of the Data Traceability Software is based on the Hyperledger Fabric framework, which lays the ground for the deployment of smart contracts and APIs to expose the functionality to the end-users. The results of this work demonstrate that such a system can increase the quality of the end-products and the trustworthiness of the monitoring process in the smart manufacturing domain. The proposed AI-enabled data traceability software can be employed by industries to accurately trace and verify records about quality through the entire production chain and take advantage of the multitude of monitoring records in their databases.

Keywords: blockchain, data quality, industry4.0, product quality

Procedia PDF Downloads 170
5095 Construction and Analysis of Samurai Sudoku

Authors: A. Danbaba

Abstract:

Samurai Sudoku consists of five Sudoku square designs each having nine treatments in each row (column or sub-block) only once such the five Sudoku designs overlaps. Two or more Samurai designs can be joint together to give an extended Samurai design. In addition, two Samurai designs, each containing five Sudoku square designs, are mutually orthogonal (Graeco). If we superimpose two Samurai designs and obtained a pair of Latin and Greek letters in each row (column or sub-block) of the five Sudoku designs only once, then we have Graeco Samurai design. In this paper, simple method of constructing Samurai designs and mutually orthogonal Samurai design are proposed. In addition, linear models and methods of data analysis for the designs are proposed.

Keywords: samurai design, graeco samurai design, sudoku design, row or column swap

Procedia PDF Downloads 258
5094 The Influence of the Concentration and Temperature on the Rheological Behavior of Carbonyl-Methylcellulose

Authors: Mohamed Rabhi, Kouider Halim Benrahou

Abstract:

The rheological properties of the carbonyl-methylcellulose (CMC), of different concentrations (25000, 50000, 60000, 80000 and 100000 ppm) and different temperatures were studied. We found that the rheological behavior of all CMC solutions presents a pseudo-plastic behavior, it follows the model of Ostwald-de Waele. The objective of this work is the modeling of flow by the CMC Cross model. The Cross model gives us the variation of the viscosity according to the shear rate. This model allowed us to adjust more clearly the rheological characteristics of CMC solutions. A comparison between the Cross model and the model of Ostwald was made. Cross the model fitting parameters were determined by a numerical simulation to make an approach between the experimental curve and those given by the two models. Our study has shown that the model of Cross, describes well the flow of "CMC" for low concentrations.

Keywords: CMC, rheological modeling, Ostwald model, cross model, viscosity

Procedia PDF Downloads 380
5093 A Survey on Taxpayer's Compliance in Prospect Theory Structure Using Hierarchical Bayesian Approach

Authors: Sahar Dehghan, Yeganeh Mousavi Jahromi, Ghahraman Abdoli

Abstract:

Since tax revenues are one of the most important sources of government revenue, it is essential to consider increasing taxpayers' compliance. One of the factors that can affect the taxpayers' compliance is the structure of the crimes and incentives envisaged in the tax law. In this research, by using the 'prospect theory', the effects of changes in the rate of crimes and the tax incentive in the direct tax law on the taxpayer’s compliance behavior have been investigated. To determine the preferences and preferences of taxpayer’s in the business sector and their degree of sensitivity to fines and incentives, a questionnaire with mixed gamble structure is designed. Estimated results using the Hierarchical Bayesian method indicate that the taxpayer’s that have been tested in this study are more sensitive to the incentives in the direct tax law, and the tax administration can use this to increase the level of collected tax and increase the level of compliance.

Keywords: tax compliance, prospect theory, value function, mixed gamble

Procedia PDF Downloads 157
5092 Impact of Reclamation on the Water Exchange in Bohai Bay

Authors: Luyao Liu, Dekui Yuan, Xu Li

Abstract:

As one of the most important bays of China, the water exchange capacity of Bohai Bay can influence the economic development and urbanization of surrounding cities. However, the rapid reclamation has influenced the weak water exchange capacity of this semi-enclosed bay in recent years. This paper sets two hydrodynamic models of Bohai Bay with two shorelines before and after reclamation. The mean value and distribution of Turn-over Time, the distribution of residual current, and the feature of the tracer path are compared. After comparison, it is found that Bohai Bay keeps these characteristics; the spending time of water exchange in the northern is longer than southern, and inshore is longer than offshore. However, the mean water exchange time becomes longer after reclamation. In addition, the material spreading is blocked because of the inwardly extending shorelines, and the direction changed from along the shoreline to towards the center after reclamation.

Keywords: Bohai Bay, water exchange, reclamation, turn-over time

Procedia PDF Downloads 125
5091 Comparison of MODIS-Based Rice Extent Map and Landsat-Based Rice Classification Map in Determining Biomass Energy Potential of Rice Hull in Nueva Ecija, Philippines

Authors: Klathea Sevilla, Marjorie Remolador, Bryan Baltazar, Imee Saladaga, Loureal Camille Inocencio, Ma. Rosario Concepcion Ang

Abstract:

The underutilization of biomass resources in the Philippines, combined with its growing population and the rise in fossil fuel prices confirms demand for alternative energy sources. The goal of this paper is to provide a comparison of MODIS-based and Landsat-based agricultural land cover maps when used in the estimation of rice hull’s available energy potential. Biomass resource assessment was done using mathematical models and remote sensing techniques employed in a GIS platform.

Keywords: biomass, geographic information system (GIS), remote sensing, renewable energy

Procedia PDF Downloads 471
5090 Predicting the Product Life Cycle of Songs on Radio - How Record Labels Can Manage Product Portfolio and Prioritise Artists by Using Machine Learning Techniques

Authors: Claus N. Holm, Oliver F. Grooss, Robert A. Alphinas

Abstract:

This research strives to predict the remaining product life cycle of a song on radio after it has been played for one or two months. The best results were achieved using a k-d tree to calculate the most similar songs to the test songs and use a Random Forest model to forecast radio plays. An 82.78% and 83.44% accuracy is achieved for the two time periods, respectively. This explorative research leads to over 4500 test metrics to find the best combination of models and pre-processing techniques. Other algorithms tested are KNN, MLP and CNN. The features only consist of daily radio plays and use no musical features.

Keywords: hit song science, product life cycle, machine learning, radio

Procedia PDF Downloads 142
5089 Islamic Extremist Groups' Usage of Populism in Social Media to Radicalize Muslim Migrants in Europe

Authors: Muhammad Irfan

Abstract:

The rise of radicalization within Islam has spawned a new era of global terror. The battlefield Successes of ISIS and the Taliban are fuelled by an ideological war waged, largely and successfully, in the media arena. This research will examine how Islamic extremist groups are using media modalities and populist narratives to influence migrant Muslim populations in Europe towards extremism. In 2014, ISIS shocked the world in exporting horrifically graphic forms of violence on social media. Their Muslim support base was largely disgusted and reviled. In response, they reconfigured their narrative by introducing populist 'hooks', astutely portraying the Muslim populous as oppressed and exploited by unjust, corrupt autocratic regimes and Western power structures. Within this crucible of real and perceived oppression, hundreds of thousands of the most desperate, vulnerable and abused migrants left their homelands, risking their lives in the hope of finding peace, justice, and prosperity in Europe. Instead, many encountered social stigmatization, detention and/or discrimination for being illegal migrants, for lacking resources and for simply being Muslim. This research will examine how Islamic extremist groups are exploiting the disenfranchisement of these migrant populations and using populist messaging on social media to influence them towards violent extremism. ISIS, in particular, formulates specific encoded messages for newly-arriving Muslims in Europe, preying upon their vulnerability. Violence is posited, as a populist response, to the tyranny of European oppression. This research will analyze the factors and indicators which propel Muslim migrants along the spectrum from resilience to violence extremism. Expected outcomes are identification of factors which influence vulnerability towards violent extremism; an early-warning detection framework; predictive analysis models; and de-radicalization frameworks. This research will provide valuable tools (practical and policy level) for European governments, security stakeholders, communities, policy-makers, and educators; it is anticipated to contribute to a de-escalation of Islamic extremism globally.

Keywords: populism, radicalization, de-radicalization, social media, ISIS, Taliban, shariah, jihad, Islam, Europe, political communication, terrorism, migrants, refugees, extremism, global terror, predictive analysis, early warning detection, models, strategic communication, populist narratives, Islamic extremism

Procedia PDF Downloads 111
5088 Shark Detection and Classification with Deep Learning

Authors: Jeremy Jenrette, Z. Y. C. Liu, Pranav Chimote, Edward Fox, Trevor Hastie, Francesco Ferretti

Abstract:

Suitable shark conservation depends on well-informed population assessments. Direct methods such as scientific surveys and fisheries monitoring are adequate for defining population statuses, but species-specific indices of abundance and distribution coming from these sources are rare for most shark species. We can rapidly fill these information gaps by boosting media-based remote monitoring efforts with machine learning and automation. We created a database of shark images by sourcing 24,546 images covering 219 species of sharks from the web application spark pulse and the social network Instagram. We used object detection to extract shark features and inflate this database to 53,345 images. We packaged object-detection and image classification models into a Shark Detector bundle. We developed the Shark Detector to recognize and classify sharks from videos and images using transfer learning and convolutional neural networks (CNNs). We applied these models to common data-generation approaches of sharks: boosting training datasets, processing baited remote camera footage and online videos, and data-mining Instagram. We examined the accuracy of each model and tested genus and species prediction correctness as a result of training data quantity. The Shark Detector located sharks in baited remote footage and YouTube videos with an average accuracy of 89\%, and classified located subjects to the species level with 69\% accuracy (n =\ eight species). The Shark Detector sorted heterogeneous datasets of images sourced from Instagram with 91\% accuracy and classified species with 70\% accuracy (n =\ 17 species). Data-mining Instagram can inflate training datasets and increase the Shark Detector’s accuracy as well as facilitate archiving of historical and novel shark observations. Base accuracy of genus prediction was 68\% across 25 genera. The average base accuracy of species prediction within each genus class was 85\%. The Shark Detector can classify 45 species. All data-generation methods were processed without manual interaction. As media-based remote monitoring strives to dominate methods for observing sharks in nature, we developed an open-source Shark Detector to facilitate common identification applications. Prediction accuracy of the software pipeline increases as more images are added to the training dataset. We provide public access to the software on our GitHub page.

Keywords: classification, data mining, Instagram, remote monitoring, sharks

Procedia PDF Downloads 102
5087 A Framework for Incorporating Non-Linear Degradation of Conductive Adhesive in Environmental Testing

Authors: Kedar Hardikar, Joe Varghese

Abstract:

Conductive adhesives have found wide-ranging applications in electronics industry ranging from fixing a defective conductor on printed circuit board (PCB) attaching an electronic component in an assembly to protecting electronics components by the formation of “Faraday Cage.” The reliability requirements for the conductive adhesive vary widely depending on the application and expected product lifetime. While the conductive adhesive is required to maintain the structural integrity, the electrical performance of the associated sub-assembly can be affected by the degradation of conductive adhesive. The degradation of the adhesive is dependent upon the highly varied use case. The conventional approach to assess the reliability of the sub-assembly involves subjecting it to the standard environmental test conditions such as high-temperature high humidity, thermal cycling, high-temperature exposure to name a few. In order to enable projection of test data and observed failures to predict field performance, systematic development of an acceleration factor between the test conditions and field conditions is crucial. Common acceleration factor models such as Arrhenius model are based on rate kinetics and typically rely on an assumption of linear degradation in time for a given condition and test duration. The application of interest in this work involves conductive adhesive used in an electronic circuit of a capacitive sensor. The degradation of conductive adhesive in high temperature and humidity environment is quantified by the capacitance values. Under such conditions, the use of established models such as Hallberg-Peck model or Eyring Model to predict time to failure in the field typically relies on linear degradation rate. In this particular case, it is seen that the degradation is nonlinear in time and exhibits a square root t dependence. It is also shown that for the mechanism of interest, the presence of moisture is essential, and the dominant mechanism driving the degradation is the diffusion of moisture. In this work, a framework is developed to incorporate nonlinear degradation of the conductive adhesive for the development of an acceleration factor. This method can be extended to applications where nonlinearity in degradation rate can be adequately characterized in tests. It is shown that depending on the expected product lifetime, the use of conventional linear degradation approach can overestimate or underestimate the field performance. This work provides guidelines for suitability of linear degradation approximation for such varied applications

Keywords: conductive adhesives, nonlinear degradation, physics of failure, acceleration factor model.

Procedia PDF Downloads 120
5086 Application of the Discrete Rationalized Haar Transform to Distributed Parameter System

Authors: Joon-Hoon Park

Abstract:

In this paper the rationalized Haar transform is applied for distributed parameter system identification and estimation. A distributed parameter system is a dynamical and mathematical model described by a partial differential equation. And system identification concerns the problem of determining mathematical models from observed data. The Haar function has some disadvantages of calculation because it contains irrational numbers, for these reasons the rationalized Haar function that has only rational numbers. The algorithm adopted in this paper is based on the transform and operational matrix of the rationalized Haar function. This approach provides more convenient and efficient computational results.

Keywords: distributed parameter system, rationalized Haar transform, operational matrix, system identification

Procedia PDF Downloads 493
5085 Simulating Drilling Using a CAD System

Authors: Panagiotis Kyratsis, Konstantinos Kakoulis

Abstract:

Nowadays, the rapid development of CAD systems’ programming environments results in the creation of multiple downstream applications, which are developed and becoming increasingly available. CAD based manufacturing simulations is gradually following the same trend. Drilling is the most popular hole-making process used in a variety of industries. A specially built piece of software that deals with the drilling kinematics is presented. The cutting forces are calculated based on the tool geometry, the cutting conditions and the tool/work piece materials. The results are verified by experimental work. Finally, the response surface methodology (RSM) is applied and mathematical models of the total thrust force and the thrust force developed because of the main cutting edges are proposed.

Keywords: CAD, application programming interface, response surface methodology, drilling, RSM

Procedia PDF Downloads 455
5084 The Use of Social Networking Sites in eLearning

Authors: Clifford De Raffaele, Luana Bugeja, Serengul Smith

Abstract:

The adaptation of social networking sites within higher education has garnered significant interest in the recent years with numerous researches considering it as a possible shift from the traditional classroom based learning paradigm. Notwithstanding this increase in research and conducted studies however, the adaption of SNS based modules have failed to proliferate within Universities. This paper, commences its contribution by analyzing the various models and theories proposed in literature and amalgamates together various effective aspects for the inclusion of social technology within e-Learning. A three phased framework is further proposed which details the necessary considerations for the successful adaptation of SNS in enhancing the students learning experience. This proposal outlines the theoretical foundations which will be analyzed in practical implementation across international university campuses.

Keywords: eLearning, higher education, social network sites, student learning

Procedia PDF Downloads 325
5083 Experimental Squeeze Flow of Bitumen: Rheological Properties

Authors: A. Kraiem, A. Ayadi

Abstract:

The squeeze flow tests were studied by many authors to measure the rheological properties of fluid. Experimental squeezing flow test with constant area between two parallel disks of bitumen is investigated in the present work. The effect of the temperature, the process of preparing the sample and the gap between the discs were discussed. The obtained results were compared with the theoretical models. The behavior of bitumen depends on the viscosity and the yield stress. Thus, the bitumen was presented as a power law for a small power law exponent and as a biviscous fluid when the viscosity ratio was smaller than one. Also, the influence of the ambient temperature is required for the compression test. Therefore, for a high temperature the yield stress decrease.

Keywords: bitumen, biviscous fluid, squeeze flow, viscosity, yield stress

Procedia PDF Downloads 124
5082 A Probability Analysis of Construction Project Schedule Using Risk Management Tool

Authors: A. L. Agarwal, D. A. Mahajan

Abstract:

Construction industry tumbled along with other industry/sectors during recent economic crash. Construction business could not regain thereafter and still pass through slowdown phase, resulted many real estate as well as infrastructure projects not completed on schedule and within budget. There are many theories, tools, techniques with software packages available in the market to analyze construction schedule. This study focuses on the construction project schedule and uncertainties associated with construction activities. The infrastructure construction project has been considered for the analysis of uncertainty on project activities affecting project duration and analysis is done using @RISK software. Different simulation results arising from three probability distribution functions are compiled to benefit construction project managers to plan more realistic schedule of various construction activities as well as project completion to document in the contract and avoid compensations or claims arising out of missing the planned schedule.

Keywords: construction project, distributions, project schedule, uncertainty

Procedia PDF Downloads 336
5081 Predictive Modelling of Curcuminoid Bioaccessibility as a Function of Food Formulation and Associated Properties

Authors: Kevin De Castro Cogle, Mirian Kubo, Maria Anastasiadi, Fady Mohareb, Claire Rossi

Abstract:

Background: The bioaccessibility of bioactive compounds is a critical determinant of the nutritional quality of various food products. Despite its importance, there is a limited number of comprehensive studies aimed at assessing how the composition of a food matrix influences the bioaccessibility of a compound of interest. This knowledge gap has prompted a growing need to investigate the intricate relationship between food matrix formulations and the bioaccessibility of bioactive compounds. One such class of bioactive compounds that has attracted considerable attention is curcuminoids. These naturally occurring phytochemicals, extracted from the roots of Curcuma longa, have gained popularity owing to their purported health benefits and also well known for their poor bioaccessibility Project aim: The primary objective of this research project is to systematically assess the influence of matrix composition on the bioaccessibility of curcuminoids. Additionally, this study aimed to develop a series of predictive models for bioaccessibility, providing valuable insights for optimising the formula for functional foods and provide more descriptive nutritional information to potential consumers. Methods: Food formulations enriched with curcuminoids were subjected to in vitro digestion simulation, and their bioaccessibility was characterized with chromatographic and spectrophotometric techniques. The resulting data served as the foundation for the development of predictive models capable of estimating bioaccessibility based on specific physicochemical properties of the food matrices. Results: One striking finding of this study was the strong correlation observed between the concentration of macronutrients within the food formulations and the bioaccessibility of curcuminoids. In fact, macronutrient content emerged as a very informative explanatory variable of bioaccessibility and was used, alongside other variables, as predictors in a Bayesian hierarchical model that predicted curcuminoid bioaccessibility accurately (optimisation performance of 0.97 R2) for the majority of cross-validated test formulations (LOOCV of 0.92 R2). These preliminary results open the door to further exploration, enabling researchers to investigate a broader spectrum of food matrix types and additional properties that may influence bioaccessibility. Conclusions: This research sheds light on the intricate interplay between food matrix composition and the bioaccessibility of curcuminoids. This study lays a foundation for future investigations, offering a promising avenue for advancing our understanding of bioactive compound bioaccessibility and its implications for the food industry and informed consumer choices.

Keywords: bioactive bioaccessibility, food formulation, food matrix, machine learning, probabilistic modelling

Procedia PDF Downloads 60
5080 Fragility Assessment for Torsionally Asymmetric Buildings in Plan

Authors: S. Feli, S. Tavousi Tafreshi, A. Ghasemi

Abstract:

The present paper aims at evaluating the response of three-dimensional buildings with in-plan stiffness irregularities that have been subjected to two-way excitation ground motion records simultaneously. This study is broadly-based fragility assessment with greater emphasis on structural response at in-plan flexible and stiff sides. To this end, three type of three-dimensional 5-story steel building structures with stiffness eccentricities, were subjected to extensive nonlinear incremental dynamic analyses (IDA) utilizing Ibarra-Krawinkler deterioration models. Fragility assessment was implemented for different configurations of braces to investigate the losses in buildings with center of resisting (CR) eccentricities.

Keywords: Ibarra-Krawinkler, fragility assessment, flexible and stiff side, center of resisting

Procedia PDF Downloads 194
5079 Real Activities Manipulation vs. Accrual Earnings Management: The Effect of Political Risk

Authors: Heba Abdelmotaal, Magdy Abdel-Kader

Abstract:

Purpose: This study explores whether a firm’s effective political risk management is preventing real and accrual earnings management . Design/methodology/approach: Based on a sample of 130 firms operating in Egypt during the period 2008-2013, two hypotheses are tested using the panel data regression models. Findings: The empirical findings indicate a significant relation between real and accrual earnings management and political risk. Originality/value: This paper provides a statistically evidence on the effects of the political risk management failure on the mangers’ engagement in the real and accrual earnings management practices, and its impact on the firm’s performance.

Keywords: political risk, risk management failure, real activities manipulation, accrual earnings management

Procedia PDF Downloads 423
5078 Whistleblowing a Contemporary Topic Concerning Businesses

Authors: Andreas Kapardis, Maria Krambia-Kapardis, Sofia Michaelides-Mateou

Abstract:

Corruption and economic crime is a serious problem affecting the sustainability of businesses in the 21st century. Nowadays, many corruption or fraud cases come to light thanks to whistleblowers. This article will first discuss the concept of whistleblowing as well as some relevant legislation enacted around the world. Secondly, it will discuss the findings of a survey of whistleblowers or could-have-been whistleblowers. Finally, suggestions for the development of a comprehensive whistleblowing framework will be considered. Whistleblowing can be described as expressing a concern about a wrongdoing within an organization, such as a corporation, an association, an institution or a union. Such concern must be in the public interest and in good faith and should relate to the cover up of matters that could potentially result in a miscarriage of justice, a crime, criminal offence and threats to health and safety. Whistleblowing has proven to be an effective anti-corruption mechanism and a powerful tool that helps deterring fraud, violations, and malpractices within organizations, corporations and the public sector. Research in the field of whistleblowing has concentrated on the reasons for whistleblowing and financial bounties; the effectiveness of whistleblowing; whistleblowing being a prosocial behavior with a psychological perspective and consequences; as a tool in protecting shareholders, saving lives and billions of dollars of public funds. Whilst, no other study of whistleblowing has been carried out on whistleblowers or intended whistleblowers. The study reported in the current paper analyses the findings of 74 whistleblowers or intended whistleblowers, the reasons behind their decision to blow the whistle, or not to proceed to blow the whistle and any regrets they may have had. In addition a profile of a whistleblower is developed concerning their age, gender, marital and family status and position in an organization. Lessons learned from the intended whistleblowers and in response to the questions if they would be willing to blow the whistle again show that enacting legislation to protect the whistleblower is not enough. Similarly, rewarding the whistleblower does not appear to provide the whistleblower with an incentive since the majority noted that “work ethics is more important than financial rewards”. We recommend the development of a comprehensive and holistic framework for the protection of the whistleblower and to ensure that remedial actions are immediately taken once a whistleblower comes forward. The suggested framework comprises (a) hard legislation in ensuring the whistleblowers follow certain principles when blowing the whistle and, in return, are protected for a period of 5 years from being fired, dismissed, bullied, harassed; (b) soft legislation in establishing an agency to firstly ensure psychological and legal advice is provided to the whistleblowers and secondly any required remedial action is immediately taken to avert the undesirable events reported by a whistleblower from occurring and, finally; (c) mechanisms to ensure the coordination of actions taken.

Keywords: whistleblowing, business ethics, legislation, business

Procedia PDF Downloads 298
5077 Studies on the Applicability of Artificial Neural Network (ANN) in Prediction of Thermodynamic Behavior of Sodium Chloride Aqueous System Containing a Non-Electrolytes

Authors: Dariush Jafari, S. Mostafa Nowee

Abstract:

In this study a ternary system containing sodium chloride as solute, water as primary solvent and ethanol as the antisolvent was considered to investigate the application of artificial neural network (ANN) in prediction of sodium solubility in the mixture of water as the solvent and ethanol as the antisolvent. The system was previously studied using by Extended UNIQUAC model by the authors of this study. The comparison between the results of the two models shows an excellent agreement between them (R2=0.99), and also approves the capability of ANN to predict the thermodynamic behavior of ternary electrolyte systems which are difficult to model.

Keywords: thermodynamic modeling, ANN, solubility, ternary electrolyte system

Procedia PDF Downloads 376
5076 Lee-Carter Mortality Forecasting Method with Dynamic Normal Inverse Gaussian Mortality Index

Authors: Funda Kul, İsmail Gür

Abstract:

Pension scheme providers have to price mortality risk by accurate mortality forecasting method. There are many mortality-forecasting methods constructed and used in literature. The Lee-Carter model is the first model to consider stochastic improvement trends in life expectancy. It is still precisely used. Mortality forecasting is done by mortality index in the Lee-Carter model. It is assumed that mortality index fits ARIMA time series model. In this paper, we propose and use dynamic normal inverse gaussian distribution to modeling mortality indes in the Lee-Carter model. Using population mortality data for Italy, France, and Turkey, the model is forecasting capability is investigated, and a comparative analysis with other models is ensured by some well-known benchmarking criterions.

Keywords: mortality, forecasting, lee-carter model, normal inverse gaussian distribution

Procedia PDF Downloads 345
5075 Tumor Detection Using Convolutional Neural Networks (CNN) Based Neural Network

Authors: Vinai K. Singh

Abstract:

In Neural Network-based Learning techniques, there are several models of Convolutional Networks. Whenever the methods are deployed with large datasets, only then can their applicability and appropriateness be determined. Clinical and pathological pictures of lobular carcinoma are thought to exhibit a large number of random formations and textures. Working with such pictures is a difficult problem in machine learning. Focusing on wet laboratories and following the outcomes, numerous studies have been published with fresh commentaries in the investigation. In this research, we provide a framework that can operate effectively on raw photos of various resolutions while easing the issues caused by the existence of patterns and texturing. The suggested approach produces very good findings that may be used to make decisions in the diagnosis of cancer.

Keywords: lobular carcinoma, convolutional neural networks (CNN), deep learning, histopathological imagery scans

Procedia PDF Downloads 123
5074 Artificial Intelligence Based Comparative Analysis for Supplier Selection in Multi-Echelon Automotive Supply Chains via GEP and ANN Models

Authors: Seyed Esmail Seyedi Bariran, Laysheng Ewe, Amy Ling

Abstract:

Since supplier selection appears as a vital decision, selecting supplier based on the best and most accurate ways has a lot of importance for enterprises. In this study, a new Artificial Intelligence approach is exerted to remove weaknesses of supplier selection. The paper has three parts. First part is choosing the appropriate criteria for assessing the suppliers’ performance. Next one is collecting the data set based on experts. Afterwards, the data set is divided into two parts, the training data set and the testing data set. By the training data set the best structure of GEP and ANN are selected and to evaluate the power of the mentioned methods the testing data set is used. The result obtained shows that the accuracy of GEP is more than ANN. Moreover, unlike ANN, a mathematical equation is presented by GEP for the supplier selection.

Keywords: supplier selection, automotive supply chains, ANN, GEP

Procedia PDF Downloads 614
5073 A Genetic-Neural-Network Modeling Approach for Self-Heating in GaN High Electron Mobility Transistors

Authors: Anwar Jarndal

Abstract:

In this paper, a genetic-neural-network (GNN) based large-signal model for GaN HEMTs is presented along with its parameters extraction procedure. The model is easy to construct and implement in CAD software and requires only DC and S-parameter measurements. An improved decomposition technique is used to model self-heating effect. Two GNN models are constructed to simulate isothermal drain current and power dissipation, respectively. The two model are then composed to simulate the drain current. The modeling procedure was applied to a packaged GaN-on-Si HEMT and the developed model is validated by comparing its large-signal simulation with measured data. A very good agreement between the simulation and measurement is obtained.

Keywords: GaN HEMT, computer-aided design and modeling, neural networks, genetic optimization

Procedia PDF Downloads 369
5072 A Blind Three-Dimensional Meshes Watermarking Using the Interquartile Range

Authors: Emad E. Abdallah, Alaa E. Abdallah, Bajes Y. Alskarnah

Abstract:

We introduce a robust three-dimensional watermarking algorithm for copyright protection and indexing. The basic idea behind our technique is to measure the interquartile range or the spread of the 3D model vertices. The algorithm starts by converting all the vertices to spherical coordinate followed by partitioning them into small groups. The proposed algorithm is slightly altering the interquartile range distribution of the small groups based on predefined watermark. The experimental results on several 3D meshes prove perceptual invisibility and the robustness of the proposed technique against the most common attacks including compression, noise, smoothing, scaling, rotation as well as combinations of these attacks.

Keywords: watermarking, three-dimensional models, perceptual invisibility, interquartile range, 3D attacks

Procedia PDF Downloads 462
5071 Model-Driven and Data-Driven Approaches for Crop Yield Prediction: Analysis and Comparison

Authors: Xiangtuo Chen, Paul-Henry Cournéde

Abstract:

Crop yield prediction is a paramount issue in agriculture. The main idea of this paper is to find out efficient way to predict the yield of corn based meteorological records. The prediction models used in this paper can be classified into model-driven approaches and data-driven approaches, according to the different modeling methodologies. The model-driven approaches are based on crop mechanistic modeling. They describe crop growth in interaction with their environment as dynamical systems. But the calibration process of the dynamic system comes up with much difficulty, because it turns out to be a multidimensional non-convex optimization problem. An original contribution of this paper is to propose a statistical methodology, Multi-Scenarios Parameters Estimation (MSPE), for the parametrization of potentially complex mechanistic models from a new type of datasets (climatic data, final yield in many situations). It is tested with CORNFLO, a crop model for maize growth. On the other hand, the data-driven approach for yield prediction is free of the complex biophysical process. But it has some strict requirements about the dataset. A second contribution of the paper is the comparison of these model-driven methods with classical data-driven methods. For this purpose, we consider two classes of regression methods, methods derived from linear regression (Ridge and Lasso Regression, Principal Components Regression or Partial Least Squares Regression) and machine learning methods (Random Forest, k-Nearest Neighbor, Artificial Neural Network and SVM regression). The dataset consists of 720 records of corn yield at county scale provided by the United States Department of Agriculture (USDA) and the associated climatic data. A 5-folds cross-validation process and two accuracy metrics: root mean square error of prediction(RMSEP), mean absolute error of prediction(MAEP) were used to evaluate the crop prediction capacity. The results show that among the data-driven approaches, Random Forest is the most robust and generally achieves the best prediction error (MAEP 4.27%). It also outperforms our model-driven approach (MAEP 6.11%). However, the method to calibrate the mechanistic model from dataset easy to access offers several side-perspectives. The mechanistic model can potentially help to underline the stresses suffered by the crop or to identify the biological parameters of interest for breeding purposes. For this reason, an interesting perspective is to combine these two types of approaches.

Keywords: crop yield prediction, crop model, sensitivity analysis, paramater estimation, particle swarm optimization, random forest

Procedia PDF Downloads 221
5070 Enhancing Large Language Models' Data Analysis Capability with Planning-and-Execution and Code Generation Agents: A Use Case for Southeast Asia Real Estate Market Analytics

Authors: Kien Vu, Jien Min Soh, Mohamed Jahangir Abubacker, Piyawut Pattamanon, Soojin Lee, Suvro Banerjee

Abstract:

Recent advances in Generative Artificial Intelligence (GenAI), in particular Large Language Models (LLMs) have shown promise to disrupt multiple industries at scale. However, LLMs also present unique challenges, notably, these so-called "hallucination" which is the generation of outputs that are not grounded in the input data that hinders its adoption into production. Common practice to mitigate hallucination problem is utilizing Retrieval Agmented Generation (RAG) system to ground LLMs'response to ground truth. RAG converts the grounding documents into embeddings, retrieve the relevant parts with vector similarity between user's query and documents, then generates a response that is not only based on its pre-trained knowledge but also on the specific information from the retrieved documents. However, the RAG system is not suitable for tabular data and subsequent data analysis tasks due to multiple reasons such as information loss, data format, and retrieval mechanism. In this study, we have explored a novel methodology that combines planning-and-execution and code generation agents to enhance LLMs' data analysis capabilities. The approach enables LLMs to autonomously dissect a complex analytical task into simpler sub-tasks and requirements, then convert them into executable segments of code. In the final step, it generates the complete response from output of the executed code. When deployed beta version on DataSense, the property insight tool of PropertyGuru, the approach yielded promising results, as it was able to provide market insights and data visualization needs with high accuracy and extensive coverage by abstracting the complexities for real-estate agents and developers from non-programming background. In essence, the methodology not only refines the analytical process but also serves as a strategic tool for real estate professionals, aiding in market understanding and enhancement without the need for programming skills. The implication extends beyond immediate analytics, paving the way for a new era in the real estate industry characterized by efficiency and advanced data utilization.

Keywords: large language model, reasoning, planning and execution, code generation, natural language processing, prompt engineering, data analysis, real estate, data sense, PropertyGuru

Procedia PDF Downloads 65
5069 Optimization of Electric Vehicle (EV) Charging Station Allocation Based on Multiple Data - Taking Nanjing (China) as an Example

Authors: Yue Huang, Yiheng Feng

Abstract:

Due to the global pressure on climate and energy, many countries are vigorously promoting electric vehicles and building charging (public) charging facilities. Faced with the supply-demand gap of existing electric vehicle charging stations and unreasonable space usage in China, this paper takes the central city of Nanjing as an example, establishes a site selection model through multivariate data integration, conducts multiple linear regression SPSS analysis, gives quantitative site selection results, and provides optimization models and suggestions for charging station layout planning.

Keywords: electric vehicle, charging station, allocation optimization, urban mobility, urban infrastructure, nanjing

Procedia PDF Downloads 78