Search results for: risk management model
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 28128

Search results for: risk management model

27408 A New Nonlinear State-Space Model and Its Application

Authors: Abdullah Eqal Al Mazrooei

Abstract:

In this work, a new nonlinear model will be introduced. The model is in the state-space form. The nonlinearity of this model is in the state equation where the state vector is multiplied by its self. This technique makes our model generalizes many famous models as Lotka-Volterra model and Lorenz model which have many applications in the real life. We will apply our new model to estimate the wind speed by using a new nonlinear estimator which suitable to work with our model.

Keywords: nonlinear systems, state-space model, Kronecker product, nonlinear estimator

Procedia PDF Downloads 691
27407 Application of the State of the Art of Hydraulic Models to Manage Coastal Problems, Case Study: The Egyptian Mediterranean Coast Model

Authors: Al. I. Diwedar, Moheb Iskander, Mohamed Yossef, Ahmed ElKut, Noha Fouad, Radwa Fathy, Mustafa M. Almaghraby, Amira Samir, Ahmed Romya, Nourhan Hassan, Asmaa Abo Zed, Bas Reijmerink, Julien Groenenboom

Abstract:

Coastal problems are stressing the coastal environment due to its complexity. The dynamic interaction between the sea and the land results in serious problems that threaten coastal areas worldwide, in addition to human interventions and activities. This makes the coastal environment highly vulnerable to natural processes like flooding, erosion, and the impact of human activities as pollution. Protecting and preserving this vulnerable coastal zone with its valuable ecosystems calls for addressing the coastal problems. This, in the end, will support the sustainability of the coastal communities and maintain the current and future generations. Consequently applying suitable management strategies and sustainable development that consider the unique characteristics of the coastal system is a must. The coastal management philosophy aims to solve the conflicts of interest between human development activities and this dynamic nature. Modeling emerges as a successful tool that provides support to decision-makers, engineers, and researchers for better management practices. Modeling tools proved that it is accurate and reliable in prediction. With its capability to integrate data from various sources such as bathymetric surveys, satellite images, and meteorological data, it offers the possibility for engineers and scientists to understand this complex dynamic system and get in-depth into the interaction between both the natural and human-induced factors. This enables decision-makers to make informed choices and develop effective strategies for sustainable development and risk mitigation of the coastal zone. The application of modeling tools supports the evaluation of various scenarios by affording the possibility to simulate and forecast different coastal processes from the hydrodynamic and wave actions and the resulting flooding and erosion. The state-of-the-art application of modeling tools in coastal management allows for better understanding and predicting coastal processes, optimizing infrastructure planning and design, supporting ecosystem-based approaches, assessing climate change impacts, managing hazards, and finally facilitating stakeholder engagement. This paper emphasizes the role of hydraulic models in enhancing the management of coastal problems by discussing the diverse applications of modeling in coastal management. It highlights the modelling role in understanding complex coastal processes, and predicting outcomes. The importance of informing decision-makers with modeling results which gives technical and scientific support to achieve sustainable coastal development and protection.

Keywords: coastal problems, coastal management, hydraulic model, numerical model, physical model

Procedia PDF Downloads 28
27406 Financial Instruments Disclosure: A Review of the Literature

Authors: Y. Tahat, T. Dunne, S. Fifield, D. Power

Abstract:

Information about a firm’s usage of Financial Instruments (FIs) plays a very important role in determining its financial position and performance. Yet accounting standard-setters have encountered problems when deciding on the FI-related disclosures which firms must make. The primary objective of this paper is to review the extant literature on FI disclosure. This objective is achieved by surveying the literature on: the corporate usage of FIs; the different accounting standards adopted concerning FIs; and empirical studies on FI disclosure. This review concludes that the current research on FI disclosure has generated a number of useful insights. In particular, the paper reports that: FIs are a very important risk management mechanism in ensuring that companies have the cash available to make value-enhancing investments, however, without a clear set of risk management objectives, using such instruments can be dangerous; accounting standards concerning FIs have resulted in enhanced transparency about the usage of these instruments; and FI-related information is a key input into investors’ decision-making processes. Finally, the paper provides a number of suggestions for future research in the area.

Keywords: financial instruments, financial reporting, accounting standards, value relevance, corporate disclosure

Procedia PDF Downloads 412
27405 Risk Assessment in Construction of K-Span Buildings in United Arab Emirates (UAE)

Authors: Imtiaz Ali, Imam Mansoor

Abstract:

Investigations as a part of the academic study were undertaken to identify and evaluate the significant risks associated with the construction of K-span buildings in the region of UAE. Primary field data was collected through questionnaires obtaining specific open and close-ended questions from carefully selected construction firms, civil engineers and, construction manager regarding risks associated to K-span building construction. Historical data available for other regions of the same construction technique was available which was compared for identifying various non-critical and critical risk parameters by comparative evaluation techniques to come up with important risks and potential sources for their control and minimization in K-Span buildings that is increasing in the region. The associated risks have been determined with their Relative Importance Index (RII) values of which Risk involved in Change of Design required by Owners carries the highest value (RII=0.79) whereas, Delayed Payment by Owner to Contractor is one of the least (RII=0.42) value. The overall findings suggest that most relative risks as quantified originate or associated with the contractors. It may be concluded that project proponents undertaking K-span projects in planning and budgeting the cost and delays should take into account of risks on high account if changes in design are also required any delays in the material by the supplier would then be a major risk in K-span project delay. Since projects are, less costly, so owners have limited budgets, then they hire small contractors, which are not highly competent contractors. So study suggests that owner should be aware of these types of risks associated with the construction of K-span buildings in order to make it cost effective.

Keywords: k-span buildings, k-span construction, risk management, relative improvement index (RII)

Procedia PDF Downloads 375
27404 Demographic Assessment and Evaluation of Degree of Lipid Control in High Risk Indian Dyslipidemia Patients

Authors: Abhijit Trailokya

Abstract:

Background: Cardiovascular diseases (CVD’s) are the major cause of morbidity and mortality in both developed and developing countries. Many clinical trials have demonstrated that low-density lipoprotein cholesterol (LDL-C) lowering, reduces the incidence of coronary and cerebrovascular events across a broad spectrum of patients at risk. Guidelines for the management of patients at risk have been established in Europe and North America. The guidelines have advocated progressively lower LDL-C targets and more aggressive use of statin therapy. In Indian patients, comprehensive data on dyslipidemia management and its treatment outcomes are inadequate. There is lack of information on existing treatment patterns, the patient’s profile being treated, and factors that determine treatment success or failure in achieving desired goals. Purpose: The present study was planned to determine the lipid control status in high-risk dyslipidemic patients treated with lipid-lowering therapy in India. Methods: This cross-sectional, non-interventional, single visit program was conducted across 483 sites in India where male and female patients with high-risk dyslipidemia aged 18 to 65 years who had visited for a routine health check-up to their respective physician at hospital or a healthcare center. Percentage of high-risk dyslipidemic patients achieving adequate LDL-C level (< 70 mg/dL) on lipid-lowering therapy and the association of lipid parameters with patient characteristics, comorbid conditions, and lipid lowering drugs were analysed. Results: 3089 patients were enrolled in the study; of which 64% were males. LDL-C data was available for 95.2% of the patients; only 7.7% of these patients achieved LDL-C levels < 70 mg/dL on lipid-lowering therapy, which may be due to inability to follow therapeutic plans, poor compliance, or inadequate counselling by physician. The physician’s lack of awareness about recent treatment guidelines also might contribute to patients’ poor adherence, not explaining adequately the benefit and risks of a medication, not giving consideration to the patient’s life style and the cost of medication. Statin was the most commonly used anti-dyslipidemic drug across population. The higher proportion of patients had the comorbid condition of CVD and diabetes mellitus across all dyslipidemic patients. Conclusion: As per the European Society of Cardiology guidelines the ideal LDL-C levels in high risk dyslipidemic patients should be less than 70%. In the present study, 7.7% of the patients achieved LDL-C levels < 70 mg/dL on lipid lowering therapy which is very less. Most of high risk dyslipidemic patients in India are on suboptimal dosage of statin. So more aggressive and high dosage statin therapy may be required to achieve target LDLC levels in high risk Indian dyslipidemic patients.

Keywords: cardiovascular disease, diabetes mellitus, dyslipidemia, LDL-C, lipid lowering drug, statins

Procedia PDF Downloads 201
27403 Screening of Strategic Management Criterions in Hospitals Using Delphi-Fuzzy Method

Authors: Helia Moayedi, Mahdi Moaidi

Abstract:

Nowadays, the managing and planning of hospitals is facing many problems. Failure to recognize the main criteria for strategic management to ensure long-term hospital performance can lead to many health problems. To achieve this goal, a qualitative-quantitate method titled Delphi-Fuzzy has been applied. This strategy makes it possible for experts to screen among the most important criteria in strategic management. To conduct this operation, a statistical society consisting of 20 experts in Ahwaz hospitals has been questioned. The final model confirms the key criterions after three stages of Delphi. This model provides the possibility to focus on the basic criteria and can determine the organization’s main orientation.

Keywords: Delphi-fuzzy method, hospital management, long-term planning, qualitative-quantitate method, screening of strategic criteria, strategic planning

Procedia PDF Downloads 131
27402 Improving the Employee Transfer Experience within an Organization

Authors: Drew Fockler

Abstract:

This research examines how to improve an employee’s experience when transferring between departments within an organization. This research includes a historical review of a Canadian retail organization. Based on this historical review, gaps are identified between current and future visions to show where problems with existing training and development practices need to be resolved to reduce front-line employee turnover within an organization. The strategies within this paper support leaders through the LEAD: Listen, Explore, Act and Develop, Change Management Model. The LEAD Change Management Model supports the change process. This research proposes three possible solutions to improve an employee who is transferring between departments. The best solution to resolve the problem of improving an employee moving between departments experience is creating a Training Manager position within the retail store. A Training Manager position could support both employees and leadership with training and development of staff who are moving between departments. Within this research, an implementation plan using the TransX Model was created. The TransX Model is a hybrid of Leader-Member Exchange Theory and Transformational Leadership Theory to facilitate this organizational change within an organization by creating a common vision. Finally, this research provides the next steps as well as future considerations to enhance the training manager role within an organization.

Keywords: employee transfers, employee engagement, human resources, employee induction, TransX model, lead change management model

Procedia PDF Downloads 77
27401 Governance, Risk Management, and Compliance Factors Influencing the Adoption of Cloud Computing in Australia

Authors: Tim Nedyalkov

Abstract:

A business decision to move to the cloud brings fundamental changes in how an organization develops and delivers its Information Technology solutions. The accelerated pace of digital transformation across businesses and government agencies increases the reliance on cloud-based services. They are collecting, managing, and retaining large amounts of data in cloud environments makes information security and data privacy protection essential. It becomes even more important to understand what key factors drive successful cloud adoption following the commencement of the Privacy Amendment Notifiable Data Breaches (NDB) Act 2017 in Australia as the regulatory changes impact many organizations and industries. This quantitative correlational research investigated the governance, risk management, and compliance factors contributing to cloud security success. The factors influence the adoption of cloud computing within an organizational context after the commencement of the NDB scheme. The results and findings demonstrated that corporate information security policies, data storage location, management understanding of data governance responsibilities, and regular compliance assessments are the factors influencing cloud computing adoption. The research has implications for organizations, future researchers, practitioners, policymakers, and cloud computing providers to meet the rapidly changing regulatory and compliance requirements.

Keywords: cloud compliance, cloud security, data governance, privacy protection

Procedia PDF Downloads 116
27400 Comparison Study of Capital Protection Risk Management Strategies: Constant Proportion Portfolio Insurance versus Volatility Target Based Investment Strategy with a Guarantee

Authors: Olga Biedova, Victoria Steblovskaya, Kai Wallbaum

Abstract:

In the current capital market environment, investors constantly face the challenge of finding a successful and stable investment mechanism. Highly volatile equity markets and extremely low bond returns bring about the demand for sophisticated yet reliable risk management strategies. Investors are looking for risk management solutions to efficiently protect their investments. This study compares a classic Constant Proportion Portfolio Insurance (CPPI) strategy to a Volatility Target portfolio insurance (VTPI). VTPI is an extension of the well-known Option Based Portfolio Insurance (OBPI) to the case where an embedded option is linked not to a pure risky asset such as e.g., S&P 500, but to a Volatility Target (VolTarget) portfolio. VolTarget strategy is a recently emerged rule-based dynamic asset allocation mechanism where the portfolio’s volatility is kept under control. As a result, a typical VTPI strategy allows higher participation rates in the market due to reduced embedded option prices. In addition, controlled volatility levels eliminate the volatility spread in option pricing, one of the frequently cited reasons for OBPI strategy fall behind CPPI. The strategies are compared within the framework of the stochastic dominance theory based on numerical simulations, rather than on the restrictive assumption of the Black-Scholes type dynamics of the underlying asset. An extended comparative quantitative analysis of performances of the above investment strategies in various market scenarios and within a range of input parameter values is presented.

Keywords: CPPI, portfolio insurance, stochastic dominance, volatility target

Procedia PDF Downloads 167
27399 The Artificial Intelligence (AI) Impact on Project Management: A Destructive or Transformative Agent

Authors: Kwame Amoah

Abstract:

Artificial intelligence (AI) has the prospect of transforming project management, significantly improving efficiency and accuracy. By automating specific tasks with defined guidelines, AI can assist project managers in making better decisions and allocating resources efficiently, with possible risk mitigation. This study explores how AI is already impacting project management and likely future AI's impact on the field. The AI's reaction has been a divided opinion; while others picture it as a destroyer of jobs, some welcome it as an innovation advocate. Both sides agree that AI will be disruptive and revolutionize PM's functions. If current research is to go by, AI or some form will replace one-third of all learning graduate PM jobs by as early as 2030. A recent survey indicates AI spending will reach $97.9 billion by the end of 2023. Considering such a profound impact, the project management profession will also see a paradigm shift driven by AI. The study examines what the project management profession will look like in the next 5-10 years after this technological disruption. The research methods incorporate existing literature, develop trend analysis, and conduct structured interviews with project management stakeholders from North America to gauge the trend. PM professionals can harness the power of AI, ensuring a smooth transition and positive outcomes. AI adoption will maximize benefits, minimize adverse consequences, and uphold ethical standards, leading to improved project performance.

Keywords: project management, disruptive teacnologies, project management function, AL applications, artificial intelligence

Procedia PDF Downloads 83
27398 Continuous-Time Convertible Lease Pricing and Firm Value

Authors: Ons Triki, Fathi Abid

Abstract:

Along with the increase in the use of leasing contracts in corporate finance, multiple studies aim to model the credit risk of the lease in order to cover the losses of the lessor of the asset if the lessee goes bankrupt. In the current research paper, a convertible lease contract is elaborated in a continuous time stochastic universe aiming to ensure the financial stability of the firm and quickly recover the losses of the counterparties to the lease in case of default. This work examines the term structure of the lease rates taking into account the credit default risk and the capital structure of the firm. The interaction between the lessee's capital structure and the equilibrium lease rate has been assessed by applying the competitive lease market argument developed by Grenadier (1996) and the endogenous structural default model set forward by Leland and Toft (1996). The cumulative probability of default was calculated by referring to Leland and Toft (1996) and Yildirim and Huan (2006). Additionally, the link between lessee credit risk and lease rate was addressed so as to explore the impact of convertible lease financing on the term structure of the lease rate, the optimal leverage ratio, the cumulative default probability, and the optimal firm value by applying an endogenous conversion threshold. The numerical analysis is suggestive that the duration structure of lease rates increases with the increase in the degree of the market price of risk. The maximal value of the firm decreases with the effect of the optimal leverage ratio. The results are indicative that the cumulative probability of default increases with the maturity of the lease contract if the volatility of the asset service flows is significant. Introducing the convertible lease contract will increase the optimal value of the firm as a function of asset volatility for a high initial service flow level and a conversion ratio close to 1.

Keywords: convertible lease contract, lease rate, credit-risk, capital structure, default probability

Procedia PDF Downloads 98
27397 A Model Approach of Good Practice Based on the Project Management Body of Knowledge® Guide in the Project Owner

Authors: Claudia Marcela Munoz Gonzalez, Diego Fernando Hernandez Losada, Hugo Alberto Herrera Fonseca

Abstract:

The project owner's role in the public-private investment consists of controlling and verifying the correct execution of the project's objectives and resources. Likewise, it is a discipline little explored in the academic field, whereby this work wishes to contribute with a model of good practices based on the project management methodology proposed by the Project Management Body of Knowledge® Guide. In the same way, highlight what are the controls that an integral project owner should take into account in its exercise and application, through the stages in which its contract runs. This proposal aims to structure its practice and integrate its functions according to a project management methodology. In addition, these practices will be applied in a case study of projects in the agricultural sector, particularly in the construction of irrigation district in Cundinamarca, Colombia.

Keywords: controls, construction of irrigation district, PMBOK®, project owner

Procedia PDF Downloads 458
27396 Managing Diversity in MNCS: A Literature Review of Existing Strategic Models for Managing Diversity and a Roadmap to Transfer Them to the Subsidiaries

Authors: Debora Gottardello, Mireia Valverde Aparicio, Juan Llopis Taverner

Abstract:

Globalization has given rise to a great diversity in the composition of people in organizations. Diversity management is therefore key to create growth in today’s competitive global marketplace. This work develops a literature review related to the existing models for managing diversity covering the period from 1980 until 2014. Furthermore, it identifies limitations in previous models. More specifically, the literature review reveals that there is a lack of information about how these models can be adapted from the headquarters to the subsidiaries. Therefore, the contribution of this paper is to suggest how the models should be adapted when they are directed to host countries. Our aim is to highlight the limitations of the developed models with regards to the translation of the diversity management practices to the subsidiaries. Accordingly, a model that will enable MNCs to ensure a global strategy is suggested. Taking advantage of the potential incorporated in a culturally diverse work team should be at the top of every international company’s aims. Executives from headquarters need to use different attitudes when transferring diversity practices towards their subsidiaries. Further studies should reassess local practices of diversity management to find out how this universal management model is translated.

Keywords: culture diversity, diversity management, human resources management, MNCs, subsidiaries, workforce diversity

Procedia PDF Downloads 255
27395 The 10-year Risk of Major Osteoporotic and Hip Fractures Among Indonesian People Living with HIV

Authors: Iqbal Pramukti, Mamat Lukman, Hasniatisari Harun, Kusman Ibrahim

Abstract:

Introduction: People living with HIV had a higher risk of osteoporotic fracture than the general population. The purpose of this study was to predict the 10-year risk of fracture among people living with HIV (PLWH) using FRAX™ and to identify characteristics related to the fracture risk. Methodology: This study consisted of 75 subjects. The ten-year probability of major osteoporotic fractures (MOF) and hip fractures was assessed using the FRAX™ algorithm. A cross-tabulation was used to identify the participant’s characteristics related to fracture risk. Results: The overall mean 10-year probability of fracture was 2.4% (1.7) for MOF and 0.4% (0.3) for hip fractures. For MOF score, participants with parents’ hip fracture history, smoking behavior and glucocorticoid use showed a higher MOF score than those who were not (3.1 vs. 2.5; 4.6 vs 2.5; and 3.4 vs 2.5, respectively). For HF score, participants with parents’ hip fracture history, smoking behavior and glucocorticoid use also showed a higher HF score than those who were not (0.5 vs. 0.3; 0.8 vs. 0.3; and 0.5 vs. 0.3, respectively). Conclusions: The 10-year risk of fracture was higher among PLWH with several factors, including the parent’s hip. Fracture history, smoking behavior and glucocorticoid used. Further analysis on determining factors using multivariate regression analysis with a larger sample size is required to confirm the factors associated with the high fracture risk.

Keywords: HIV, PLWH, osteoporotic fractures, hip fractures, 10-year risk of fracture, FRAX

Procedia PDF Downloads 49
27394 A Framework for Auditing Multilevel Models Using Explainability Methods

Authors: Debarati Bhaumik, Diptish Dey

Abstract:

Multilevel models, increasingly deployed in industries such as insurance, food production, and entertainment within functions such as marketing and supply chain management, need to be transparent and ethical. Applications usually result in binary classification within groups or hierarchies based on a set of input features. Using open-source datasets, we demonstrate that popular explainability methods, such as SHAP and LIME, consistently underperform inaccuracy when interpreting these models. They fail to predict the order of feature importance, the magnitudes, and occasionally even the nature of the feature contribution (negative versus positive contribution to the outcome). Besides accuracy, the computational intractability of SHAP for binomial classification is a cause of concern. For transparent and ethical applications of these hierarchical statistical models, sound audit frameworks need to be developed. In this paper, we propose an audit framework for technical assessment of multilevel regression models focusing on three aspects: (i) model assumptions & statistical properties, (ii) model transparency using different explainability methods, and (iii) discrimination assessment. To this end, we undertake a quantitative approach and compare intrinsic model methods with SHAP and LIME. The framework comprises a shortlist of KPIs, such as PoCE (Percentage of Correct Explanations) and MDG (Mean Discriminatory Gap) per feature, for each of these three aspects. A traffic light risk assessment method is furthermore coupled to these KPIs. The audit framework will assist regulatory bodies in performing conformity assessments of AI systems using multilevel binomial classification models at businesses. It will also benefit businesses deploying multilevel models to be future-proof and aligned with the European Commission’s proposed Regulation on Artificial Intelligence.

Keywords: audit, multilevel model, model transparency, model explainability, discrimination, ethics

Procedia PDF Downloads 94
27393 An Integrated Real-Time Hydrodynamic and Coastal Risk Assessment Model

Authors: M. Reza Hashemi, Chris Small, Scott Hayward

Abstract:

The Northeast Coast of the US faces damaging effects of coastal flooding and winds due to Atlantic tropical and extratropical storms each year. Historically, several large storm events have produced substantial levels of damage to the region; most notably of which were the Great Atlantic Hurricane of 1938, Hurricane Carol, Hurricane Bob, and recently Hurricane Sandy (2012). The objective of this study was to develop an integrated modeling system that could be used as a forecasting/hindcasting tool to evaluate and communicate the risk coastal communities face from these coastal storms. This modeling system utilizes the ADvanced CIRCulation (ADCIRC) model for storm surge predictions and the Simulating Waves Nearshore (SWAN) model for the wave environment. These models were coupled, passing information to each other and computing over the same unstructured domain, allowing for the most accurate representation of the physical storm processes. The coupled SWAN-ADCIRC model was validated and has been set up to perform real-time forecast simulations (as well as hindcast). Modeled storm parameters were then passed to a coastal risk assessment tool. This tool, which is generic and universally applicable, generates spatial structural damage estimate maps on an individual structure basis for an area of interest. The required inputs for the coastal risk model included a detailed information about the individual structures, inundation levels, and wave heights for the selected region. Additionally, calculation of wind damage to structures was incorporated. The integrated coastal risk assessment system was then tested and applied to Charlestown, a small vulnerable coastal town along the southern shore of Rhode Island. The modeling system was applied to Hurricane Sandy and a synthetic storm. In both storm cases, effect of natural dunes on coastal risk was investigated. The resulting damage maps for the area (Charlestown) clearly showed that the dune eroded scenarios affected more structures, and increased the estimated damage. The system was also tested in forecast mode for a large Nor’Easters: Stella (March 2017). The results showed a good performance of the coupled model in forecast mode when compared to observations. Finally, a nearshore model XBeach was then nested within this regional grid (ADCIRC-SWAN) to simulate nearshore sediment transport processes and coastal erosion. Hurricane Irene (2011) was used to validate XBeach, on the basis of a unique beach profile dataset at the region. XBeach showed a relatively good performance, being able to estimate eroded volumes along the beach transects with a mean error of 16%. The validated model was then used to analyze the effectiveness of several erosion mitigation methods that were recommended in a recent study of coastal erosion in New England: beach nourishment, coastal bank (engineered core), and submerged breakwater as well as artificial surfing reef. It was shown that beach nourishment and coastal banks perform better to mitigate shoreline retreat and coastal erosion.

Keywords: ADCIRC, coastal flooding, storm surge, coastal risk assessment, living shorelines

Procedia PDF Downloads 116
27392 Formal Verification of Cache System Using a Novel Cache Memory Model

Authors: Guowei Hou, Lixin Yu, Wei Zhuang, Hui Qin, Xue Yang

Abstract:

Formal verification is proposed to ensure the correctness of the design and make functional verification more efficient. As cache plays a vital role in the design of System on Chip (SoC), and cache with Memory Management Unit (MMU) and cache memory unit makes the state space too large for simulation to verify, then a formal verification is presented for such system design. In the paper, a formal model checking verification flow is suggested and a new cache memory model which is called “exhaustive search model” is proposed. Instead of using large size ram to denote the whole cache memory, exhaustive search model employs just two cache blocks. For cache system contains data cache (Dcache) and instruction cache (Icache), Dcache memory model and Icache memory model are established separately using the same mechanism. At last, the novel model is employed to the verification of a cache which is module of a custom-built SoC system that has been applied in practical, and the result shows that the cache system is verified correctly using the exhaustive search model, and it makes the verification much more manageable and flexible.

Keywords: cache system, formal verification, novel model, system on chip (SoC)

Procedia PDF Downloads 496
27391 A Case Study of Brownfield Revitalization in Taiwan

Authors: Jen Wang, Wei-Chia Hsu, Zih-Sin Wang, Ching-Ping Chu, Bo-Shiou Guo

Abstract:

In the late 19th century, the Jinguashi ore deposit in northern Taiwan was discovered, and accompanied with flourishing mining activities. However, tons of contaminants including heavy metals, sulfur dioxide, and total petroleum hydrocarbons (TPH) were released to surroundings and caused environmental problems. Site T was one of copper smelter located on the coastal hill near Jinguashi ore deposit. In over ten years of operation, variety contaminants were emitted that it polluted the surrounding soil and groundwater quality. In order to exhaust fumes produced from smelting process, three stacks were built along the hill behind the factory. The sediment inside the stacks contains high concentration of heavy metals such as arsenic, lead, copper, etc. Moreover, soil around the discarded stacks suffered a serious contamination when deposition leached from the ruptures of stacks. Consequently, Site T (including the factory and its surroundings) was declared as a pollution remediation site that visiting the site and land-use activities on it are forbidden. However, the natural landscape and cultural attractions of Site T are spectacular that it attracts a lot of visitors annually. Moreover, land resources are extremely precious in Taiwan. In addition, Taiwan Environmental Protection Administration (EPA) is actively promoting the contaminated land revitalization policy. Therefore, this study took Site T as case study for brownfield revitalization planning to the limits of activate and remediate the natural resources. Land-use suitability analysis and risk mapping were applied in this study to make appropriate risk management measures and redevelopment plan for the site. In land-use suitability analysis, surrounding factors into consideration such as environmentally sensitive areas, biological resources, land use, contamination, culture, and landscapes were collected to assess the development of each area; health risk mapping was introduced to show the image of risk assessments results based on the site contamination investigation. According to land-use suitability analysis, the site was divided into four zones: priority area (for high-efficiency development), secondary area (for co-development with priority area), conditional area (for reusing existing building) and limited area (for Eco-tourism and education). According to the investigation, polychlorinated biphenyls (PCB), heavy metals and TPH were considered as target contaminants while oral, inhalation and dermal would be the major exposure pathways in health risk assessment. In accordance with health risk map, the highest risk was found in the southwest and eastern side. Based on the results, the development plan focused on zoning and land use. Site T was recommended be divides to public facility zone, public architectonic art zone, viewing zone, existing building preservation zone, historic building zone, and cultural landscape zone for various purpose. In addition, risk management measures including sustained remediation, extinguish exposure and administration management are applied to ensure particular places are suitable for visiting and protect the visitors’ health. The consolidated results are corroborated available by analyzing aspects of law, land acquired method, maintenance and management and public participation. Therefore, this study has a certain reference value to promote the contaminated land revitalization policy in Taiwan.

Keywords: brownfield revitalization, land-use suitability analysis, health risk map, risk management

Procedia PDF Downloads 184
27390 Prevalence Of High-Risk Obstructive Sleep Apnoea by Berlin Questionnaire in Patients with Type 2 Diabetes Mellitus: A Study from Tertiary Care Hospital

Authors: Leela Lakshmi Sai Srinivas Parasa, Pushpavalli Naidu, Bhaskar Kambhampati

Abstract:

Background: Obstructive Sleep Apnea (OSA) is a common but underdiagnosed sleep disorder characterized by repeated upper airway collapse during sleep. It frequently coexists with Type 2 Diabetes Mellitus (T2DM), a chronic metabolic condition with significant global and national health implications. The interplay between these conditions increases the risk of severe complications such as stroke, hypertension, and cardiovascular diseases. Understanding the prevalence of high-risk OSA in T2DM patients is essential for improving clinical outcomes and guiding public health interventions. Objective: This study aimed to assess the prevalence of high-risk OSA among T2DM patients using the Berlin questionnaire and to identify demographic and clinical predictors of risk. Methodology: This cross-sectional study of 100 patients with Type2 diabetes was undertaken at a tertiary care hospital in India for a period of 2 months. OSA was evaluated using the Berlin questionnaire. Diabetes mellitus was diagnosed according to WHO criteria as fasting blood sugar level > 7 mmol/L or random blood sugar >126 mg/dl on 2 separate occasions. Particulars of patients were recorded following which detailed history and physical examination were done. Data was analyzed using spss version 21. Results: The prevalence of high-risk OSA was 72% among the study population. The mean age was 54.9 +- 12.47 years. Age groups were divided into 3 groups (less than 50, 50-65, and greater than 65) The highest prevalence was observed in the 50-65 age group, accounting for 59.7% of high-risk cases. Gender distribution revealed a higher proportion of high-risk females (88.46%) compared to males (66.21%). Normal-weight individuals (BMI 18.5–24.9 kg/m²) represented 74.02% of high-risk OSA cases, while overweight individuals (BMI 25–29.9 kg/m²) accounted for 65.21%. Additionally, 90.9% of hypertensive patients were identified as high-risk for OSA. Conclusion: The study highlights a high prevalence of OSA risk among T2DM patients, especially those with hypertension and those in the middle-aged group. These findings emphasize the importance of routine OSA screening and tailored interventions for diabetic patients to reduce associated health risks. Proactive management can help mitigate the compounded effects of these two conditions, improving overall patient outcomes and reducing the burden on healthcare systems.

Keywords: obstructive sleep apnea, type 2 diabetes mellitus, berlin questionnaire, prevalence, hypertension

Procedia PDF Downloads 5
27389 Risk Tolerance and Individual Worthiness Based on Simultaneous Analysis of the Cognitive Performance and Emotional Response to a Multivariate Situational Risk Assessment

Authors: Frederic Jumelle, Kelvin So, Didan Deng

Abstract:

A method and system for neuropsychological performance test, comprising a mobile terminal, used to interact with a cloud server which stores user information and is logged into by the user through the terminal device; the user information is directly accessed through the terminal device and is processed by artificial neural network, and the user information comprises user facial emotions information, performance test answers information and user chronometrics. This assessment is used to evaluate the cognitive performance and emotional response of the subject to a series of dichotomous questions describing various situations of daily life and challenging the users' knowledge, values, ethics, and principles. In industrial applications, the timing of this assessment will depend on the users' need to obtain a service from a provider, such as opening a bank account, getting a mortgage or an insurance policy, authenticating clearance at work, or securing online payments.

Keywords: artificial intelligence, neurofinance, neuropsychology, risk management

Procedia PDF Downloads 138
27388 Pricing the Risk Associated to Weather of Variable Renewable Energy Generation

Authors: Jorge M. Uribe

Abstract:

We propose a methodology for setting the price of an insurance contract targeted to manage the risk associated with weather conditions that affect variable renewable energy generation. The methodology relies on conditional quantile regressions to estimate the weather risk of a solar panel. It is illustrated using real daily radiation and weather data for three cities in Spain (Valencia, Barcelona and Madrid) from February 2/2004 to January 22/2019. We also adapt the concepts of value at risk and expected short fall from finance to this context, to provide a complete panorama of what we label as weather risk. The methodology is easy to implement and can be used by insurance companies to price a contract with the aforementioned characteristics when data about similar projects and accurate cash flow projections are lacking. Our methodology assigns a higher price to an insurance product with the stated characteristics in Madrid, compared to Valencia and Barcelona. This is consistent with Madrid showing the largest interquartile range of operational deficits and it is unrelated to the average value deficit, which illustrates the importance of our proposal.

Keywords: insurance, weather, vre, risk

Procedia PDF Downloads 148
27387 Pain Management Program in Helping Community-Dwelling Older Adults and Their Informal Caregivers to Manage Pain and Related Situations

Authors: Mimi My Tse

Abstract:

The prevalence of chronic non-cancer pain is high among community-dwelling older adults. Pain affects physical and psychosocial abilities. Older adults tend to be less mobile and have a high tendency to fall risk. In addition, older adults with pain are depressed, anxious, and not too willing to join social activities. This will make them feel very lonely and social isolation. Instead of giving pain management education and programs to older adults/clients, both older adults and their caregivers, it is sad to find that the majority of existing services are given to older adults only. Given the importance of family members in increasing compliance with health-promoting programs, we proposed to offer pain management programs to both older adults with his/her caregiver as a “dyad.” We used the Health Promotion Model and implemented a dyadic pain management program (DPM). The DPM is an 8-week group-based program. The DPM comprises 4 weeks of center-based, face-to-face activities and 4 weeks of digital-based activities delivered via a WhatsApp group. There were 30 dyads (15 in the experimental group with DPM and 15 in the control group with pain education pamphlets). Upon the completion of DPM, pain intensity and pain interference were significantly lower in the intervention group as compared to the control group. At the same time, physical function showed significant improvement and lower depression scores in the intervention group. In conclusion, the study highlights the potential benefits of involving caregivers in the management of chronic pain for older adults. This approach should be widely promoted in managing chronic pain situations for community-dwelling older adults and their caregivers.

Keywords: pain, older adults, dyadic approach, education

Procedia PDF Downloads 78
27386 Distribution of Maximum Loss of Fractional Brownian Motion with Drift

Authors: Ceren Vardar Acar, Mine Caglar

Abstract:

In finance, the price of a volatile asset can be modeled using fractional Brownian motion (fBm) with Hurst parameter H>1/2. The Black-Scholes model for the values of returns of an asset using fBm is given as, 〖Y_t=Y_0 e^((r+μ)t+σB)〗_t^H, 0≤t≤T where Y_0 is the initial value, r is constant interest rate, μ is constant drift and σ is constant diffusion coefficient of fBm, which is denoted by B_t^H where t≥0. Black-Scholes model can be constructed with some Markov processes such as Brownian motion. The advantage of modeling with fBm to Markov processes is its capability of exposing the dependence between returns. The real life data for a volatile asset display long-range dependence property. For this reason, using fBm is a more realistic model compared to Markov processes. Investors would be interested in any kind of information on the risk in order to manage it or hedge it. The maximum possible loss is one way to measure highest possible risk. Therefore, it is an important variable for investors. In our study, we give some theoretical bounds on the distribution of maximum possible loss of fBm. We provide both asymptotical and strong estimates for the tail probability of maximum loss of standard fBm and fBm with drift and diffusion coefficients. In the investment point of view, these results explain, how large values of possible loss behave and its bounds.

Keywords: maximum drawdown, maximum loss, fractional brownian motion, large deviation, Gaussian process

Procedia PDF Downloads 483
27385 Correlations between Obesity Indices and Cardiometabolic Risk Factors in Obese Subgroups in Severely Obese Women

Authors: Seung Hun Lee, Sang Yeoup Lee

Abstract:

Objectives: To investigate associations between degrees of obesity using correlations between obesity indices and cardiometabolic risk factors. Methods: BMI, waist circumference (WC), fasting insulin, fasting glucose, lipids, and visceral adipose tissue (VAT) area using computed tomographic images were measured in 113 obese female without cardiovascular disease (CVD). Correlations between obesity indices and cardiometabolic risk factors were analyzed in obese subgroups defined using sequential obesity indices. Results: Mean BMI and WC were 29.6 kg/m2 and 92.8 cm. BMI showed significant correlations with all five cardiometabolic risk factors until the BMI cut-off point reached 27 kg/m2, but when it exceeded 30 kg/m2, correlations no longer existed. WC was significantly correlated with all five cardiometabolic risk factors up to a value of 85 cm, but when WC exceeded 90 cm, correlations no longer existed. Conclusions: Our data suggest that moderate weight-loss goals may not be enough to ameliorate cardiometabolic markers in severely obese patients. Therefore, individualized weight-loss goals should be recommended to such patients to improve health benefits.

Keywords: correlation, cardiovascular disease, risk factors, obesity

Procedia PDF Downloads 357
27384 Measuring Flood Risk concerning with the Flood Protection Embankment in Big Flooding Events of Dhaka Metropolitan Zone

Authors: Marju Ben Sayed, Shigeko Haruyama

Abstract:

Among all kinds of natural disaster, the flood is a common feature in rapidly urbanizing Dhaka city. In this research, assessment of flood risk of Dhaka metropolitan area has been investigated by using an integrated approach of GIS, remote sensing and socio-economic data. The purpose of the study is to measure the flooding risk concerning with the flood protection embankment in big flooding events (1988, 1998 and 2004) and urbanization of Dhaka metropolitan zone. In this research, we considered the Dhaka city into two parts; East Dhaka (outside the flood protection embankment) and West Dhaka (inside the flood protection embankment). Using statistical data, we explored the socio-economic status of the study area population by comparing the density of population, land price and income level. We have drawn the cross section profile of the flood protection embankment into three different points for realizing the flooding risk in the study area, especially in the big flooding year (1988, 1998 and 2004). According to the physical condition of the study area, the land use/land cover map has been classified into five classes. Comparing with each land cover unit, historical weather station data and the socio-economic data, the flooding risk has been evaluated. Moreover, we compared between DEM data and each land cover units to find out the relationship with flood. It is expected that, this study could contribute to effective flood forecasting, relief and emergency management for a future flood event in Dhaka city.

Keywords: land use, land cover change, socio-economic, Dhaka city, GIS, flood

Procedia PDF Downloads 296
27383 Normalizing Logarithms of Realized Volatility in an ARFIMA Model

Authors: G. L. C. Yap

Abstract:

Modelling realized volatility with high-frequency returns is popular as it is an unbiased and efficient estimator of return volatility. A computationally simple model is fitting the logarithms of the realized volatilities with a fractionally integrated long-memory Gaussian process. The Gaussianity assumption simplifies the parameter estimation using the Whittle approximation. Nonetheless, this assumption may not be met in the finite samples and there may be a need to normalize the financial series. Based on the empirical indices S&P500 and DAX, this paper examines the performance of the linear volatility model pre-treated with normalization compared to its existing counterpart. The empirical results show that by including normalization as a pre-treatment procedure, the forecast performance outperforms the existing model in terms of statistical and economic evaluations.

Keywords: Gaussian process, long-memory, normalization, value-at-risk, volatility, Whittle estimator

Procedia PDF Downloads 354
27382 Bio-Hub Ecosystems: Investment Risk Analysis Using Monte Carlo Techno-Economic Analysis

Authors: Kimberly Samaha

Abstract:

In order to attract new types of investors into the emerging Bio-Economy, new methodologies to analyze investment risk are needed. The Bio-Hub Ecosystem model was developed to address a critical area of concern within the global energy market regarding the use of biomass as a feedstock for power plants. This study looked at repurposing existing biomass-energy plants into Circular Zero-Waste Bio-Hub Ecosystems. A Bio-Hub model that first targets a ‘whole-tree’ approach and then looks at the circular economics of co-hosting diverse industries (wood processing, aquaculture, agriculture) in the vicinity of the Biomass Power Plants facilities. This study modeled the economics and risk strategies of cradle-to-cradle linkages to incorporate the value-chain effects on capital/operational expenditures and investment risk reductions using a proprietary techno-economic model that incorporates investment risk scenarios utilizing the Monte Carlo methodology. The study calculated the sequential increases in profitability for each additional co-host on an operating forestry-based biomass energy plant in West Enfield, Maine. Phase I starts with the base-line of forestry biomass to electricity only and was built up in stages to include co-hosts of a greenhouse and a land-based shrimp farm. Phase I incorporates CO2 and heat waste streams from the operating power plant in an analysis of lowering and stabilizing the operating costs of the agriculture and aquaculture co-hosts. Phase II analysis incorporated a jet-fuel biorefinery and its secondary slip-stream of biochar which would be developed into two additional bio-products: 1) A soil amendment compost for agriculture and 2) A biochar effluent filter for the aquaculture. The second part of the study applied the Monte Carlo risk methodology to illustrate how co-location derisks investment in an integrated Bio-Hub versus individual investments in stand-alone projects of energy, agriculture or aquaculture. The analyzed scenarios compared reductions in both Capital and Operating Expenditures, which stabilizes profits and reduces the investment risk associated with projects in energy, agriculture, and aquaculture. The major findings of this techno-economic modeling using the Monte Carlo technique resulted in the masterplan for the first Bio-Hub to be built in West Enfield, Maine. In 2018, the site was designated as an economic opportunity zone as part of a Federal Program, which allows for Capital Gains tax benefits for investments on the site. Bioenergy facilities are currently at a critical juncture where they have an opportunity to be repurposed into efficient, profitable and socially responsible investments, or be idled and scrapped. The Bio-hub Ecosystems techno-economic analysis model is a critical model to expedite new standards for investments in circular zero-waste projects. Profitable projects will expedite adoption and advance the critical transition from the current ‘take-make-dispose’ paradigm inherent in the energy, forestry and food industries to a more sustainable Bio-Economy paradigm that supports local and rural communities.

Keywords: bio-economy, investment risk, circular design, economic modelling

Procedia PDF Downloads 101
27381 Adherence to Dietary Approaches to Stop Hypertension-Style Diet and Risk of Mortality from Cancer: A Systematic Review and Meta-Analysis of Cohort Studies

Authors: Roohallah Fallah-Moshkani, Mohammad Ali Mohsenpour, Reza Ghiasvand, Hossein Khosravi-Boroujeni, Seyed Mehdi Ahmadi, Paula Brauer, Amin Salehi-Abargouei

Abstract:

Purpose: Several investigations have proposed the protective association between dietary approaches to stop hypertension (DASH) style diet and risk of cancers; however, they have led to inconsistent results. The present study aimed to systematically review the prospective cohort studies conducted in this regard and, if possible, to quantify the overall effect of using meta-analysis. Methods: PubMed, EMBASE, Scopus, and Google Scholar were searched for cohort studies published up to December 2017. Relative risks (RRs) which were reported for fully adjusted models and their confidence intervals were extracted for meta-analysis. Random effects model was incorporated to combine the RRs. Results: Sixteen studies were eligible to be included in the systematic review from which 8 reports were conducted on the effect of DASH on the risk of mortality from all cancer types, four on the risk of colorectal cancer, and three on the risk of colon and rectal cancer. Four studies examined the association with other cancers (breast, hepatic, endometrial, and lung cancer). Meta-analysis showed that high concordance with DASH significantly decreases the risk of all cancer types (RR=0.83, 95% confidence interval (95%CI):0.80-0.85); furthermore participants who highly adhered to the DASH had lower risk of developing colorectal (RR=0.79, 95%CI: 0.75-0.83), colon (RR=0.81, 95%CI: 0.74-0.87) and rectal (RR=0.79, 95%CI: 0.63-0.98) cancer compared to those with the lowest adherence. Conclusions: DASH-style diet should be suggested as a healthy approach to protect from cancer in the community. Prospective studies exploring the effect on other cancer types and from regions other than the United States are highly recommended.

Keywords: cancer, DASH-style diet, dietary patterns, meta-analysis, systematic review

Procedia PDF Downloads 188
27380 Modeling Driving Distraction Considering Psychological-Physical Constraints

Authors: Yixin Zhu, Lishengsa Yue, Jian Sun, Lanyue Tang

Abstract:

Modeling driving distraction in microscopic traffic simulation is crucial for enhancing simulation accuracy. Current driving distraction models are mainly derived from physical motion constraints under distracted states, in which distraction-related error terms are added to existing microscopic driver models. However, the model accuracy is not very satisfying, due to a lack of modeling the cognitive mechanism underlying the distraction. This study models driving distraction based on the Queueing Network Human Processor model (QN-MHP). This study utilizes the queuing structure of the model to perform task invocation and switching for distracted operation and control of the vehicle under driver distraction. Based on the assumption of the QN-MHP model about the cognitive sub-network, server F is a structural bottleneck. The latter information must wait for the previous information to leave server F before it can be processed in server F. Therefore, the waiting time for task switching needs to be calculated. Since the QN-MHP model has different information processing paths for auditory information and visual information, this study divides driving distraction into two types: auditory distraction and visual distraction. For visual distraction, both the visual distraction task and the driving task need to go through the visual perception sub-network, and the stimuli of the two are asynchronous, which is called stimulus on asynchrony (SOA), so when calculating the waiting time for switching tasks, it is necessary to consider it. In the case of auditory distraction, the auditory distraction task and the driving task do not need to compete for the server resources of the perceptual sub-network, and their stimuli can be synchronized without considering the time difference in receiving the stimuli. According to the Theory of Planned Behavior for drivers (TPB), this study uses risk entropy as the decision criterion for driver task switching. A logistic regression model is used with risk entropy as the independent variable to determine whether the driver performs a distraction task, to explain the relationship between perceived risk and distraction. Furthermore, to model a driver’s perception characteristics, a neurophysiological model of visual distraction tasks is incorporated into the QN-MHP, and executes the classical Intelligent Driver Model. The proposed driving distraction model integrates the psychological cognitive process of a driver with the physical motion characteristics, resulting in both high accuracy and interpretability. This paper uses 773 segments of distracted car-following in Shanghai Naturalistic Driving Study data (SH-NDS) to classify the patterns of distracted behavior on different road facilities and obtains three types of distraction patterns: numbness, delay, and aggressiveness. The model was calibrated and verified by simulation. The results indicate that the model can effectively simulate the distracted car-following behavior of different patterns on various roadway facilities, and its performance is better than the traditional IDM model with distraction-related error terms. The proposed model overcomes the limitations of physical-constraints-based models in replicating dangerous driving behaviors, and internal characteristics of an individual. Moreover, the model is demonstrated to effectively generate more dangerous distracted driving scenarios, which can be used to construct high-value automated driving test scenarios.

Keywords: computational cognitive model, driving distraction, microscopic traffic simulation, psychological-physical constraints

Procedia PDF Downloads 91
27379 Beyond Classic Program Evaluation and Review Technique: A Generalized Model for Subjective Distributions with Flexible Variance

Authors: Byung Cheol Kim

Abstract:

The Program Evaluation and Review Technique (PERT) is widely used for project management, but it struggles with subjective distributions, particularly due to its assumptions of constant variance and light tails. To overcome these limitations, we propose the Generalized PERT (G-PERT) model, which enhances PERT by incorporating variability in three-point subjective estimates. Our methodology extends the original PERT model to cover the full range of unimodal beta distributions, enabling the model to handle thick-tailed distributions and offering formulas for computing mean and variance. This maintains the simplicity of PERT while providing a more accurate depiction of uncertainty. Our empirical analysis demonstrates that the G-PERT model significantly improves performance, particularly when dealing with heavy-tail subjective distributions. In comparative assessments with alternative models such as triangular and lognormal distributions, G-PERT shows superior accuracy and flexibility. These results suggest that G-PERT offers a more robust solution for project estimation while still retaining the user-friendliness of the classic PERT approach.

Keywords: PERT, subjective distribution, project management, flexible variance

Procedia PDF Downloads 18