Search results for: adopted methods
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 17684

Search results for: adopted methods

16394 Investigation and Comprehensive Benefit Analysis of 11 Typical Polar-Based Agroforestry Models Based on Analytic Hierarchy Process in Anhui Province, Eastern China

Authors: Zhihua Cao, Hongfei Zhao, Zhongneng Wu

Abstract:

The development of polar-based agroforestry was necessary due to the influence of the timber market environment in China, which can promote the coordinated development of forestry and agriculture, and gain remarkable ecological, economic and social benefits. The main agroforestry models of the main poplar planting area in Huaibei plain and along the Yangtze River plain were carried out. 11 typical management models of poplar were selected to sum up: pure poplar forest, poplar-rape-soybean, poplar-wheat-soybean, poplar-rape-cotton, poplar-wheat, poplar-chicken, poplar-duck, poplar-sheep, poplar-Agaricus blazei, poplar-oil peony, poplar-fish, represented by M0-M10, respectively. 12 indexes related with economic, ecological and social benefits (annual average cost, net income, ratio of output to investment, payback period of investment, land utilization ratio, utilization ratio of light energy, improvement and system stability of ecological and production environment, product richness, labor capacity, cultural quality of labor force, sustainability) were screened out to carry on the comprehensive evaluation and analysis to 11 kinds of typical agroforestry models based on analytic hierarchy process (AHP). The results showed that the economic benefit of each agroforestry model was in the order of: M8 > M6 > M9 > M7 > M5 > M10 > M4 > M1 > M2 > M3 > M0. The economic benefit of poplar-A. blazei model was the highest (332, 800 RMB / hm²), followed by poplar-duck and poplar-oil peony model (109, 820RMB /hm², 5, 7226 RMB /hm²). The order of comprehensive benefit was: M8 > M4 > M9 > M6 > M1 > M2 > M3 > M7 > M5 > M10 > M0. The economic benefit and comprehensive benefit of each agroforestry model were higher than that of pure poplar forest. The comprehensive benefit of poplar-A. blazei model was the highest, and that of poplar-wheat model ranked second, while its economic benefit was not high. Next were poplar-oil peony and poplar-duck models. It was suggested that the model of poplar-wheat should be adopted in the plain along the Yangtze River, and the whole cycle mode of poplar-grain, popalr-A. blazei, or poplar-oil peony should be adopted in Huaibei plain, northern Anhui. Furthermore, wheat, rape, and soybean are the main crops before the stand was closed; the agroforestry model of edible fungus or Chinese herbal medicine can be carried out when the stand was closed in order to maximize the comprehensive benefit. The purpose of this paper is to provide a reference for forest farmers in the selection of poplar agroforestry model in the future and to provide the basic data for the sustainable and efficient study of poplar agroforestry in Anhui province, eastern China.

Keywords: agroforestry, analytic hierarchy process (AHP), comprehensive benefit, model, poplar

Procedia PDF Downloads 165
16393 BIM Modeling of Site and Existing Buildings: Case Study of ESTP Paris Campus

Authors: Rita Sassine, Yassine Hassani, Mohamad Al Omari, Stéphanie Guibert

Abstract:

Building Information Modelling (BIM) is the process of creating, managing, and centralizing information during the building lifecycle. BIM can be used all over a construction project, from the initiation phase to the planning and execution phases to the maintenance and lifecycle management phase. For existing buildings, BIM can be used for specific applications such as lifecycle management. However, most of the existing buildings don’t have a BIM model. Creating a compatible BIM for existing buildings is very challenging. It requires special equipment for data capturing and efforts to convert these data into a BIM model. The main difficulties for such projects are to define the data needed, the level of development (LOD), and the methodology to be adopted. In addition to managing information for an existing building, studying the impact of the built environment is a challenging topic. So, integrating the existing terrain that surrounds buildings into the digital model is essential to be able to make several simulations as flood simulation, energy simulation, etc. Making a replication of the physical model and updating its information in real-time to make its Digital Twin (DT) is very important. The Digital Terrain Model (DTM) represents the ground surface of the terrain by a set of discrete points with unique height values over 2D points based on reference surface (e.g., mean sea level, geoid, and ellipsoid). In addition, information related to the type of pavement materials, types of vegetation and heights and damaged surfaces can be integrated. Our aim in this study is to define the methodology to be used in order to provide a 3D BIM model for the site and the existing building based on the case study of “Ecole Spéciale des Travaux Publiques (ESTP Paris)” school of engineering campus. The property is located on a hilly site of 5 hectares and is composed of more than 20 buildings with a total area of 32 000 square meters and a height between 50 and 68 meters. In this work, the campus precise levelling grid according to the NGF-IGN69 altimetric system and the grid control points are computed according to (Réseau Gédésique Français) RGF93 – Lambert 93 french system with different methods: (i) Land topographic surveying methods using robotic total station, (ii) GNSS (Global Network Satellite sytem) levelling grid with NRTK (Network Real Time Kinematic) mode, (iii) Point clouds generated by laser scanning. These technologies allow the computation of multiple building parameters such as boundary limits, the number of floors, the floors georeferencing, the georeferencing of the 4 base corners of each building, etc. Once the entry data are identified, the digital model of each building is done. The DTM is also modeled. The process of altimetric determination is complex and requires efforts in order to collect and analyze multiple data formats. Since many technologies can be used to produce digital models, different file formats such as DraWinG (DWG), LASer (LAS), Comma-separated values (CSV), Industry Foundation Classes (IFC) and ReViT (RVT) will be generated. Checking the interoperability between BIM models is very important. In this work, all models are linked together and shared on 3DEXPERIENCE collaborative platform.

Keywords: building information modeling, digital terrain model, existing buildings, interoperability

Procedia PDF Downloads 112
16392 Contextualizing Theory Z of Motivation Among Indian Universities of Higher Education

Authors: Janani V., Tanika Singh, Bala Subramanian R., Santosh Kumar Sharma

Abstract:

Higher education across the globe is undergoing a sea change. This has created a varied management of higher education in Indian universities, and therefore, we find no universal law regarding HR policies and practices in these universities. As a result, faculty retention is very low, which is a serious concern for educational leaders such as vice-chancellors or directors working in the higher education sector. We can understand this phenomenon in the light of various management theories, among which theory z proposed by William Ouchi is a prominent one. With this backdrop, the present article strives to contextualize theory z in Indian higher education. For the said purpose, qualitative methodology has been adopted, and accordingly, propositions have been generated. We believe that this article will motivate other researchers to empirically test the generated propositions and thereby contribute in the existing literature.

Keywords: education, managemenet, motivation, Theory X, Theory Y, Theory Z, faculty members, universities, India

Procedia PDF Downloads 114
16391 Diagnosis of Choledocholithiasis with Endosonography

Authors: A. Kachmazova, A. Shadiev, Y. Teterin, P. Yartcev

Abstract:

Introduction: Biliary calculi disease (LCS) still occupies the leading position among urgent diseases of the abdominal cavity, manifesting itself from asymptomatic course to life-threatening states. Nowadays arsenal of diagnostic methods for choledocholithiasis is quite wide: ultrasound, hepatobiliscintigraphy (HBSG), magnetic resonance imaging (MRI), endoscopic retrograde cholangiography (ERCP). Among them, transabdominal ultrasound (TA ultrasound) is the most accessible and routine diagnostic method. Nowadays ERCG is the "gold" standard in diagnosis and one-stage treatment of biliary tract obstruction. However, transpapillary techniques are accompanied by serious postoperative complications (postmanipulative pancreatitis (3-5%), endoscopic papillosphincterotomy bleeding (2%), cholangitis (1%)), the lethality being 0.4%. GBSG and MRI are also quite informative methods in the diagnosis of choledocholithiasis. Small size of concrements, their localization in intrapancreatic and retroduodenal part of common bile duct significantly reduces informativity of all diagnostic methods described above, that demands additional studying of this problem. Materials and Methods: 890 patients with the diagnosis of cholelithiasis (calculous cholecystitis) were admitted to the Sklifosovsky Scientific Research Institute of Hospital Medicine in the period from August, 2020 to June, 2021. Of them 115 people with mechanical jaundice caused by concrements in bile ducts. Results: Final EUS diagnosis was made in all patients (100,0%). In all patients in whom choledocholithiasis diagnosis was revealed or confirmed after EUS, ERCP was performed urgently (within two days from the moment of its detection) as the X-ray operation room was provided; it confirmed the presence of concrements. All stones were removed by lithoextraction using Dormia basket. The postoperative period in these patients had no complications. Conclusions: EUS is the most informative and safe diagnostic method, which allows to detect choledocholithiasis in patients with discrepancies between clinical-laboratory and instrumental methods of diagnosis in shortest time, that in its turn will help to decide promptly on the further tactics of patient treatment. We consider it reasonable to include EUS in the diagnostic algorithm for choledocholithiasis. Disclosure: Nothing to disclose.

Keywords: endoscopic ultrasonography, choledocholithiasis, common bile duct, concrement, ERCP

Procedia PDF Downloads 85
16390 Italian Central Guarantee Fund: An Analysis of the Guaranteed SMEs’ Default Risk

Authors: M. C. Arcuri, L. Gai, F. Ielasi

Abstract:

Italian Central Guarantee Fund (CGF) has the purpose to facilitate Small and Medium-sized Enterprises (SMEs)’ access to credit. The aim of the paper is to study the evaluation method adopted by the CGF with regard to SMEs requiring its intervention. This is even more important in the light of the recent CGF reform. We analyse an initial sample of more than 500.000 guarantees from 2012 to 2018. We distinguish between a counter-guarantee delivered to a mutual guarantee institution and a guarantee directly delivered to a bank. We investigate the impact of variables related to the operations and the SMEs on Altman Z’’-score and the score consistent with CGF methodology. We verify that the type of intervention affects the scores and the initial condition changes with the new assessment criterions. 

Keywords: banks, default risk, Italian guarantee fund, mutual guarantee institutions

Procedia PDF Downloads 175
16389 Research Study on the Environmental Conditions in the Foreign

Authors: Vahid Bairami Rad, Shapoor Norazar, Moslem Talebi Asl

Abstract:

The fast growing accessibility and capability of emerging technologies have fashioned enormous possibilities of designing, developing and implementing innovative teaching methods in the classroom. Using teaching methods and technology together have a fantastic results, because the global technological scenario has paved the way to new pedagogies in teaching-learning process. At the other side methods by focusing on students and the ways of learning in them, that can demonstrate logical ways of improving student achievement in English as a foreign language in Iran. The sample of study was 90 students of 10th grade of high school located in Ardebil. A pretest-posttest equivalent group designed to compare the achievement of groups. Students divided to 3 group, Control base, computer base, method and technology base. Pretest and post test contain 30 items each from English textbook were developed and administrated, then obtained data were analyzed. The results showed that there was an important difference. The 3rd group performance was better than other groups. On the basis of this result it was obviously counseled that teaching-learning capabilities.

Keywords: method, technology based environment, computer based environment, english as a foreign language, student achievement

Procedia PDF Downloads 474
16388 The Modelling of Real Time Series Data

Authors: Valeria Bondarenko

Abstract:

We proposed algorithms for: estimation of parameters fBm (volatility and Hurst exponent) and for the approximation of random time series by functional of fBm. We proved the consistency of the estimators, which constitute the above algorithms, and proved the optimal forecast of approximated time series. The adequacy of estimation algorithms, approximation, and forecasting is proved by numerical experiment. During the process of creating software, the system has been created, which is displayed by the hierarchical structure. The comparative analysis of proposed algorithms with the other methods gives evidence of the advantage of approximation method. The results can be used to develop methods for the analysis and modeling of time series describing the economic, physical, biological and other processes.

Keywords: mathematical model, random process, Wiener process, fractional Brownian motion

Procedia PDF Downloads 358
16387 Effect of Building Construction Sizes on Project Delivery Methods in Nigeria

Authors: Nuruddeen Usman, Mohammad Sani

Abstract:

The performance of project delivery methods has been an issue of concern to various stakeholders in the construction industry. The contracting system of project delivery is the traditional system used in the delivery of most public projects in Nigeria. The direct labor system is used most times as an alternative to the traditional system. There were so many complain about the performance of contracting system and the suitability of direct labor as an alternative to the delivery of public projects. Therefore, this paper is aimed at investigating the effect of project size on the project delivery methods in the completed public buildings. Questionnaires were self-administered to managerial staff in the study area and analyzed using descriptive statistics. The findings reveals that contracting system was choosing for large size building construction project delivery with higher frequency (F) of 40 (76.9%) against direct labor with 12 (23.1%). While the small size project, the result revealed a frequency (F) of 26 (50%) for contracting system and direct labor system respectively. Base on the research findings, the contracting system, was recommended for all sizes of building construction project delivery while direct labor system can only use as an alternative for small size building construction projects delivery.

Keywords: construction size, contracting system, direct labour, effect

Procedia PDF Downloads 457
16386 A Comparison Between Different Discretization Techniques for the Doyle-Fuller-Newman Li+ Battery Model

Authors: Davide Gotti, Milan Prodanovic, Sergio Pinilla, David Muñoz-Torrero

Abstract:

Since its proposal, the Doyle-Fuller-Newman (DFN) lithium-ion battery model has gained popularity in the electrochemical field. In fact, this model provides the user with theoretical support for designing the lithium-ion battery parameters, such as the material particle or the diffusion coefficient adjustment direction. However, the model is mathematically complex as it is composed of several partial differential equations (PDEs) such as Fick’s law of diffusion, the MacInnes and Ohm’s equations, among other phenomena. Thus, to efficiently use the model in a time-domain simulation environment, the selection of the discretization technique is of a pivotal importance. There are several numerical methods available in the literature that can be used to carry out this task. In this study, a comparison between the explicit Euler, Crank-Nicolson, and Chebyshev discretization methods is proposed. These three methods are compared in terms of accuracy, stability, and computational times. Firstly, the explicit Euler discretization technique is analyzed. This method is straightforward to implement and is computationally fast. In this work, the accuracy of the method and its stability properties are shown for the electrolyte diffusion partial differential equation. Subsequently, the Crank-Nicolson method is considered. It represents a combination of the implicit and explicit Euler methods that has the advantage of being of the second order in time and is intrinsically stable, thus overcoming the disadvantages of the simpler Euler explicit method. As shown in the full paper, the Crank-Nicolson method provides accurate results when applied to the DFN model. Its stability does not depend on the integration time step, thus it is feasible for both short- and long-term tests. This last remark is particularly important as this discretization technique would allow the user to implement parameter estimation and optimization techniques such as system or genetic parameter identification methods using this model. Finally, the Chebyshev discretization technique is implemented in the DFN model. This discretization method features swift convergence properties and, as other spectral methods used to solve differential equations, achieves the same accuracy with a smaller number of discretization nodes. However, as shown in the literature, these methods are not suitable for handling sharp gradients, which are common during the first instants of the charge and discharge phases of the battery. The numerical results obtained and presented in this study aim to provide the guidelines on how to select the adequate discretization technique for the DFN model according to the type of application to be performed, highlighting the pros and cons of the three methods. Specifically, the non-eligibility of the simple Euler method for longterm tests will be presented. Afterwards, the Crank-Nicolson and the Chebyshev discretization methods will be compared in terms of accuracy and computational times under a wide range of battery operating scenarios. These include both long-term simulations for aging tests, and short- and mid-term battery charge/discharge cycles, typically relevant in battery applications like grid primary frequency and inertia control and electrical vehicle breaking and acceleration.

Keywords: Doyle-Fuller-Newman battery model, partial differential equations, discretization, numerical methods

Procedia PDF Downloads 24
16385 Stochastic Default Risk Estimation Evidence from the South African Financial Market

Authors: Mesias Alfeus, Kirsty Fitzhenry, Alessia Lederer

Abstract:

The present paper provides empirical studies to estimate defaultable bonds in the South African financial market. The main goal is to estimate the unobservable factors affecting bond yields for South African major banks. The maximum likelihood approach is adopted for the estimation methodology. Extended Kalman filtering techniques are employed in order to tackle the situation that the factors cannot be observed directly. Multi-dimensional Cox-Ingersoll-Ross (CIR)-type factor models are considered. Results show that default risk increased sharply in the South African financial market during COVID-19 and the CIR model with jumps exhibits a better performance.

Keywords: default intensity, unobservable state variables, CIR, α-CIR, extended kalman filtering

Procedia PDF Downloads 111
16384 Developing and Managing an Institutional Repository in a Nigerian University Library: The Futa Experience

Authors: Belau Olatunde Gbadamosi, Oluchi Okere

Abstract:

Spurred by the ease of access to and the cost-effectiveness of open-source software such as DSpace, EPrints, and Greenstone Digital Libraries for hosting digital content, many libraries have added institutional repositories (IRs) to their repertoire of digital assets. This paper adopts a qualitative approach based on focus group discussions and the system development life cycle model (SDLC) to describe the experience of Albert Ilemobade Library (the Federal University of Technology Akure, Nigeria (FUTA) in the development of their IR - FUTASpace. Peculiar challenges experienced in the course of the development and solutions adopted are also reported. This study will serve as a reference point to other institutions, particularly those operating in developing countries, which may be poorly funded.

Keywords: institutional repository, digital libraries, university libraries, DSpace

Procedia PDF Downloads 175
16383 Design of Black-Seed Pulp biomass-Derived New Bio-Sorbent by Combining Methods of Mineral Acids and High-Temperature for Arsenic Removal

Authors: Mozhgan Mohammadi, Arezoo Ghadi

Abstract:

Arsenic is known as a potential threat to the environment. Therefore, the aim of this research is to assess the arsenic removal efficiency from an aqueous solution, with a new biosorbent composed of a black seed pulp (BSP). To treat BSP, the combination of two methods (i.e. treating with mineral acids and use at high temperature) was used and designed bio-sorbent called BSP-activated/carbonized. The BSP-activated and BSP-carbonized were also prepared using HCL and 400°C temperature, respectively, to compare the results of each three methods. Followed by, adsorption parameters such as pH, initial ion concentration, biosorbent dosage, contact time, and temperature were assessed. It was found that the combination method has provided higher adsorption capacity so that up to ~99% arsenic removal was observed with BSP-activated/carbonized at pH of 7.0 and 40°C. The adsorption capacity for BSP-carbonized and BSP-activated were 87.92% (pH: 7, 60°C) and 78.50% (pH: 6, 90°C), respectively. Moreover, adsorption kinetics data indicated the best fit with the pseudo-second-order model. The maximum biosorption capacity, by the Langmuir isotherm model, was also recorded for BSP-activated/carbonized (53.47 mg/g). It is notable that arsenic adsorption on studied bio sorbents takes place as spontaneous and through chemisorption along with the endothermic nature of the biosorption process and reduction of random collision in the solid-liquid phase.

Keywords: black seed pulp, bio-sorbents, treatment of sorbents, adsorption isotherms

Procedia PDF Downloads 95
16382 Structural Transformation after 2000 in Turkey Economy Evaluation as Theoretical in the Context of Inflation and Foreign Trade

Authors: Sadife Güngör, Sevilay Konya, Zeynep Karaçor

Abstract:

Inflation and foreign trade are the most important economic indicator of a country. In this study, Turkey's economy with the policies adopted after 2000, given how performs an economic transformation. This transformation of the economy is discussed with inflation and foreign trade. In this context, attention is drawn to 2001 Strong Economy and Transition Program and 2006 Inflation Targeting Regime. The evaluation was performed of after the year 2000 inflation and foreign trade figures in Turkey economy. When we looked the progress, after 2000 in Turkey economy, we can say a new process was built up.

Keywords: inflation, foreign trade, 2001 strong economy programme, 2006 inflation targeting regime

Procedia PDF Downloads 357
16381 Development of Chronic Obstructive Pulmonary Disease (COPD) Proforma (E-ICP) to Improve Guideline Adherence in Emergency Department: Modified Delphi Study

Authors: Hancy Issac, Gerben Keijzers, Ian Yang, Clint Moloney, Jackie Lea, Melissa Taylor

Abstract:

Introduction: Chronic obstructive pulmonary disease guideline non-adherence is associated with a reduction in health-related quality of life in patients (HRQoL). Improving guideline adherence has the potential to mitigate fragmented care thereby sustaining pulmonary function, preventing acute exacerbations, reducing economic health burdens, and enhancing HRQoL. The development of an electronic proforma stemming from expert consensus, including digital guideline resources and direct interdisciplinary referrals is hypothesised to improve guideline adherence and patient outcomes for emergency department (ED) patients with COPD. Aim: The aim of this study was to develop consensus among ED and respiratory staff for the correct composition of a COPD electronic proforma that aids in guideline adherence and management in the ED. Methods: This study adopted a mixed-method design to develop the most important indicators of care in the ED. The study involved three phases: (1) a systematic literature review and qualitative interdisciplinary staff interviews to assess barriers and solutions for guideline adherence and qualitative interdisciplinary staff interviews, (2) a modified Delphi panel to select interventions for the proforma, and (3) a consensus process through three rounds of scoring through a quantitative survey (ED and Respiratory consensus) and qualitative thematic analysis on each indicator. Results: The electronic proforma achieved acceptable and good internal consistency through all iterations from national emergency department and respiratory department interdisciplinary experts. Cronbach’s alpha score for internal consistency (α) in iteration 1 emergency department cohort (EDC) (α = 0.80 [CI = 0.89%]), respiratory department cohort (RDC) (α = 0.95 [CI = 0.98%]). Iteration 2 reported EDC (α = 0.85 [CI = 0.97%]) and RDC (α = 0.86 [CI = 0.97%]). Iteration 3 revealed EDC (α = 0.73 [CI = 0.91%]) and RDC (α = 0.86 [CI = 0.95%]), respectively. Conclusion: Electronic proformas have the potential to facilitate direct referrals from the ED leading to reduced hospital admissions, reduced length of hospital stays, holistic care, improved health care and quality of life and improved interdisciplinary guideline adherence.

Keywords: COPD, electronic proforma, modified delphi study, interdisciplinary, guideline adherence, COPD-X plan

Procedia PDF Downloads 60
16380 Open Science Philosophy, Research and Innovation

Authors: C.Ardil

Abstract:

Open Science translates the understanding and application of various theories and practices in open science philosophy, systems, paradigms and epistemology. Open Science originates with the premise that universal scientific knowledge is a product of a collective scholarly and social collaboration involving all stakeholders and knowledge belongs to the global society. Scientific outputs generated by public research are a public good that should be available to all at no cost and without barriers or restrictions. Open Science has the potential to increase the quality, impact and benefits of science and to accelerate advancement of knowledge by making it more reliable, more efficient and accurate, better understandable by society and responsive to societal challenges, and has the potential to enable growth and innovation through reuse of scientific results by all stakeholders at all levels of society, and ultimately contribute to growth and competitiveness of global society. Open Science is a global movement to improve accessibility to and reusability of research practices and outputs. In its broadest definition, it encompasses open access to publications, open research data and methods, open source, open educational resources, open evaluation, and citizen science. The implementation of open science provides an excellent opportunity to renegotiate the social roles and responsibilities of publicly funded research and to rethink the science system as a whole. Open Science is the practice of science in such a way that others can collaborate and contribute, where research data, lab notes and other research processes are freely available, under terms that enable reuse, redistribution and reproduction of the research and its underlying data and methods. Open Science represents a novel systematic approach to the scientific process, shifting from the standard practices of publishing research results in scientific publications towards sharing and using all available knowledge at an earlier stage in the research process, based on cooperative work and diffusing scholarly knowledge with no barriers and restrictions. Open Science refers to efforts to make the primary outputs of publicly funded research results (publications and the research data) publicly accessible in digital format with no limitations. Open Science is about extending the principles of openness to the whole research cycle, fostering, sharing and collaboration as early as possible, thus entailing a systemic change to the way science and research is done. Open Science is the ongoing transition in how open research is carried out, disseminated, deployed, and transformed to make scholarly research more open, global, collaborative, creative and closer to society. Open Science involves various movements aiming to remove the barriers for sharing any kind of output, resources, methods or tools, at any stage of the research process. Open Science embraces open access to publications, research data, source software, collaboration, peer review, notebooks, educational resources, monographs, citizen science, or research crowdfunding. The recognition and adoption of open science practices, including open science policies that increase open access to scientific literature and encourage data and code sharing, is increasing in the open science philosophy. Revolutionary open science policies are motivated by ethical, moral or utilitarian arguments, such as the right to access digital research literature for open source research or science data accumulation, research indicators, transparency in the field of academic practice, and reproducibility. Open science philosophy is adopted primarily to demonstrate the benefits of open science practices. Researchers use open science applications for their own advantage in order to get more offers, increase citations, attract media attention, potential collaborators, career opportunities, donations and funding opportunities. In open science philosophy, open data findings are evidence that open science practices provide significant benefits to researchers in scientific research creation, collaboration, communication, and evaluation according to more traditional closed science practices. Open science considers concerns such as the rigor of peer review, common research facts such as financing and career development, and the sacrifice of author rights. Therefore, researchers are recommended to implement open science research within the framework of existing academic evaluation and incentives. As a result, open science research issues are addressed in the areas of publishing, financing, collaboration, resource management and sharing, career development, discussion of open science questions and conclusions.

Keywords: Open Science, Open Science Philosophy, Open Science Research, Open Science Data

Procedia PDF Downloads 132
16379 Application of Complete Ensemble Empirical Mode Decomposition with Adaptive Noise and Multipoint Optimal Minimum Entropy Deconvolution in Railway Bearings Fault Diagnosis

Authors: Yao Cheng, Weihua Zhang

Abstract:

Although the measured vibration signal contains rich information on machine health conditions, the white noise interferences and the discrete harmonic coming from blade, shaft and mash make the fault diagnosis of rolling element bearings difficult. In order to overcome the interferences of useless signals, a new fault diagnosis method combining Complete Ensemble Empirical Mode Decomposition with adaptive noise (CEEMDAN) and Multipoint Optimal Minimum Entropy Deconvolution (MOMED) is proposed for the fault diagnosis of high-speed train bearings. Firstly, the CEEMDAN technique is applied to adaptively decompose the raw vibration signal into a series of finite intrinsic mode functions (IMFs) and a residue. Compared with Ensemble Empirical Mode Decomposition (EEMD), the CEEMDAN can provide an exact reconstruction of the original signal and a better spectral separation of the modes, which improves the accuracy of fault diagnosis. An effective sensitivity index based on the Pearson's correlation coefficients between IMFs and raw signal is adopted to select sensitive IMFs that contain bearing fault information. The composite signal of the sensitive IMFs is applied to further analysis of fault identification. Next, for propose of identifying the fault information precisely, the MOMED is utilized to enhance the periodic impulses in composite signal. As a non-iterative method, the MOMED has better deconvolution performance than the classical deconvolution methods such Minimum Entropy Deconvolution (MED) and Maximum Correlated Kurtosis Deconvolution (MCKD). Third, the envelope spectrum analysis is applied to detect the existence of bearing fault. The simulated bearing fault signals with white noise and discrete harmonic interferences are used to validate the effectiveness of the proposed method. Finally, the superiorities of the proposed method are further demonstrated by high-speed train bearing fault datasets measured from test rig. The analysis results indicate that the proposed method has strong practicability.

Keywords: bearing, complete ensemble empirical mode decomposition with adaptive noise, fault diagnosis, multipoint optimal minimum entropy deconvolution

Procedia PDF Downloads 374
16378 An Analysis on the Appropriateness and Effectiveness of CCTV Location for Crime Prevention

Authors: Tae-Heon Moon, Sun-Young Heo, Sang-Ho Lee, Youn-Taik Leem, Kwang-Woo Nam

Abstract:

This study aims to investigate the possibility of crime prevention through CCTV by analyzing the appropriateness of the CCTV location, whether it is installed in the hotspot of crime-prone areas, and exploring the crime prevention effect and transition effect. The real crime and CCTV locations of case city were converted into the spatial data by using GIS. The data was analyzed by hotspot analysis and weighted displacement quotient(WDQ). As study methods, it analyzed existing relevant studies for identifying the trends of CCTV and crime studies based on big data from 1800 to 2014 and understanding the relation between CCTV and crime. Second, it investigated the current situation of nationwide CCTVs and analyzed the guidelines of CCTV installation and operation to draw attention to the problems and indicating points of domestic CCTV use. Third, it investigated the crime occurrence in case areas and the current situation of CCTV installation in the spatial aspects, and analyzed the appropriateness and effectiveness of CCTV installation to suggest a rational installation of CCTV and the strategic direction of crime prevention. The results demonstrate that there was no significant effect in the installation of CCTV on crime prevention. This indicates that CCTV should be installed and managed in a more scientific way reflecting local crime situations. In terms of CCTV, the methods of spatial analysis such as GIS, which can evaluate the installation effect, and the methods of economic analysis like cost-benefit analysis should be developed. In addition, these methods should be distributed to local governments across the nation for the appropriate installation of CCTV and operation. This study intended to find a design guideline of the optimum CCTV installation. In this regard, this study is meaningful in that it will contribute to the creation of a safe city.

Keywords: CCTV, safe city, crime prevention, spatial analysis

Procedia PDF Downloads 438
16377 Lightweight Synergy IoT Framework for Smart Home Healthcare for the Elderly

Authors: Huawei Ma, Wencai Du, Shengbin Liang

Abstract:

Smart Home Healthcare technologies for the elderly represent a transformative paradigm that leverages emerging technologies to provide the elderly’ health indicators and daily life monitoring, emergency calls, environmental monitoring, behavior perception, and other services to ensure the health and safety of the elderly who are aging in their own home. However, the excessive complexity in the main adopted framework has affected the acceptance and adoption of the elderly. Therefore, this paper proposes a lightweight synergy architecture of IoT data and service for elderly home smart health environment. It includes the modeling of IoT applications and their workflows, data interoperability, interaction, and storage paradigms to meet the growing needs of older people so that they can lead an active, fulfilling, and quality life.

Keywords: smart home healthcare, IoT, independent living, lightweight framework

Procedia PDF Downloads 53
16376 Energy Models for Analyzing the Economic Wide Impact of the Environmental Policies

Authors: Majdi M. Alomari, Nafesah I. Alshdaifat, Mohammad S. Widyan

Abstract:

Different countries have introduced different schemes and policies to counter global warming. The rationale behind the proposed policies and the potential barriers to successful implementation of the policies adopted by the countries were analyzed and estimated based on different models. It is argued that these models enhance the transparency and provide a better understanding to the policy makers. However, these models are underpinned with several structural and baseline assumptions. These assumptions, modeling features and future prediction of emission reductions and other implication such as cost and benefits of a transition to a low-carbon economy and its economy wide impacts were discussed. On the other hand, there are potential barriers in the form political, financial, and cultural and many others that pose a threat to the mitigation options.

Keywords: energy models, environmental policy instruments, mitigating CO2 emission, economic wide impact

Procedia PDF Downloads 524
16375 Computational Approach to the Interaction of Neurotoxins and Kv1.3 Channel

Authors: Janneth González, George Barreto, Ludis Morales, Angélica Sabogal

Abstract:

Sea anemone neurotoxins are peptides that interact with Na+ and K+ channels, resulting in specific alterations on their functions. Some of these neurotoxins (1ROO, 1BGK, 2K9E, 1BEI) are important for the treatment of nearly eighty autoimmune disorders due to their specificity for Kv1.3 channel. The aim of this study was to identify the common residues among these neurotoxins by computational methods, and establish whether there is a pattern useful for the future generation of a treatment for autoimmune diseases. Our results showed eight new key common residues between the studied neurotoxins interacting with a histidine ring and the selectivity filter of the receptor, thus showing a possible pattern of interaction. This knowledge may serve as an input for the design of more promising drugs for autoimmune treatments.

Keywords: neurotoxins, potassium channel, Kv1.3, computational methods, autoimmune diseases

Procedia PDF Downloads 374
16374 Investigating Effects of Vehicle Speed and Road PSDs on Response of a 35-Ton Heavy Commercial Vehicle (HCV) Using Mathematical Modelling

Authors: Amal G. Kurian

Abstract:

The use of mathematical modeling has seen a considerable boost in recent times with the development of many advanced algorithms and mathematical modeling capabilities. The advantages this method has over other methods are that they are much closer to standard physics theories and thus represent a better theoretical model. They take lesser solving time and have the ability to change various parameters for optimization, which is a big advantage, especially in automotive industry. This thesis work focuses on a thorough investigation of the effects of vehicle speed and road roughness on a heavy commercial vehicle ride and structural dynamic responses. Since commercial vehicles are kept in operation continuously for longer periods of time, it is important to study effects of various physical conditions on the vehicle and its user. For this purpose, various experimental as well as simulation methodologies, are adopted ranging from experimental transfer path analysis to various road scenario simulations. To effectively investigate and eliminate several causes of unwanted responses, an efficient and robust technique is needed. Carrying forward this motivation, the present work focuses on the development of a mathematical model of a 4-axle configuration heavy commercial vehicle (HCV) capable of calculating responses of the vehicle on different road PSD inputs and vehicle speeds. Outputs from the model will include response transfer functions and PSDs and wheel forces experienced. A MATLAB code will be developed to implement the objectives in a robust and flexible manner which can be exploited further in a study of responses due to various suspension parameters, loading conditions as well as vehicle dimensions. The thesis work resulted in quantifying the effect of various physical conditions on ride comfort of the vehicle. An increase in discomfort is seen with velocity increase; also the effect of road profiles has a considerable effect on comfort of the driver. Details of dominant modes at each frequency are analysed and mentioned in work. The reduction in ride height or deflection of tire and suspension with loading along with load on each axle is analysed and it is seen that the front axle supports a greater portion of vehicle weight while more of payload weight comes on fourth and third axles. The deflection of the vehicle is seen to be well inside acceptable limits.

Keywords: mathematical modeling, HCV, suspension, ride analysis

Procedia PDF Downloads 259
16373 Countercyclical Capital Buffer in the Polish Banking System

Authors: Mateusz Mokrogulski, Piotr Śliwka

Abstract:

The aim of this paper is the identification of periods of excessive credit growth in the Polish banking sector in years 2007-2014 using different methodologies. Due to the lack of precise guidance in CRD IV regarding methods of calculating the credit gap and related deviations from the long-term trends, a few filtering methods are applied, e.g. Hodrick-Prescott and Baxter-King. The solutions based on the switching model are also proposed. The next step represent computations of both the credit gap, and the counter cyclical capital buffer (CCB) rates on a quarterly basis. The calculations are carried out for the entire banking sector in Poland, as well as for its components (commercial and co-operative banks), and different types of loans. The calculations show vividly that in the analysed period there were the times of excessive credit growth. However, the results are different for the above mentioned sub-sectors. Of paramount importance here are mortgage loans, where the outcomes are distorted by high exchange rate fluctuations. The research on the CCB is now going to gain popularity as the buffer will soon become one of the tools of the macro prudential policy under CRD IV. Although the presented method is focused on the Polish banking sector, it can also be applied to other member states. Especially to the Central and Eastern European countries, that are usually characterized by smaller banking sectors compared to EU-15.

Keywords: countercyclical capital buffer, CRD IV, filtering methods, mortgage loans

Procedia PDF Downloads 322
16372 Balancing Independence and Guidance: Cultivating Student Agency in Blended Learning

Authors: Yeo Leng Leng

Abstract:

Blended learning, with its combination of online and face-to-face instruction, presents a unique set of challenges and opportunities in terms of cultivating student agency. While it offers flexibility and personalized learning pathways, it also demands a higher degree of self-regulation and motivation from students. This paper presents the design of blended learning in a Chinese lesson and discusses the framework involved. It also talks about the Edtech tools adopted to engage the students. Some of the students’ works will be showcased. A qualitative case study research method was employed in this paper to find out more about students’ learning experiences and to give them a voice. The purpose is to seek improvement in the blended learning design of the Chinese lessons and to encourage students’ self-directed learning.

Keywords: blended learning, student agency, ed-tech tools, self-directed learning

Procedia PDF Downloads 78
16371 Seismic Vulnerability of Structures Designed in Accordance with the Allowable Stress Design and Load Resistant Factor Design Methods

Authors: Mohammadreza Vafaei, Amirali Moradi, Sophia C. Alih

Abstract:

The method selected for the design of structures not only can affect their seismic vulnerability but also can affect their construction cost. For the design of steel structures, two distinct methods have been introduced by existing codes, namely allowable stress design (ASD) and load resistant factor design (LRFD). This study investigates the effect of using the aforementioned design methods on the seismic vulnerability and construction cost of steel structures. Specifically, a 20-story building equipped with special moment resisting frame and an eccentrically braced system was selected for this study. The building was designed for three different intensities of peak ground acceleration including 0.2 g, 0.25 g, and 0.3 g using the ASD and LRFD methods. The required sizes of beams, columns, and braces were obtained using response spectrum analysis. Then, the designed frames were subjected to nine natural earthquake records which were scaled to the designed response spectrum. For each frame, the base shear, story shears, and inter-story drifts were calculated and then were compared. Results indicated that the LRFD method led to a more economical design for the frames. In addition, the LRFD method resulted in lower base shears and larger inter-story drifts when compared with the ASD method. It was concluded that the application of the LRFD method not only reduced the weights of structural elements but also provided a higher safety margin against seismic actions when compared with the ASD method.

Keywords: allowable stress design, load resistant factor design, nonlinear time history analysis, seismic vulnerability, steel structures

Procedia PDF Downloads 270
16370 Simulation Based Performance Comparison of Different Control Methods of ZSI Feeding Industrial Drives

Authors: Parag Nihawan, Ravinder Singh Bhatia, Dinesh Kumar Jain

Abstract:

Industrial drives are source of serious power quality problems. In this, two typical industrial drives have been dealt with, namely, FOC induction motor drives and DTC induction motor drive. The Z-source inverter is an emerging topology of power electronic converters which is capable of buck boost characteristics. The performances of different control methods based Z-source inverters feeding these industrial drives have been investigated, in this work. The test systems have been modeled and simulated in MATLAB/SIMULINK. The results obtained after carrying out these simulations have been used to draw the conclusions.

Keywords: Z-source inverter, total harmonic distortion, direct torque control, field orientation control

Procedia PDF Downloads 582
16369 Total Quality Management in Companies Manufacturing

Authors: Malki Nadia Fatima Zahra, Kellal Cheimaa, Brahimi Houria

Abstract:

Aim of the study is to show the role of total Quality Management on firm performance; the research relied on the views of sample managers working in the Marinel pharmaceutical company. The research aims to achieve many objectives, including increasing awareness of the concepts of Total Quality Management on Firm Performance, especially in the manufacturing firm, providing a future vision of the possibility of success, and the actual application of the Principles of Total Quality Management in the manufacturing company. The research adopted a default model was built after a review and analysis of the literature review in the context of one hypothesis main points at the origin of a group of sub-hypotheses. The research presented a set of conclusions, and the most important of these conclusions was there is a relationship between the Principles of TQM and Firm Performance.

Keywords: total quality management, TQM dimension, firm performance, strategies

Procedia PDF Downloads 69
16368 Comparative Analysis of the Computer Methods' Usage for Calculation of Hydrocarbon Reserves in the Baltic Sea

Authors: Pavel Shcherban, Vlad Golovanov

Abstract:

Nowadays, the depletion of hydrocarbon deposits on the land of the Kaliningrad region leads to active geological exploration and development of oil and natural gas reserves in the southeastern part of the Baltic Sea. LLC 'Lukoil-Kaliningradmorneft' implements a comprehensive program for the development of the region's shelf in 2014-2023. Due to heterogeneity of reservoir rocks in various open fields, as well as with ambiguous conclusions on the contours of deposits, additional geological prospecting and refinement of the recoverable oil reserves are carried out. The key element is use of an effective technique of computer stock modeling at the first stage of processing of the received data. The following step uses information for the cluster analysis, which makes it possible to optimize the field development approaches. The article analyzes the effectiveness of various methods for reserves' calculation and computer modelling methods of the offshore hydrocarbon fields. Cluster analysis allows to measure influence of the obtained data on the development of a technical and economic model for mining deposits. The relationship between the accuracy of the calculation of recoverable reserves and the need of modernization of existing mining infrastructure, as well as the optimization of the scheme of opening and development of oil deposits, is observed.

Keywords: cluster analysis, computer modelling of deposits, correction of the feasibility study, offshore hydrocarbon fields

Procedia PDF Downloads 166
16367 Evaluation of Current Methods in Modelling and Analysis of Track with Jointed Rails

Authors: Hossein Askarinejad, Manicka Dhanasekar

Abstract:

In railway tracks, two adjacent rails are either welded or connected using bolted jointbars. In recent years the number of bolted rail joints is reduced by introduction of longer rail sections and by welding the rails at location of some joints. However, significant number of bolted rail joints remains in railways around the world as they are required to allow for rail thermal expansion or to provide electrical insulation in some sections of track. Regardless of the quality and integrity of the jointbar and bolt connections, the bending stiffness of jointbars is much lower than the rail generating large deflections under the train wheels. In addition, the gap or surface discontinuity on the rail running surface leads to generation of high wheel-rail impact force at the joint gap. These fundamental weaknesses have caused high rate of failure in track components at location of rail joints resulting in significant economic and safety issues in railways. The mechanical behavior of railway track at location of joints has not been fully understood due to various structural and material complexities. Although there have been some improvements in the methods for analysis of track at jointed rails in recent years, there are still uncertainties concerning the accuracy and reliability of the current methods. In this paper the current methods in analysis of track with a rail joint are critically evaluated and the new advances and recent research outcomes in this area are discussed. This research is part of a large granted project on rail joints which was defined by Cooperative Research Centre (CRC) for Rail Innovation with supports from Australian Rail Track Corporation (ARTC) and Queensland Rail (QR).

Keywords: jointed rails, railway mechanics, track dynamics, wheel-rail interaction

Procedia PDF Downloads 350
16366 Urban Noise and Air Quality: Correlation between Air and Noise Pollution; Sensors, Data Collection, Analysis and Mapping in Urban Planning

Authors: Massimiliano Condotta, Paolo Ruggeri, Chiara Scanagatta, Giovanni Borga

Abstract:

Architects and urban planners, when designing and renewing cities, have to face a complex set of problems, including the issues of noise and air pollution which are considered as hot topics (i.e., the Clean Air Act of London and the Soundscape definition). It is usually taken for granted that these problems go by together because the noise pollution present in cities is often linked to traffic and industries, and these produce air pollutants as well. Traffic congestion can create both noise pollution and air pollution, because NO₂ is mostly created from the oxidation of NO, and these two are notoriously produced by processes of combustion at high temperatures (i.e., car engines or thermal power stations). We can see the same process for industrial plants as well. What have to be investigated – and is the topic of this paper – is whether or not there really is a correlation between noise pollution and air pollution (taking into account NO₂) in urban areas. To evaluate if there is a correlation, some low-cost methodologies will be used. For noise measurements, the OpeNoise App will be installed on an Android phone. The smartphone will be positioned inside a waterproof box, to stay outdoor, with an external battery to allow it to collect data continuously. The box will have a small hole to install an external microphone, connected to the smartphone, which will be calibrated to collect the most accurate data. For air, pollution measurements will be used the AirMonitor device, an Arduino board to which the sensors, and all the other components, are plugged. After assembling the sensors, they will be coupled (one noise and one air sensor) and placed in different critical locations in the area of Mestre (Venice) to map the existing situation. The sensors will collect data for a fixed period of time to have an input for both week and weekend days, in this way it will be possible to see the changes of the situation during the week. The novelty is that data will be compared to check if there is a correlation between the two pollutants using graphs that should show the percentage of pollution instead of the values obtained with the sensors. To do so, the data will be converted to fit on a scale that goes up to 100% and will be shown thru a mapping of the measurement using GIS methods. Another relevant aspect is that this comparison can help to choose which are the right mitigation solutions to be applied in the area of the analysis because it will make it possible to solve both the noise and the air pollution problem making only one intervention. The mitigation solutions must consider not only the health aspect but also how to create a more livable space for citizens. The paper will describe in detail the methodology and the technical solution adopted for the realization of the sensors, the data collection, noise and pollution mapping and analysis.

Keywords: air quality, data analysis, data collection, NO₂, noise mapping, noise pollution, particulate matter

Procedia PDF Downloads 212
16365 Assessment of Adequacy of Pile Load Determination Formulas

Authors: Ashraf Eid

Abstract:

Many formulas are set to estimate the pile load capacity based on the in-situ pile load tests. However, discrepancy is obvious between the results of these applications. Some formulas are more adequate than others with respect to soil formation and its characteristics. In this research, attempts were undertaken to evaluate the adequacy of the most well-known formulas based on a series of pile load tests carried out in Port Said city in the northeast of Egypt for major residential projects. Comparisons were undertaken between the different formulas supported by the results of in-situ Cone Penetration Tests (CPT). Based on this study, a guide for engineers for using the proper formula can be adopted with consideration of soil type and characteristics. The Egyptian Code which relies on the results of some formulas is involved in the study as a guiding aspect in the pile design.

Keywords: pile load formula, load test of piles, CPT, Egyptian code

Procedia PDF Downloads 152