Search results for: software cumulative failure prediction
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 9049

Search results for: software cumulative failure prediction

7549 Prediction of Physical Properties and Sound Absorption Performance of Automotive Interior Materials

Authors: Un-Hwan Park, Jun-Hyeok Heo, In-Sung Lee, Seong-Jin Cho, Tae-Hyeon Oh, Dae-Kyu Park

Abstract:

Sound absorption coefficient is considered important when designing because noise affects emotion quality of car. It is designed with lots of experiment tunings in the field because it is unreliable to predict it for multi-layer material. In this paper, we present the design of sound absorption for automotive interior material with multiple layers using estimation software of sound absorption coefficient for reverberation chamber. Additionally, we introduce the method for estimation of physical properties required to predict sound absorption coefficient of car interior materials with multiple layers too. It is calculated by inverse algorithm. It is very economical to get information about physical properties without expensive equipment. Correlation test is carried out to ensure reliability for accuracy. The data to be used for the correlation is sound absorption coefficient measured in the reverberation chamber. In this way, it is considered economical and efficient to design automotive interior materials. And design optimization for sound absorption coefficient is also easy to implement when it is designed.

Keywords: sound absorption coefficient, optimization design, inverse algorithm, automotive interior material, multiple layers nonwoven, scaled reverberation chamber, sound impedance tubes

Procedia PDF Downloads 292
7548 Ensuring Quality in DevOps Culture

Authors: Sagar Jitendra Mahendrakar

Abstract:

Integrating quality assurance (QA) practices into DevOps culture has become increasingly important in modern software development environments. Collaboration, automation and continuous feedback characterize the seamless integration of DevOps development and operations teams to achieve rapid and reliable software delivery. In this context, quality assurance plays a key role in ensuring that software products meet the highest quality, performance and reliability standards throughout the development life cycle. This brief explores key principles, challenges, and best practices related to quality assurance in a DevOps culture. This emphasizes the importance of quality transfer in the development process, as quality control processes are integrated in every step of the DevOps process. Automation is the cornerstone of DevOps quality assurance, enabling continuous testing, integration and deployment and providing rapid feedback for early problem identification and resolution. In addition, the summary addresses the cultural and organizational challenges of implementing quality assurance in DevOps, emphasizing the need to foster collaboration, break down silos, and promote a culture of continuous improvement. It also discusses the importance of toolchain integration and capability development to support effective QA practices in DevOps environments. Moreover, the abstract discusses the cultural and organizational challenges in implementing QA within DevOps, emphasizing the need for fostering collaboration, breaking down silos, and nurturing a culture of continuous improvement. It also addresses the importance of toolchain integration and skills development to support effective QA practices within DevOps environments. Overall, this collection works at the intersection of QA and DevOps culture, providing insights into how organizations can use DevOps principles to improve software quality, accelerate delivery, and meet the changing demands of today's dynamic software. landscape.

Keywords: quality engineer, devops, automation, tool

Procedia PDF Downloads 38
7547 Numerical Investigation of Hot Oil Velocity Effect on Force Heat Convection and Impact of Wind Velocity on Convection Heat Transfer in Receiver Tube of Parabolic Trough Collector System

Authors: O. Afshar

Abstract:

A solar receiver is designed for operation under extremely uneven heat flux distribution, cyclic weather, and cloud transient cycle conditions, which can include large thermal stress and even receiver failure. In this study, the effect of different oil velocity on convection coefficient factor and impact of wind velocity on local Nusselt number by Finite Volume Method will be analyzed. This study is organized to give an overview of the numerical modeling using a MATLAB software, as an accurate, time efficient and economical way of analyzing the heat transfer trends over stationary receiver tube for different Reynolds number. The results reveal when oil velocity is below 0.33m/s, the value of convection coefficient is negligible at low temperature. The numerical graphs indicate that when oil velocity increases up to 1.2 m/s, heat convection coefficient increases significantly. In fact, a reduction in oil velocity causes a reduction in heat conduction through the glass envelope. In addition, the different local Nusselt number is reduced when the wind blows toward the concave side of the collector and it has a significant effect on heat losses reduction through the glass envelope.

Keywords: receiver tube, heat convection, heat conduction, Nusselt number

Procedia PDF Downloads 342
7546 Empirical Investigation of Antecedents of Perceived Recovery Service Quality: Evidence from Retail Banking in United Arab Emirates

Authors: Vimi Jham

Abstract:

The banking sector has undergone tremendous change in all forms of service it provides to its customers. The efforts of the banks is to avoid customer defection and lead to customer satisfaction. The purpose of the study was to examine the linkages among the constructs such as customer perceived service quality, perceived service recovery quality and customer satisfaction in the banking industry. The moderating effect of negative brand perception due to service failure on recovery satisfaction were investigated. Random sampling methods are used to draw the sample from the population. Data was collected from 262 banking customers and were analyzed with the help of structural equation modelling approach using Smart PLS to understand the relationship among variables being studied. The results of the study contribute to the research by proving that customer service recovery satisfaction is dependent on customer perceived service quality and the moderating effect of negative brand perception due to service failure was insignificant.

Keywords: service recovery satisfaction, perceived service recovery quality, perceived service quality, structural equation modelling

Procedia PDF Downloads 271
7545 Clash of Civilizations without Civilizational Groups: Revisiting Samuel P. Huntington's Clash of Civilizations Theory

Authors: Jamal Abdi

Abstract:

This paper offers a critique of Samuel P. Huntington's Clash of Civilizations thesis. The overriding argument is that Huntington's thesis is characterized by failure to distinguish between 'groups' and 'categories'. Multinational civilizations overcoming their internal collective action problems, which would enable them to pursue a unified strategy vis-à-vis the West, is a rather foundational assumption in his theory. Without assigning sufficient intellectual attention to the processes through which multinational civilizations may gain the capacity for concerted action, i.e., become a group, he contended that the post-cold-war world would be shaped in large measure by interactions among seven or eight major civilizations. Thus, failure in providing a convincing analysis of multi-national civilizations' transition from categories to groups is a significant weakness in Huntington's clash theory. It is also suggested that so-called Islamic terrorism and the war on terror is not to be taken as an expression of the presence of clash between a Western and an Islamic civilization, as terrorist organizations would be superfluous in a world characterized by clash of civilizations. Consequences of multinational civilizations becoming a group are discussed in relation to contemporary Western superiority.

Keywords: clash of civilizations, groups, categories, groupism

Procedia PDF Downloads 187
7544 Modeling of the Biodegradation Performance of a Membrane Bioreactor to Enhance Water Reuse in Agri-food Industry - Poultry Slaughterhouse as an Example

Authors: masmoudi Jabri Khaoula, Zitouni Hana, Bousselmi Latifa, Akrout Hanen

Abstract:

Mathematical modeling has become an essential tool for sustainable wastewater management, particularly for the simulation and the optimization of complex processes involved in activated sludge systems. In this context, the activated sludge model (ASM3h) was used for the simulation of a Biological Membrane Reactor (MBR) as it includes the integration of biological wastewater treatment and physical separation by membrane filtration. In this study, the MBR with a useful volume of 12.5 L was fed continuously with poultry slaughterhouse wastewater (PSWW) for 50 days at a feed rate of 2 L/h and for a hydraulic retention time (HRT) of 6.25h. Throughout its operation, High removal efficiency was observed for the removal of organic pollutants in terms of COD with 84% of efficiency. Moreover, the MBR has generated a treated effluent which fits with the limits of discharge into the public sewer according to the Tunisian standards which were set in March 2018. In fact, for the nitrogenous compounds, average concentrations of nitrate and nitrite in the permeat reached 0.26±0.3 mg. L-1 and 2.2±2.53 mg. L-1, respectively. The simulation of the MBR process was performed using SIMBA software v 5.0. The state variables employed in the steady state calibration of the ASM3h were determined using physical and respirometric methods. The model calibration was performed using experimental data obtained during the first 20 days of the MBR operation. Afterwards, kinetic parameters of the model were adjusted and the simulated values of COD, N-NH4+and N- NOx were compared with those reported from the experiment. A good prediction was observed for the COD, N-NH4+and N- NOx concentrations with 467 g COD/m³, 110.2 g N/m³, 3.2 g N/m³ compared to the experimental data which were 436.4 g COD/m³, 114.7 g N/m³ and 3 g N/m³, respectively. For the validation of the model under dynamic simulation, the results of the experiments obtained during the second treatment phase of 30 days were used. It was demonstrated that the model simulated the conditions accurately by yielding a similar pattern on the variation of the COD concentration. On the other hand, an underestimation of the N-NH4+ concentration was observed during the simulation compared to the experimental results and the measured N-NO3 concentrations were lower than the predicted ones, this difference could be explained by the fact that the ASM models were mainly designed for the simulation of biological processes in the activated sludge systems. In addition, more treatment time could be required by the autotrophic bacteria to achieve a complete and stable nitrification. Overall, this study demonstrated the effectiveness of mathematical modeling in the prediction of the performance of the MBR systems with respect to organic pollution, the model can be further improved for the simulation of nutrients removal for a longer treatment period.

Keywords: activated sludge model (ASM3h), membrane bioreactor (MBR), poultry slaughter wastewater (PSWW), reuse

Procedia PDF Downloads 41
7543 Statistical Comparison of Ensemble Based Storm Surge Forecasting Models

Authors: Amin Salighehdar, Ziwen Ye, Mingzhe Liu, Ionut Florescu, Alan F. Blumberg

Abstract:

Storm surge is an abnormal water level caused by a storm. Accurate prediction of a storm surge is a challenging problem. Researchers developed various ensemble modeling techniques to combine several individual forecasts to produce an overall presumably better forecast. There exist some simple ensemble modeling techniques in literature. For instance, Model Output Statistics (MOS), and running mean-bias removal are widely used techniques in storm surge prediction domain. However, these methods have some drawbacks. For instance, MOS is based on multiple linear regression and it needs a long period of training data. To overcome the shortcomings of these simple methods, researchers propose some advanced methods. For instance, ENSURF (Ensemble SURge Forecast) is a multi-model application for sea level forecast. This application creates a better forecast of sea level using a combination of several instances of the Bayesian Model Averaging (BMA). An ensemble dressing method is based on identifying best member forecast and using it for prediction. Our contribution in this paper can be summarized as follows. First, we investigate whether the ensemble models perform better than any single forecast. Therefore, we need to identify the single best forecast. We present a methodology based on a simple Bayesian selection method to select the best single forecast. Second, we present several new and simple ways to construct ensemble models. We use correlation and standard deviation as weights in combining different forecast models. Third, we use these ensembles and compare with several existing models in literature to forecast storm surge level. We then investigate whether developing a complex ensemble model is indeed needed. To achieve this goal, we use a simple average (one of the simplest and widely used ensemble model) as benchmark. Predicting the peak level of Surge during a storm as well as the precise time at which this peak level takes place is crucial, thus we develop a statistical platform to compare the performance of various ensemble methods. This statistical analysis is based on root mean square error of the ensemble forecast during the testing period and on the magnitude and timing of the forecasted peak surge compared to the actual time and peak. In this work, we analyze four hurricanes: hurricanes Irene and Lee in 2011, hurricane Sandy in 2012, and hurricane Joaquin in 2015. Since hurricane Irene developed at the end of August 2011 and hurricane Lee started just after Irene at the beginning of September 2011, in this study we consider them as a single contiguous hurricane event. The data set used for this study is generated by the New York Harbor Observing and Prediction System (NYHOPS). We find that even the simplest possible way of creating an ensemble produces results superior to any single forecast. We also show that the ensemble models we propose generally have better performance compared to the simple average ensemble technique.

Keywords: Bayesian learning, ensemble model, statistical analysis, storm surge prediction

Procedia PDF Downloads 297
7542 Coastal Water Characteristics along the Saudi Arabian Coastline

Authors: Yasser O. Abualnaja1, Alexandra Pavlidou2, Taha Boksmati3, Ahmad Alharbi3, Hammad Alsulmi3, Saleh Omar Maghrabi3, Hassan Mowalad3, Rayan Mutwalli3, James H. Churchill4, Afroditi Androni2, Dionysios Ballas2, Ioannis Hatzianestis2, Harilaos Kontoyiannis2, Angeliki Konstantinopoulou2, Georgios Krokkos1, 5, Georgios Pappas2, Vassilis P. Papadopoulos2, Konstantinos Parinos2, Elvira Plakidi2, Eleni Rousselaki2, Dimitris Velaoras2, Panagiota Zachioti2, Theodore Zoulias2, Ibrahim Hoteit5.

Abstract:

The coastal areas along the Kingdom of Saudi Arabia on both the Red Sea and Arabian Gulf have been witnessing in the past decades an unprecedented economic growth and a rapid increase in anthropogenic activities. Therefore, the Saudi Arabian government has decided to frame a strategy for sustainable development of the coastal and marine environments, which comes in the context of the Vision 2030, aimed at providing the first comprehensive ‘Status Quo Assessment’ of the Kingdom’s coastal and marine environments. This strategy will serve as a baseline assessment for future monitoring activities; this baseline is relied on scientific evidence of the drivers, pressures, and their impact on the environments of the Red Sea and Arabian Gulf. A key element of the assessment was the cumulative pressures of the hotspots analysis, which was developed following the principles of the Driver-Pressure-State-Impact-Response (DPSIR) framework and using the cumulative pressure and impact assessment methodology. Ten hotspot sites were identified, eight in the Red Sea and two in the Arabian Gulf. Thus, multidisciplinary research cruises were conducted throughout the Red Sea and the Arabian Gulf coastal and marine environments in June/July 2021 and September 2021, respectively, in order to understand the relative impact of hydrography and the various pressures on the quality of seawater and sediments. The main objective was to record the physical and biogeochemical parameters along the coastal waters of the Kingdom, tracing the dispersion of contaminants related to specific pressures. The assessment revealed the effect of hydrography on the trophic status of the southern marine coastal areas of the Red Sea. Jeddah Lagoon system seems to face significant eutrophication and pollution challenges, whereas sediments are enriched in some heavy metals in many areas of the Red Sea and the Arabian Gulf. This multidisciplinary research in the Red Sea and the Arabian Gulf coastal waters will pave the way for future detailed environmental monitoring strategies for the Saudi Arabian marine environment.

Keywords: arabian gulf, contaminants, hotspot, red sea

Procedia PDF Downloads 97
7541 Investigation of the Multiaxial Pedicle Screw Tulip Design Using Finite Element Analysis

Authors: S. Daqiqeh Rezaei, S. Mohajerzadeh, M. R. Sharifi

Abstract:

Pedicle screws are used to stabilize vertebrae and treat several types of spinal diseases and injuries. Multiaxial pedicle screws are a type of pedicle screw that increase surgical versatility, but they also increase design complexity. Failure of multiaxial pedicle screws caused by static loading, dynamic loading and fatigue can lead to irreparable damage to the patient. Inappropriate deformation of the multiaxial pedicle screw tulip can cause system failure. Investigation of deformation and stress in these tulips can be employed to optimize multiaxial pedicle screw design. The sensitivity of this matter necessitates precise analyzing and modeling of pedicle screws. In this work, three commercial multiaxial pedicle screw tulips and a newly designed tulip are investigated using finite element analysis. Employing video measuring machine (VMM), tulips are modeled. Afterwards, utilizing ANSYS, static analysis is performed on these models. In the end, stresses and displacements of the models are compared.

Keywords: pedicle screw, multiaxial pedicle screw, finite element analysis, static analysis

Procedia PDF Downloads 351
7540 Design of Quality Assessment System for On-Orbit 3D Printing Based on 3D Reconstruction Technology

Authors: Jianning Tang, Trevor Hocksun Kwan, Xiaofeng Wu

Abstract:

With the increasing demand for space use in multiple sectors (navigation, telecommunication, imagery, etc.), the deployment and maintenance demand of satellites are growing. Considering the high launching cost and the restrictions on weight and size of the payload when using launch vehicle, the technique of on-orbit manufacturing has obtained more attention because of its significant potential to support future space missions. 3D printing is the most promising manufacturing technology that could be applied in space. However, due to the lack of autonomous quality assessment, the operation of conventional 3D printers still relies on human presence to supervise the printing process. This paper is proposed to develop an automatic 3D reconstruction system aiming at detecting failures on the 3D printed objects through application of point cloud technology. Based on the data obtained from the point cloud, the 3D printer could locate the failure and repair the failure. The system will increase automation and provide 3D printing with more feasibilities for space use without human interference.

Keywords: 3D printing, quality assessment, point cloud, on-orbit manufacturing

Procedia PDF Downloads 104
7539 Integrating Computer-Aided Manufacturing and Computer-Aided Design for Streamlined Carpentry Production in Ghana

Authors: Benson Tette, Thomas Mensah

Abstract:

As a developing country, Ghana has a high potential to harness the economic value of every industry. Two of the industries that produce below capacity are handicrafts (for instance, carpentry) and information technology (i.e., computer science). To boost production and maintain competitiveness, the carpentry sector in Ghana needs more effective manufacturing procedures that are also more affordable. This issue can be resolved using computer-aided manufacturing (CAM) technology, which automates the fabrication process and decreases the amount of time and labor needed to make wood goods. Yet, the integration of CAM in carpentry-related production is rarely explored. To streamline the manufacturing process, this research investigates the equipment and technology that are currently used in the Ghanaian carpentry sector for automated fabrication. The research looks at the various CAM technologies, such as Computer Numerical Control routers, laser cutters, and plasma cutters, that are accessible to Ghanaian carpenters yet unexplored. We also investigate their potential to enhance the production process. To achieve the objective, 150 carpenters, 15 software engineers, and 10 policymakers were interviewed using structured questionnaires. The responses provided by the 175 respondents were processed to eliminate outliers and omissions were corrected using multiple imputations techniques. The processed responses were analyzed through thematic analysis. The findings showed that adaptation and integration of CAD software with CAM technologies would speed up the design-to-manufacturing process for carpenters. It must be noted that achieving such results entails first; examining the capabilities of current CAD software, then determining what new functions and resources are required to improve the software's suitability for carpentry tasks. Responses from both carpenters and computer scientists showed that it is highly practical and achievable to streamline the design-to-manufacturing process through processes such as modifying and combining CAD software with CAM technology. Making the carpentry-software integration program more useful for carpentry projects would necessitate investigating the capabilities of the current CAD software and identifying additional features in the Ghanaian ecosystem and tools that are required. In conclusion, the Ghanaian carpentry sector has a chance to increase productivity and competitiveness through the integration of CAM technology with CAD software. Carpentry companies may lower labor costs and boost production capacity by automating the fabrication process, giving them a competitive advantage. This study offers implementation-ready and representative recommendations for successful implementation as well as important insights into the equipment and technologies available for automated fabrication in the Ghanaian carpentry sector.

Keywords: carpentry, computer-aided manufacturing (CAM), Ghana, information technology(IT)

Procedia PDF Downloads 75
7538 Software Transactional Memory in a Dynamic Programming Language at Virtual Machine Level

Authors: Szu-Kai Hsu, Po-Ching Lin

Abstract:

As more and more multi-core processors emerge, traditional sequential programming paradigm no longer suffice. Yet only few modern dynamic programming languages can leverage such advantage. Ruby, for example, despite its wide adoption, only includes threads as a simple parallel primitive. The global virtual machine lock of official Ruby runtime makes it impossible to exploit full parallelism. Though various alternative Ruby implementations do eliminate the global virtual machine lock, they only provide developers dated locking mechanism for data synchronization. However, traditional locking mechanism error-prone by nature. Software Transactional Memory is one of the promising alternatives among others. This paper introduces a new virtual machine: GobiesVM to provide a native software transactional memory based solution for dynamic programming languages to exploit parallelism. We also proposed a simplified variation of Transactional Locking II algorithm. The empirical results of our experiments show that support of STM at virtual machine level enables developers to write straightforward code without compromising parallelism or sacrificing thread safety. Existing source code only requires minimal or even none modi cation, which allows developers to easily switch their legacy codebase to a parallel environment. The performance evaluations of GobiesVM also indicate the difference between sequential and parallel execution is significant.

Keywords: global interpreter lock, ruby, software transactional memory, virtual machine

Procedia PDF Downloads 267
7537 The Ability of Forecasting the Term Structure of Interest Rates Based on Nelson-Siegel and Svensson Model

Authors: Tea Poklepović, Zdravka Aljinović, Branka Marasović

Abstract:

Due to the importance of yield curve and its estimation it is inevitable to have valid methods for yield curve forecasting in cases when there are scarce issues of securities and/or week trade on a secondary market. Therefore in this paper, after the estimation of weekly yield curves on Croatian financial market from October 2011 to August 2012 using Nelson-Siegel and Svensson models, yield curves are forecasted using Vector auto-regressive model and Neural networks. In general, it can be concluded that both forecasting methods have good prediction abilities where forecasting of yield curves based on Nelson Siegel estimation model give better results in sense of lower Mean Squared Error than forecasting based on Svensson model Also, in this case Neural networks provide slightly better results. Finally, it can be concluded that most appropriate way of yield curve prediction is neural networks using Nelson-Siegel estimation of yield curves.

Keywords: Nelson-Siegel Model, neural networks, Svensson Model, vector autoregressive model, yield curve

Procedia PDF Downloads 306
7536 Photo-Fenton Decolorization of Methylene Blue Adsolubilized on Co2+ -Embedded Alumina Surface: Comparison of Process Modeling through Response Surface Methodology and Artificial Neural Network

Authors: Prateeksha Mahamallik, Anjali Pal

Abstract:

In the present study, Co(II)-adsolubilized surfactant modified alumina (SMA) was prepared, and methylene blue (MB) degradation was carried out on Co-SMA surface by visible light photo-Fenton process. The entire reaction proceeded on solid surface as MB was embedded on Co-SMA surface. The reaction followed zero order kinetics. Response surface methodology (RSM) and artificial neural network (ANN) were used for modeling the decolorization of MB by photo-Fenton process as a function of dose of Co-SMA (10, 20 and 30 g/L), initial concentration of MB (10, 20 and 30 mg/L), concentration of H2O2 (174.4, 348.8 and 523.2 mM) and reaction time (30, 45 and 60 min). The prediction capabilities of both the methodologies (RSM and ANN) were compared on the basis of correlation coefficient (R2), root mean square error (RMSE), standard error of prediction (SEP), relative percent deviation (RPD). Due to lower value of RMSE (1.27), SEP (2.06) and RPD (1.17) and higher value of R2 (0.9966), ANN was proved to be more accurate than RSM in order to predict decolorization efficiency.

Keywords: adsolubilization, artificial neural network, methylene blue, photo-fenton process, response surface methodology

Procedia PDF Downloads 241
7535 Numerical Evaluation of Shear Strength for Cold-Formed Steel Shear Wall Panel

Authors: Rouaz Idriss, Bourahla Nour-Eddine, Kahlouche Farah, Rafa Sid Ali

Abstract:

The stability of structures made of light-gauge steel depends highly on the contribution of Shear Wall Panel (SWP) systems under horizontal forces due to wind or earthquake loads. Steel plate sheathing is often used with these panels made of cold formed steel (CFS) to improve its shear strength. In order to predict the shear strength resistance, two methods are presented in this paper. In the first method, the steel plate sheathing is modeled with plats strip taking into account only the tension and compression force due to the horizontal load, where both track and stud are modeled according to the geometrical and mechanical characteristics of the specimen used in the experiments. The theoretical background and empirical formulations of this method are presented in this paper. However, the second method is based on a micro modeling of the cold formed steel Shear Wall Panel “CFS-SWP” using Abaqus software. A nonlinear analysis was carried out with an in-plan monotonic load. Finally, the comparison between these two methods shows that the micro modeling with Abaqus gives better prediction of shear resistance of SWP than strips method. However, the latter is easier and less time consuming than the micro modeling method.

Keywords: cold formed steel 'CFS', shear wall panel, strip method, finite elements

Procedia PDF Downloads 292
7534 Two-Sided Information Dissemination in Takeovers: Disclosure and Media

Authors: Eda Orhun

Abstract:

Purpose: This paper analyzes a target firm’s decision to voluntarily disclose information during a takeover event and the effect of such disclosures on the outcome of the takeover. Such voluntary disclosures especially in the form of earnings forecasts made around takeover events may affect shareholders’ decisions about the target firm’s value and in return takeover result. This study aims to shed light on this question. Design/methodology/approach: The paper tries to understand the role of voluntary disclosures by target firms during a takeover event in the likelihood of takeover success both theoretically and empirically. A game-theoretical model is set up to analyze the voluntary disclosure decision of a target firm to inform the shareholders about its real worth. The empirical implication of model is tested by employing binary outcome models where the disclosure variable is obtained by identifying the target firms in the sample that provide positive news by issuing increasing management earnings forecasts. Findings: The model predicts that a voluntary disclosure of positive information by the target decreases the likelihood that the takeover succeeds. The empirical analysis confirms this prediction by showing that positive earnings forecasts by target firms during takeover events increase the probability of takeover failure. Overall, it is shown that information dissemination through voluntary disclosures by target firms is an important factor affecting takeover outcomes. Originality/Value: This study is the first to the author's knowledge that studies the impact of voluntary disclosures by the target firm during a takeover event on the likelihood of takeover success. The results contribute to information economics, corporate finance and M&As literatures.

Keywords: takeovers, target firm, voluntary disclosures, earnings forecasts, takeover success

Procedia PDF Downloads 303
7533 Experimental and FEA Study for Reduction of Damage in Sheet Metal Forming

Authors: Amitkumar R. Shelar, B. P. Ronge, Sridevi Seshabhattar, R. M. Wabale

Abstract:

This paper gives knowledge about the behavior of cold rolled steel IS 513_2008 CR2_D having grade D for the reduction of ductile damage. CR specifies Cold Rolled and D for Drawing grade. Problems encountered during sheet metal forming operations are dent, wrinkles, thinning, spring back, insufficient stretching etc. In this paper, wrinkle defect was studied experimentally and by using FE software on one of the auto components due to which its functionality was decreased. Experimental result and simulation result were found to be in agreement.

Keywords: deep drawing, FE software-LS DYNA, friction, wrinkling

Procedia PDF Downloads 475
7532 Strengthening of Reinforced Concrete Beams Using Steel Plates

Authors: Ghusen al-Kafri, Mohammed Ali Abdallah Elsageer, Ahmed Mohamed Hadya Alsdaai, Abdeimanam Salhien Salih Khalifa

Abstract:

In this paper, external reinforcement to enhance a reinforced concrete structure performance has been done using externally bonded steel plate. This technique has been reported effective in enhancing the strength of reinforced concrete beam, a study to determine the effectiveness of steel plate as an external reinforcement was carried out. A total of two groups of beams and one group content five beams, each 750 mm long, 150 mm wide, and 150 mm deep were cast, strengthened and tested till failure under two point loads. One beam was act as a control beam without strengthening and other four beams were strengthened with steel plate at a different arrangement. Other group beams were strengthened with steel plate in shear zone and also strengthened at bottom as first group. The behaviours of the strengthened beams were studied through their load-deflection characteristic upon bending, cracking and mode of failure. The results confirmed that all steel plate arrangements enhanced the strength of the reinforced concrete beam, the positioning of the steel plate affect the moment carrying capacity of the beam.

Keywords: beams, bending, beflection, steel plates

Procedia PDF Downloads 397
7531 TRAC: A Software Based New Track Circuit for Traffic Regulation

Authors: Jérôme de Reffye, Marc Antoni

Abstract:

Following the development of the ERTMS system, we think it is interesting to develop another software-based track circuit system which would fit secondary railway lines with an easy-to-work implementation and a low sensitivity to rail-wheel impedance variations. We called this track circuit 'Track Railway by Automatic Circuits.' To be internationally implemented, this system must not have any mechanical component and must be compatible with existing track circuit systems. For example, the system is independent from the French 'Joints Isolants Collés' that isolate track sections from one another, and it is equally independent from component used in Germany called 'Counting Axles,' in French 'compteur d’essieux.' This track circuit is fully interoperable. Such universality is obtained by replacing the train detection mechanical system with a space-time filtering of train position. The various track sections are defined by the frequency of a continuous signal. The set of frequencies related to the track sections is a set of orthogonal functions in a Hilbert Space. Thus the failure probability of track sections separation is precisely calculated on the basis of signal-to-noise ratio. SNR is a function of the level of traction current conducted by rails. This is the reason why we developed a very powerful algorithm to reject noise and jamming to obtain an SNR compatible with the precision required for the track circuit and SIL 4 level. The SIL 4 level is thus reachable by an adjustment of the set of orthogonal functions. Our major contributions to railway engineering signalling science are i) Train space localization is precisely defined by a calibration system. The operation bypasses the GSM-R radio system of the ERTMS system. Moreover, the track circuit is naturally protected against radio-type jammers. After the calibration operation, the track circuit is autonomous. ii) A mathematical topology adapted to train space localization by following the train through a linear time filtering of the received signal. Track sections are numerically defined and can be modified with a software update. The system was numerically simulated, and results were beyond our expectations. We achieved a precision of one meter. Rail-ground and rail-wheel impedance sensitivity analysis gave excellent results. Results are now complete and ready to be published. This work was initialised as a research project of the French Railways developed by the Pi-Ramses Company under SNCF contract and required five years to obtain the results. This track circuit is already at Level 3 of the ERTMS system, and it will be much cheaper to implement and to work. The traffic regulation is based on variable length track sections. As the traffic growths, the maximum speed is reduced, and the track section lengths are decreasing. It is possible if the elementary track section is correctly defined for the minimum speed and if every track section is able to emit with variable frequencies.

Keywords: track section, track circuits, space-time crossing, adaptive track section, automatic railway signalling

Procedia PDF Downloads 316
7530 A Methodological Concept towards a Framework Development for Social Software Adoption in Higher Education System

Authors: Kenneth N. Ohei, Roelien Brink

Abstract:

For decades, teaching and learning processes have centered on the traditional approach (Web 1.0) that promoted teacher-directed pedagogical practices. Currently, there is a realization that the traditional approach is not adequate to effectively address and improve all student-learning outcomes. The subsequent incorporation of social software, Information, and Communication Technology (ICT) tools in universities may serve as complementary to support educational goals, offering students the affordability and opportunity to educational choices and learning platforms. Consequently, educators’ inability to incorporate these instructional ICT tools in their teaching and learning practices remains a challenge. This will signify that educators still lack the ICT skills required to administer lectures and bridging learning gaps. This study probes a methodological concept with the aim of developing a framework towards the adoption of social software in HES to help facilitate business processes and can build social presence among students. A mixed method will be appropriate to develop a comprehensive framework needed in Higher Educational System (HES). After research have been conducted, the adoption of social software will be based on the developed comprehensive framework which is supposed to impact positively on education and approach of delivery, improves learning experience, engagement and finally, increases educational opportunities and easy access to educational contents.

Keywords: blended and integrated learning, learning experience and engagement, higher educational system, HES, information and communication technology, ICT, social presence, Web 1.0, Web 2.0, Web 3.0

Procedia PDF Downloads 147
7529 Corrosion Fatigue of Al-Mg Alloy 5052 in Sodium Chloride Solution Contains Some Inhibitors

Authors: Khalid Ahmed Eldwaib

Abstract:

In this study, Al-Mg alloy 5052 was used as the testing material. Corrosion fatigue life was studied for the alloy in 3.5% NaCl (pH=1, 3, 5, 7, 9, and 11), and 3.5% NaCl (pH=1) with inhibitors. The compound inhibitors were composed mainly of phosphate (PO4³-), adding a certain proportion of other nontoxic inhibitors so as to select alternatives to environmentally hazardous chromate (Cr2O7²-). The inhibitors were sodium dichromate Na2Cr2O7, sodium phosphate Na3PO4, sodium molybdate Na2MoO4, and sodium citrate Na3C6H5O7. The total amount of inhibiting pigments was at different concentrations (250,500,750, and 1000 ppm) in the solutions. Corrosion fatigue behavior was studied by using plane-bending corrosion fatigue machine with stress ratio R=0.5 and under the constant frequency of 13.3 Hz. Results show that in 3.5% NaCl the highest fatigue life (number of cycles to failure Nf) is obtained at pH=5 where the oxide film on aluminum has very low solubility, and the lowest number of cycles is obtained at pH=1, where the media is too aggressive (extremely acidic). When the concentration of inhibitor increases the cycles to failure increase. The surface morphology and fracture section of the specimens had been characterized through scanning electron microscope (SEM).

Keywords: Al-Mg alloy 5052, corrosion, fatigue, inhibitors

Procedia PDF Downloads 441
7528 The Effects of Urbanization on Peri-Urban Livelihood in Ghana: A Case of Kumasi Peri-Urban Communities

Authors: Charles Kwaku Oppong

Abstract:

The research linked urban expansion resulting from urbanization with changing morphology processes happening in peri-urban communities. Two villages of Kumasi City peri-urban were used as a case study. Appropriate analytical framework and methodology (literature review and empirical evidence) were employed to ensure that all pertinent issues of peri-urban interface are brought to light. It was discovered from the study that since peri-urban livelihood is linked with assets base; it has been found that stock of asset, as well as transformation processes, were major factors in the shaping of livelihoods strategies. For that reason, success or failure of household livelihoods was seen to relate to the kind of livelihood strategy employed. With efforts to mitigate for livelihoods failure due to peri-urban development, households' recourse to remittances, land disposal, and other means as an alternative livelihood approach. The study calls for local government policy interventions in regulating peri-urban transformation process and providing safety nets for the vulnerable.

Keywords: urban expansion, peri-urban interface, livelihoods, asset

Procedia PDF Downloads 232
7527 Configuration as a Service in Multi-Tenant Enterprise Resource Planning System

Authors: Mona Misfer Alshardan, Djamal Ziani

Abstract:

Enterprise resource planning (ERP) systems are the organizations tickets to the global market. With the implementation of ERP, organizations can manage and coordinate all functions, processes, resources and data from different departments by a single software. However, many organizations consider the cost of traditional ERP to be expensive and look for alternative affordable solutions within their budget. One of these alternative solutions is providing ERP over a software as a service (SaaS) model. This alternative could be considered as a cost effective solution compared to the traditional ERP system. A key feature of any SaaS system is the multi-tenancy architecture where multiple customers (tenants) share the system software. However, different organizations have different requirements. Thus, the SaaS developers accommodate each tenant’s unique requirements by allowing tenant-level customization or configuration. While customization requires source code changes and in most cases a programming experience, the configuration process allows users to change many features within a predefined scope in an easy and controlled manner. The literature provides many techniques to accomplish the configuration process in different SaaS systems. However, the nature and complexity of SaaS ERP needs more attention to the details regarding the configuration process which is merely described in previous researches. Thus, this research is built on strong knowledge regarding the configuration in SaaS to define specifically the configuration borders in SaaS ERP and to design a configuration service with the consideration of the different configuration aspects. The proposed architecture will ensure the easiness of the configuration process by using wizard technology. Also, the privacy and performance are guaranteed by adopting the databases isolation technique.

Keywords: configuration, software as a service, multi-tenancy, ERP

Procedia PDF Downloads 380
7526 Comparison between Hardy-Cross Method and Water Software to Solve a Pipe Networking Design Problem for a Small Town

Authors: Ahmed Emad Ahmed, Zeyad Ahmed Hussein, Mohamed Salama Afifi, Ahmed Mohammed Eid

Abstract:

Water has a great importance in life. In order to deliver water from resources to the users, many procedures should be taken by the water engineers. One of the main procedures to deliver water to the community is by designing pressurizer pipe networks for water. The main aim of this work is to calculate the water demand of a small town and then design a simple water network to distribute water resources among the town with the smallest losses. Literature has been mentioned to cover the main point related to water distribution. Moreover, the methodology has introduced two approaches to solve the research problem, one by the iterative method of Hardy-cross and the other by water software Pipe Flow. The results have introduced two main designs to satisfy the same research requirements. Finally, the researchers have concluded that the use of water software provides more abilities and options for water engineers.

Keywords: looping pipe networks, hardy cross networks accuracy, relative error of hardy cross method

Procedia PDF Downloads 144
7525 Settlement Prediction in Cape Flats Sands Using Shear Wave Velocity – Penetration Resistance Correlations

Authors: Nanine Fouche

Abstract:

The Cape Flats is a low-lying sand-covered expanse of approximately 460 square kilometres, situated to the southeast of the central business district of Cape Town in the Western Cape of South Africa. The aeolian sands masking this area are often loose and compressible in the upper 1m to 1.5m of the surface, and there is a general exceedance of the maximum allowable settlement in these sands. The settlement of shallow foundations on Cape Flats sands is commonly predicted using the results of in-situ tests such as the SPT or DPSH due to the difficulty of retrieving undisturbed samples for laboratory testing. Varying degrees of accuracy and reliability are associated with these methods. More recently, shear wave velocity (Vs) profiles obtained from seismic testing, such as continuous surface wave tests (CSW), are being used for settlement prediction. Such predictions have the advantage of considering non-linear stress-strain behaviour of soil and the degradation of stiffness with increasing strain. CSW tests are rarely executed in the Cape Flats, whereas SPT’s are commonly performed. For this reason, and to facilitate better settlement predictions in Cape Flats sand, equations representing shear wave velocity (Vs) as a function of SPT blow count (N60) and vertical effective stress (v’) were generated by statistical regression of site investigation data. To reveal the most appropriate method of overburden correction, analyses were performed with a separate overburden term (Pa/σ’v) as well as using stress corrected shear wave velocity and SPT blow counts (correcting Vs. and N60 to Vs1and (N1)60respectively). Shear wave velocity profiles and SPT blow count data from three sites masked by Cape Flats sands were utilised to generate 80 Vs-SPT N data pairs for analysis. Investigated terrains included sites in the suburbs of Athlone, Muizenburg, and Atlantis, all underlain by windblown deposits comprising fine and medium sand with varying fines contents. Elastic settlement analysis was also undertaken for the Cape Flats sands, using a non-linear stepwise method based on small-strain stiffness estimates, which was obtained from the best Vs-N60 model and compared to settlement estimates using the general elastic solution with stiffness profiles determined using Stroud’s (1989) and Webb’s (1969) SPT N60-E transformation models. Stroud’s method considers strain level indirectly whereasWebb’smethod does not take account of the variation in elastic modulus with strain. The expression of Vs. in terms of N60 and Pa/σv’ derived from the Atlantis data set revealed the best fit with R2 = 0.83 and a standard error of 83.5m/s. Less accurate Vs-SPT N relations associated with the combined data set is presumably the result of inversion routines used in the analysis of the CSW results showcasing significant variation in relative density and stiffness with depth. The regression analyses revealed that the inclusion of a separate overburden term in the regression of Vs and N60, produces improved fits, as opposed to the stress corrected equations in which the R2 of the regression is notably lower. It is the correction of Vs and N60 to Vs1 and (N1)60 with empirical constants ‘n’ and ‘m’ prior to regression, that introduces bias with respect to overburden pressure. When comparing settlement prediction methods, both Stroud’s method (considering strain level indirectly) and the small strain stiffness method predict higher stiffnesses for medium dense and dense profiles than Webb’s method, which takes no account of strain level in the determination of soil stiffness. Webb’s method appears to be suitable for loose sands only. The Versak software appears to underestimate differences in settlement between square and strip footings of similar width. In conclusion, settlement analysis using small-strain stiffness data from the proposed Vs-N60 model for Cape Flats sands provides a way to take account of the non-linear stress-strain behaviour of the sands when calculating settlement.

Keywords: sands, settlement prediction, continuous surface wave test, small-strain stiffness, shear wave velocity, penetration resistance

Procedia PDF Downloads 161
7524 Multi-Faceted Growth in Creative Industries

Authors: Sanja Pfeifer, Nataša Šarlija, Marina Jeger, Ana Bilandžić

Abstract:

The purpose of this study is to explore the different facets of growth among micro, small and medium-sized firms in Croatia and to analyze the differences between models designed for all micro, small and medium-sized firms and those in creative industries. Three growth prediction models were designed and tested using the growth of sales, employment and assets of the company as dependent variables. The key drivers of sales growth are: prudent use of cash, industry affiliation and higher share of intangible assets. Growth of assets depends on retained profits, internal and external sources of financing, as well as industry affiliation. Growth in employment is closely related to sources of financing, in particular, debt and it occurs less frequently than growth in sales and assets. The findings confirm the assumption that growth strategies of small and medium-sized enterprises (SMEs) in creative industries have specific differences in comparison to SMEs in general. Interestingly, only 2.2% of growing enterprises achieve growth in employment, assets and sales simultaneously.

Keywords: creative industries, growth prediction model, growth determinants, growth measures

Procedia PDF Downloads 315
7523 Graph Clustering Unveiled: ClusterSyn - A Machine Learning Framework for Predicting Anti-Cancer Drug Synergy Scores

Authors: Babak Bahri, Fatemeh Yassaee Meybodi, Changiz Eslahchi

Abstract:

In the pursuit of effective cancer therapies, the exploration of combinatorial drug regimens is crucial to leverage synergistic interactions between drugs, thereby improving treatment efficacy and overcoming drug resistance. However, identifying synergistic drug pairs poses challenges due to the vast combinatorial space and limitations of experimental approaches. This study introduces ClusterSyn, a machine learning (ML)-powered framework for classifying anti-cancer drug synergy scores. ClusterSyn employs a two-step approach involving drug clustering and synergy score prediction using a fully connected deep neural network. For each cell line in the training dataset, a drug graph is constructed, with nodes representing drugs and edge weights denoting synergy scores between drug pairs. Drugs are clustered using the Markov clustering (MCL) algorithm, and vectors representing the similarity of drug pairs to each cluster are input into the deep neural network for synergy score prediction (synergy or antagonism). Clustering results demonstrate effective grouping of drugs based on synergy scores, aligning similar synergy profiles. Subsequently, neural network predictions and synergy scores of the two drugs on others within their clusters are used to predict the synergy score of the considered drug pair. This approach facilitates comparative analysis with clustering and regression-based methods, revealing the superior performance of ClusterSyn over state-of-the-art methods like DeepSynergy and DeepDDS on diverse datasets such as Oniel and Almanac. The results highlight the remarkable potential of ClusterSyn as a versatile tool for predicting anti-cancer drug synergy scores.

Keywords: drug synergy, clustering, prediction, machine learning., deep learning

Procedia PDF Downloads 59
7522 Development of Risk-Based Dam Safety Framework in Climate Change Condition for Batu Dam, Malaysia

Authors: Wan Noorul Hafilah Binti Wan Ariffin

Abstract:

Dam safety management is the crucial infrastructure as dam failure has a catastrophic effect on the community. Dam safety management is the effective framework of key actions and activities for the dam owner to manage the safety of the dam for its entire life cycle. However, maintaining dam safety is a challenging task as there are changes in current dam states. These changes introduce new risks to the dam's safety, which had not been considered when the dam was designed. A new framework has to be developed to adapt to the changes in the dam risk and make the dams resilient. This study proposes a risk-based decision-making adaptation framework for dam safety management. The research focuses on climate change's impact on hydrological situations as it causes floods and damages the dam structure. The risk analysis framework is adopted to improve the dam management strategies. The proposed study encompasses four phases. To start with, measuring the effect by assessing the impact of climate change on embankment dam, the second phase is to analyze the potential embankment dam failures. The third is analyzing the different components of risks related to the dam and, finally, developing a robust decision-making framework.

Keywords: climate change, embankment dam, failure, risk-informed decision making

Procedia PDF Downloads 133
7521 Comparative Analysis of Predictive Models for Customer Churn Prediction in the Telecommunication Industry

Authors: Deepika Christopher, Garima Anand

Abstract:

To determine the best model for churn prediction in the telecom industry, this paper compares 11 machine learning algorithms, namely Logistic Regression, Support Vector Machine, Random Forest, Decision Tree, XGBoost, LightGBM, Cat Boost, AdaBoost, Extra Trees, Deep Neural Network, and Hybrid Model (MLPClassifier). It also aims to pinpoint the top three factors that lead to customer churn and conducts customer segmentation to identify vulnerable groups. According to the data, the Logistic Regression model performs the best, with an F1 score of 0.6215, 81.76% accuracy, 68.95% precision, and 56.57% recall. The top three attributes that cause churn are found to be tenure, Internet Service Fiber optic, and Internet Service DSL; conversely, the top three models in this article that perform the best are Logistic Regression, Deep Neural Network, and AdaBoost. The K means algorithm is applied to establish and analyze four different customer clusters. This study has effectively identified customers that are at risk of churn and may be utilized to develop and execute strategies that lower customer attrition.

Keywords: attrition, retention, predictive modeling, customer segmentation, telecommunications

Procedia PDF Downloads 40
7520 Implementation of Correlation-Based Data Analysis as a Preliminary Stage for the Prediction of Geometric Dimensions Using Machine Learning in the Forming of Car Seat Rails

Authors: Housein Deli, Loui Al-Shrouf, Hammoud Al Joumaa, Mohieddine Jelali

Abstract:

When forming metallic materials, fluctuations in material properties, process conditions, and wear lead to deviations in the component geometry. Several hundred features sometimes need to be measured, especially in the case of functional and safety-relevant components. These can only be measured offline due to the large number of features and the accuracy requirements. The risk of producing components outside the tolerances is minimized but not eliminated by the statistical evaluation of process capability and control measurements. The inspection intervals are based on the acceptable risk and are at the expense of productivity but remain reactive and, in some cases, considerably delayed. Due to the considerable progress made in the field of condition monitoring and measurement technology, permanently installed sensor systems in combination with machine learning and artificial intelligence, in particular, offer the potential to independently derive forecasts for component geometry and thus eliminate the risk of defective products - actively and preventively. The reliability of forecasts depends on the quality, completeness, and timeliness of the data. Measuring all geometric characteristics is neither sensible nor technically possible. This paper, therefore, uses the example of car seat rail production to discuss the necessary first step of feature selection and reduction by correlation analysis, as otherwise, it would not be possible to forecast components in real-time and inline. Four different car seat rails with an average of 130 features were selected and measured using a coordinate measuring machine (CMM). The run of such measuring programs alone takes up to 20 minutes. In practice, this results in the risk of faulty production of at least 2000 components that have to be sorted or scrapped if the measurement results are negative. Over a period of 2 months, all measurement data (> 200 measurements/ variant) was collected and evaluated using correlation analysis. As part of this study, the number of characteristics to be measured for all 6 car seat rail variants was reduced by over 80%. Specifically, direct correlations for almost 100 characteristics were proven for an average of 125 characteristics for 4 different products. A further 10 features correlate via indirect relationships so that the number of features required for a prediction could be reduced to less than 20. A correlation factor >0.8 was assumed for all correlations.

Keywords: long-term SHM, condition monitoring, machine learning, correlation analysis, component prediction, wear prediction, regressions analysis

Procedia PDF Downloads 20