Search results for: automated equipment
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 2348

Search results for: automated equipment

1778 Statistical Analysis to Compare between Smart City and Traditional Housing

Authors: Taha Anjamrooz, Sareh Rajabi, Ayman Alzaatreh

Abstract:

Smart cities are playing important roles in real life. Integration and automation between different features of modern cities and information technologies improve smart city efficiency, energy management, human and equipment resource management, life quality and better utilization of resources for the customers. One of difficulties in this path, is use, interface and link between software, hardware, and other IT technologies to develop and optimize processes in various business fields such as construction, supply chain management and transportation in parallel to cost-effective and resource reduction impacts. Also, Smart cities are certainly intended to demonstrate a vital role in offering a sustainable and efficient model for smart houses while mitigating environmental and ecological matters. Energy management is one of the most important matters within smart houses in the smart cities and communities, because of the sensitivity of energy systems, reduction in energy wastage and maximization in utilizing the required energy. Specially, the consumption of energy in the smart houses is important and considerable in the economic balance and energy management in smart city as it causes significant increment in energy-saving and energy-wastage reduction. This research paper develops features and concept of smart city in term of overall efficiency through various effective variables. The selected variables and observations are analyzed through data analysis processes to demonstrate the efficiency of smart city and compare the effectiveness of each variable. There are ten chosen variables in this study to improve overall efficiency of smart city through increasing effectiveness of smart houses using an automated solar photovoltaic system, RFID System, smart meter and other major elements by interfacing between software and hardware devices as well as IT technologies. Secondly to enhance aspect of energy management by energy-saving within smart house through efficient variables. The main objective of smart city and smart houses is to reproduce energy and increase its efficiency through selected variables with a comfortable and harmless atmosphere for the customers within a smart city in combination of control over the energy consumption in smart house using developed IT technologies. Initially the comparison between traditional housing and smart city samples is conducted to indicate more efficient system. Moreover, the main variables involved in measuring overall efficiency of system are analyzed through various processes to identify and prioritize the variables in accordance to their influence over the model. The result analysis of this model can be used as comparison and benchmarking with traditional life style to demonstrate the privileges of smart cities. Furthermore, due to expensive and expected shortage of natural resources in near future, insufficient and developed research study in the region, and available potential due to climate and governmental vision, the result and analysis of this study can be used as key indicator to select most effective variables or devices during construction phase and design

Keywords: smart city, traditional housing, RFID, photovoltaic system, energy efficiency, energy saving

Procedia PDF Downloads 104
1777 Urgent Need for E -Waste Management in Mongolia

Authors: Enkhjargal Bat-Ochir

Abstract:

The global market of electrical and electronic equipment (EEE) has increasing rapidly while the lifespan of these products has become increasingly shorter. So, e-waste is becoming the world’s fastest growing waste stream. E-waste is a huge problem when it’s not properly disposed of, as these devices contain substances that are harmful to the environment and to human health as they contaminate the land, water, and air. This paper tends to highlight e-waste problem and harmful effects and can grasp the extent of the problem and take the necessary measures to solve it in Mongolia and to improve standards and human health.

Keywords: e -waste, recycle, electrical, Mongolia

Procedia PDF Downloads 413
1776 A Study of Issues and Mitigations on Distributed Denial of Service and Medical Internet of Things Devices

Authors: Robin Singh, Jing-Chiou Liou

Abstract:

The Internet of Things (IoT) devices are being used heavily as part of our everyday routines. Through improved communication and automated procedures, its popularity has assisted users in raising the quality of work. These devices are used in healthcare in order to better collect the patient’s data for their treatment. They are generally considered safe and secure. However, there is some possibility that some loopholes do exist which manufacturers do need to identify before some hacker takes advantage of them. For this study, we focused on two medical IoT devices which are pacemakers and hearing aids. The aim of this paper is to identify if there is any likelihood of these medical devices being hijacked and used as a botnet in Distributed Denial-Of Service attacks. Moreover, some mitigation strategies are being proposed to better secure

Keywords: cybersecurity, DDoS, IoT, medical devices

Procedia PDF Downloads 77
1775 Study on Eco-Feedback of Thermal Comfort and Cost Efficiency for Low Energy Residence

Authors: Y. Jin, N. Zhang, X. Luo, W. Zhang

Abstract:

China with annual increasing 0.5-0.6 billion squares city residence has brought in enormous energy consumption by HVAC facilities and other appliances. In this regard, governments and researchers are encouraging renewable energy like solar energy, geothermal energy using in houses. However, high cost of equipment and low energy conversion result in a very low acceptable to residents. So what’s the equilibrium point of eco-feedback to reach economic benefit and thermal comfort? That is the main question should be answered. In this paper, the objective is an on-site solar PV and heater house, which has been evaluated as a low energy building. Since HVAC system is considered as main energy consumption equipment, the residence with 24-hour monitoring system set to measure temperature, wind velocity and energy in-out value with no HVAC system for one month of summer and winter. Thermal comfort time period will be analyzed and confirmed; then the air-conditioner will be started within thermal discomfort time for the following one summer and winter month. The same data will be recorded to calculate the average energy consumption monthly for a purpose of whole day thermal comfort. Finally, two analysis work will be done: 1) Original building thermal simulation by computer at design stage with actual measured temperature after construction will be contrastive analyzed; 2) The cost of renewable energy facilities and power consumption converted to cost efficient rate to assess the feasibility of renewable energy input for residence. The results of the experiment showed that a certain deviation exists between actual measured data and simulated one for human thermal comfort, especially in summer period. Moreover, the cost-effectiveness is high for a house in targeting city Guilin now with at least 11 years of cost-covering. The conclusion proves that an eco-feedback of a low energy residence is never only consideration of its energy net value, but also the cost efficiency that is the critical factor to push renewable energy acceptable by the public.

Keywords: cost efficiency, eco-feedback, low energy residence, thermal comfort

Procedia PDF Downloads 251
1774 Extraction of Phycocyanin from Spirulina platensis by Isoelectric Point Precipitation and Salting Out for Scale Up Processes

Authors: Velasco-Rendón María Del Carmen, Cuéllar-Bermúdez Sara Paulina, Parra-Saldívar Roberto

Abstract:

Phycocyanin is a blue pigment protein with fluorescent activity produced by cyanobacteria. It has been recently studied to determine its anticancer, antioxidant and antiinflamatory potential. Since 2014 it was approved as a Generally Recognized As Safe (GRAS) proteic pigment for the food industry. Therefore, phycocyanin shows potential for the food, nutraceutical, pharmaceutical and diagnostics industry. Conventional phycocyanin extraction includes buffer solutions and ammonium sulphate followed by chromatography or ATPS for protein separation. Therefore, further purification steps are time-requiring, energy intensive and not suitable for scale-up processing. This work presents an alternative to conventional methods that also allows large scale application with commercially available equipment. The extraction was performed by exposing the dry biomass to mechanical cavitation and salting out with NaCl to use an edible reagent. Also, isoelectric point precipitation was used by addition of HCl and neutralization with NaOH. The results were measured and compared in phycocyanin concentration, purity and extraction yield. Results showed that the best extraction condition was the extraction by salting out with 0.20 M NaCl after 30 minutes cavitation, with a concentration in the supernatant of 2.22 mg/ml, a purity of 3.28 and recovery from crude extract of 81.27%. Mechanical cavitation presumably increased the solvent-biomass contact, making the crude extract visibly dark blue after centrifugation. Compared to other systems, our process has less purification steps, similar concentrations in the phycocyanin-rich fraction and higher purity. The contaminants present in our process edible NaCl or low pHs that can be neutralized. It also can be adapted to a semi-continuous process with commercially available equipment. This characteristics make this process an appealing alternative for phycocyanin extraction as a pigment for the food industry.

Keywords: extraction, phycocyanin, precipitation, scale-up

Procedia PDF Downloads 431
1773 Business Intelligence Proposal to Improve Decision Making in Companies Using Google Cloud Platform and Microsoft Power BI

Authors: Joel Vilca Tarazona, Igor Aguilar-Alonso

Abstract:

The problem of this research related to business intelligence is the lack of a tool that supports automated and efficient financial analysis for decision-making and allows an evaluation of the financial statements, which is why the availability of the information is difficult. Relevant information to managers and users as an instrument in decision making financial, and administrative. For them, a business intelligence solution is proposed that will reduce information access time, personnel costs, and process automation, proposing a 4-layer architecture based on what was reviewed by the research methodology.

Keywords: decision making, business intelligence, Google Cloud, Microsoft Power BI

Procedia PDF Downloads 91
1772 The Platform for Digitization of Georgian Documents

Authors: Erekle Magradze, Davit Soselia, Levan Shughliashvili, Irakli Koberidze, Shota Tsiskaridze, Victor Kakhniashvili, Tamar Chaghiashvili

Abstract:

Since the beginning of active publishing activity in Georgia, voluminous printed material has been accumulated, the digitization of which is an important task. Digitized materials will be available to the audience, and it will be possible to find text in them and conduct various factual research. Digitizing scanned documents means scanning documents, extracting text from the scanned documents, and processing the text into a corresponding language model to detect inaccuracies and grammatical errors. Implementing these stages requires a unified, scalable, and automated platform, where the digital service developed for each stage will perform the task assigned to it; at the same time, it will be possible to develop these services dynamically so that there is no interruption in the work of the platform.

Keywords: NLP, OCR, BERT, Kubernetes, transformers

Procedia PDF Downloads 136
1771 Optimal Control of Generators and Series Compensators within Multi-Space-Time Frame

Authors: Qian Chen, Lin Xu, Ping Ju, Zhuoran Li, Yiping Yu, Yuqing Jin

Abstract:

The operation of power grid is becoming more and more complex and difficult due to its rapid development towards high voltage, long distance, and large capacity. For instance, many large-scale wind farms have connected to power grid, where their fluctuation and randomness is very likely to affect the stability and safety of the grid. Fortunately, many new-type equipments based on power electronics have been applied to power grid, such as UPFC (Unified Power Flow Controller), TCSC (Thyristor Controlled Series Compensation), STATCOM (Static Synchronous Compensator) and so on, which can help to deal with the problem above. Compared with traditional equipment such as generator, new-type controllable devices, represented by the FACTS (Flexible AC Transmission System), have more accurate control ability and respond faster. But they are too expensive to use widely. Therefore, on the basis of the comparison and analysis of the controlling characteristics between traditional control equipment and new-type controllable equipment in both time and space scale, a coordinated optimizing control method within mutil-time-space frame is proposed in this paper to bring both kinds of advantages into play, which can better both control ability and economical efficiency. Firstly, the coordination of different space sizes of grid is studied focused on the fluctuation caused by large-scale wind farms connected to power grid. With generator, FSC (Fixed Series Compensation) and TCSC, the coordination method on two-layer regional power grid vs. its sub grid is studied in detail. The coordination control model is built, the corresponding scheme is promoted, and the conclusion is verified by simulation. By analysis, interface power flow can be controlled by generator and the specific line power flow between two-layer regions can be adjusted by FSC and TCSC. The smaller the interface power flow adjusted by generator, the bigger the control margin of TCSC, instead, the total consumption of generator is much higher. Secondly, the coordination of different time sizes is studied to further the amount of the total consumption of generator and the control margin of TCSC, where the minimum control cost can be acquired. The coordination method on two-layer ultra short-term correction vs. AGC (Automatic Generation Control) is studied with generator, FSC and TCSC. The optimal control model is founded, genetic algorithm is selected to solve the problem, and the conclusion is verified by simulation. Finally, the aforementioned method within multi-time-space scale is analyzed with practical cases, and simulated on PSASP (Power System Analysis Software Package) platform. The correctness and effectiveness are verified by the simulation result. Moreover, this coordinated optimizing control method can contribute to the decrease of control cost and will provide reference to the following studies in this field.

Keywords: FACTS, multi-space-time frame, optimal control, TCSC

Procedia PDF Downloads 259
1770 Guidelines to Designing Generic Protocol for Responding to Chemical, Biological, Radiological and Nuclear Incidents

Authors: Mohammad H. Yarmohammadian, Mehdi Nasr Isfahani, Elham Anbari

Abstract:

Introduction: The awareness of using chemical, biological, and nuclear agents in everyday industrial and non-industrial incidents has increased recently; release of these materials can be accidental or intentional. Since hospitals are the forefronts of confronting Chemical, Biological, Radiological and Nuclear( CBRN) incidents, the goal of the present research was to provide a generic protocol for CBRN incidents through a comparative review of CBRN protocols and guidelines of different countries and reviewing various books, handbooks and papers. Method: The integrative approach or research synthesis was adopted in this study. First a simple narrative review of programs, books, handbooks, and papers about response to CBRN incidents in different countries was carried out. Then the most important and functional information was discussed in the form of a generic protocol in focus group sessions and subsequently confirmed. Results: Findings indicated that most of the countries had various protocols, guidelines, and handbooks for hazardous materials or CBRN incidents. The final outcome of the research synthesis was a 50 page generic protocol whose main topics included introduction, definition and classification of CBRN agents, four major phases of incident and disaster management cycle, hospital response management plan, equipment, and recommended supplies and antidotes for decontamination (radiological/nuclear, chemical, biological); each of these also had subtopics. Conclusion: In the majority of international protocols, guidelines, handbooks and also international and Iranian books and papers, there is an emphasis on the importance of incident command system, determining the safety degree of decontamination zones, maps of decontamination zones, decontamination process, triage classifications, personal protective equipment, and supplies and antidotes for decontamination; these are the least requirements for such incidents and also consistent with the provided generic protocol.

Keywords: hospital, CBRN, decontamination, generic protocol, CBRN Incidents

Procedia PDF Downloads 291
1769 Music Listening in Dementia: Current Developments and the Potential for Automated Systems in the Home: Scoping Review and Discussion

Authors: Alexander Street, Nina Wollersberger, Paul Fernie, Leonardo Muller, Ming Hung HSU, Helen Odell-Miller, Jorg Fachner, Patrizia Di Campli San Vito, Stephen Brewster, Hari Shaji, Satvik Venkatesh, Paolo Itaborai, Nicolas Farina, Alexis Kirke, Sube Banerjee, Eduardo Reck Miranda

Abstract:

Escalating neuropsychiatric symptoms (NPS) in people with dementia may lead to earlier care home admission. Music listening has been reported to stimulate cognitive function, potentially reducing agitation in this population. We present a scoping review, reporting on current developments and discussing the potential for music listening with related technology in managing agitation in dementia care. Of two searches for music listening studies, one focused on older people or people living with dementia where music listening interventions, including technology, were delivered in participants’ homes or in institutions to address neuropsychiatric symptoms, quality of life and independence. The second included any population focusing on the use of music technology for health and wellbeing. In search one 70/251 full texts were included. The majority reported either statistical significance (6, 8.5%), significance (17, 24.2%) or improvements (26, 37.1%). Agitation was specifically reported in 36 (51.4%). The second search included 51/99 full texts, reporting improvement (28, 54.9%), significance (11, 21.5%), statistical significance (1, 1.9%) and no difference compared to the control (6, 11.7%). The majority in the first focused on mood and agitation, and the second on mood and psychophysiological responses. Five studies used AI or machine learning systems to select music, all involving healthy controls and reporting benefits. Most studies in both reviews were not conducted in a home environment (review 1 = 12; 17.1%; review 2 = 11; 21.5%). Preferred music listening may help manage NPS in the care home settings. Based on these and other data extracted in the review, a reasonable progression would be to co-design and test music listening systems and protocols for NPS in all settings, including people’s homes. Machine learning and automated technology for music selection and arousal adjustment, driven by live biodata, have not been explored in dementia care. Such approaches may help deliver the right music at the appropriate time in the required dosage, reducing the use of medication and improving quality of life.

Keywords: music listening, dementia, agitation, scoping review, technology

Procedia PDF Downloads 108
1768 Stun Practices in Swine in the Valle De Aburrá and Animal Welfare

Authors: Natalia Uribe Corrales, Carolina Cano Arroyave, Santiago Henao Villegas

Abstract:

Introduction: Stunning is an important stage in the meat industry due to the repercussions on the characteristics of the carcass. It has been demonstrated that inadequate stun can lead to hematomas, fractures and promote the appearance of pale, soft and exudative meat due to the stress caused in animals. In Colombia, gas narcosis and electrical stunning are the two authorized methods in pigs. Objective: To describe the practices of stunning in the Valle de Aburrá and its relation with animal welfare. Methods: A descriptive cross - sectional study was carried out in Valle de Aburrá slaughterhouses, which were authorized by National Institute for Food and Medicine Surveillance (INVIMA). Variables such as stunning method, presence of vocalization, falls, slips, rhythmic breathing, corneal reflex and attempts to incorporate after stunning, stun time and time between stun and bleeding were analyzed. Results: 225 pigs were analyzed, finding that 50.2% had electrical stun, whose amperage and voltage were 1.23 (A) and 120 (V) respectively; 49.8% of the animals were stunned with CO2 chamber whose concentration was always above 95%, the mean desensitization time was 16.8 seconds (d.e.5.37); the mean time of stunning - bleeding was 47.9 seconds (d.e.13.9); similarly, it was found that 27.1% had vocalizations after stunning; 12% had falls; 10.7% showed rhythmic breathing; 33.3% exhibited corneal reflex; and 10.7% had reincorporation attempts. Conclusions: The methods of stunning used in the Valle de Aburrá, although performed with those permitted by law, are shortcomings in relation to the amperage and voltage used for each type of pig, as well, it is found that welfare animal is being violated to find signology of an inadequate desensitization. It is necessary to promote compliance with the principles of stunning according to Animal Welfare, and keep in mind that in electrical desensitization, the calibration of the equipment must be guaranteed (pressure according to the type of animal or current applied and the position where the electrodes are) and in the narcosis the equipment should be calibrated to ensure proper gas concentration and exposure time.

Keywords: animal welfare, pigs, quality of meat, stun methods

Procedia PDF Downloads 221
1767 Building and Tree Detection Using Multiscale Matched Filtering

Authors: Abdullah H. Özcan, Dilara Hisar, Yetkin Sayar, Cem Ünsalan

Abstract:

In this study, an automated building and tree detection method is proposed using DSM data and true orthophoto image. A multiscale matched filtering is used on DSM data. Therefore, first watershed transform is applied. Then, Otsu’s thresholding method is used as an adaptive threshold to segment each watershed region. Detected objects are masked with NDVI to separate buildings and trees. The proposed method is able to detect buildings and trees without entering any elevation threshold. We tested our method on ISPRS semantic labeling dataset and obtained promising results.

Keywords: building detection, local maximum filtering, matched filtering, multiscale

Procedia PDF Downloads 314
1766 Profiling Risky Code Using Machine Learning

Authors: Zunaira Zaman, David Bohannon

Abstract:

This study explores the application of machine learning (ML) for detecting security vulnerabilities in source code. The research aims to assist organizations with large application portfolios and limited security testing capabilities in prioritizing security activities. ML-based approaches offer benefits such as increased confidence scores, false positives and negatives tuning, and automated feedback. The initial approach using natural language processing techniques to extract features achieved 86% accuracy during the training phase but suffered from overfitting and performed poorly on unseen datasets during testing. To address these issues, the study proposes using the abstract syntax tree (AST) for Java and C++ codebases to capture code semantics and structure and generate path-context representations for each function. The Code2Vec model architecture is used to learn distributed representations of source code snippets for training a machine-learning classifier for vulnerability prediction. The study evaluates the performance of the proposed methodology using two datasets and compares the results with existing approaches. The Devign dataset yielded 60% accuracy in predicting vulnerable code snippets and helped resist overfitting, while the Juliet Test Suite predicted specific vulnerabilities such as OS-Command Injection, Cryptographic, and Cross-Site Scripting vulnerabilities. The Code2Vec model achieved 75% accuracy and a 98% recall rate in predicting OS-Command Injection vulnerabilities. The study concludes that even partial AST representations of source code can be useful for vulnerability prediction. The approach has the potential for automated intelligent analysis of source code, including vulnerability prediction on unseen source code. State-of-the-art models using natural language processing techniques and CNN models with ensemble modelling techniques did not generalize well on unseen data and faced overfitting issues. However, predicting vulnerabilities in source code using machine learning poses challenges such as high dimensionality and complexity of source code, imbalanced datasets, and identifying specific types of vulnerabilities. Future work will address these challenges and expand the scope of the research.

Keywords: code embeddings, neural networks, natural language processing, OS command injection, software security, code properties

Procedia PDF Downloads 100
1765 A Simple Approach to Establish Urban Energy Consumption Map Using the Combination of LiDAR and Thermal Image

Authors: Yu-Cheng Chen, Tzu-Ping Lin, Feng-Yi Lin, Chih-Yu Chen

Abstract:

Due to the urban heat island effect caused by highly development of city, the heat stress increased in recent year rapidly. Resulting in a sharp raise of the energy used in urban area. The heat stress during summer time exacerbated the usage of air conditioning and electric equipment, which caused more energy consumption and anthropogenic heat. Therefore, an accurate and simple method to measure energy used in urban area can be helpful for the architectures and urban planners to develop better energy efficiency goals. This research applies the combination of airborne LiDAR data and thermal imager to provide an innovate method to estimate energy consumption. Owing to the high resolution of remote sensing data, the accurate current volume and total floor area and the surface temperature of building derived from LiDAR and thermal imager can be herein obtained to predict energy used. In the estimate process, the LiDAR data will be divided into four type of land cover which including building, road, vegetation, and other obstacles. In this study, the points belong to building were selected to overlay with the land use information; therefore, the energy consumption can be estimated precisely with the real value of total floor area and energy use index for different use of building. After validating with the real energy used data from the government, the result shows the higher building in high development area like commercial district will present in higher energy consumption, caused by the large quantity of total floor area and more anthropogenic heat. Furthermore, because of the surface temperature can be warm up by electric equipment used, this study also applies the thermal image of building to find the hot spots of energy used and make the estimation method more complete.

Keywords: urban heat island, urban planning, LiDAR, thermal imager, energy consumption

Procedia PDF Downloads 234
1764 Implementation and Challenges of Assessment Methods in the Case of Physical Education Class in Some Selected Preparatory Schools of Kirkos Sub-City

Authors: Kibreab Alene Fenite

Abstract:

The purpose of this study is to investigate the implementation and challenges of different assessment methods for physical education class in some selected preparatory schools of kirkos sub city. The participants in this study are teachers, students, department heads and school principals from 4 selected schools. Of the total 8 schools offering in kirkos sub city 4 schools (Dandi Boru, Abiyot Kirse, Assay, and Adey Ababa) are selected by using simple random sampling techniques and from these schools all (100%) of teachers, 100% of department heads and school principals are taken as a sample as their number is manageable. From the total 2520 students, 252 (10%) of students are selected using simple random sampling. Accordingly, 13 teachers, 252 students, 4 department heads and 4 school principals are taken as a sample from the 4 selected schools purposefully. As a method of data gathering tools; questionnaire and interview are employed. To analyze the collected data, both quantitative and qualitative methods are used. The result of the study revealed that assessment in physical education does not implement properly: lack of sufficient materials, inadequate time allotment, large class size, and lack of collaboration and working together of teachers towards assessing the performance of students, absence of guidelines to assess the physical education subject, no different assessment method that is implementing on students with disabilities in line with their special need are found as major challenges in implementing the current assessment method of physical education. To overcome these problems the following recommendations have been forwarded. These are: the necessary facilities and equipment should be available; In order to make reliable, accurate, objective and relevant assessment, teachers of physical education should be familiarized with different assessment techniques; Physical education assessment guidelines should be prepared, and guidelines should include different types of assessment methods; qualified teachers should be employed, and different teaching room must be build.

Keywords: assessment, challenges, equipment, guidelines, implementation, performance

Procedia PDF Downloads 276
1763 Automated Transformation of 3D Point Cloud to BIM Model: Leveraging Algorithmic Modeling for Efficient Reconstruction

Authors: Radul Shishkov, Orlin Davchev

Abstract:

The digital era has revolutionized architectural practices, with building information modeling (BIM) emerging as a pivotal tool for architects, engineers, and construction professionals. However, the transition from traditional methods to BIM-centric approaches poses significant challenges, particularly in the context of existing structures. This research introduces a technical approach to bridge this gap through the development of algorithms that facilitate the automated transformation of 3D point cloud data into detailed BIM models. The core of this research lies in the application of algorithmic modeling and computational design methods to interpret and reconstruct point cloud data -a collection of data points in space, typically produced by 3D scanners- into comprehensive BIM models. This process involves complex stages of data cleaning, feature extraction, and geometric reconstruction, which are traditionally time-consuming and prone to human error. By automating these stages, our approach significantly enhances the efficiency and accuracy of creating BIM models for existing buildings. The proposed algorithms are designed to identify key architectural elements within point clouds, such as walls, windows, doors, and other structural components, and to translate these elements into their corresponding BIM representations. This includes the integration of parametric modeling techniques to ensure that the generated BIM models are not only geometrically accurate but also embedded with essential architectural and structural information. Our methodology has been tested on several real-world case studies, demonstrating its capability to handle diverse architectural styles and complexities. The results showcase a substantial reduction in time and resources required for BIM model generation while maintaining high levels of accuracy and detail. This research contributes significantly to the field of architectural technology by providing a scalable and efficient solution for the integration of existing structures into the BIM framework. It paves the way for more seamless and integrated workflows in renovation and heritage conservation projects, where the accuracy of existing conditions plays a critical role. The implications of this study extend beyond architectural practices, offering potential benefits in urban planning, facility management, and historic preservation.

Keywords: BIM, 3D point cloud, algorithmic modeling, computational design, architectural reconstruction

Procedia PDF Downloads 50
1762 Study on Optimization of Air Infiltration at Entrance of a Commercial Complex in Zhejiang Province

Authors: Yujie Zhao, Jiantao Weng

Abstract:

In the past decade, with the rapid development of China's economy, the purchasing power and physical demand of residents have been improved, which results in the vast emergence of public buildings like large shopping malls. However, the architects usually focus on the internal functions and streamlines of these buildings, ignoring the impact of the environment on the subjective feelings of building users. Only in Zhejiang province, the infiltration of cold air in winter frequently occurs at the entrance of sizeable commercial complex buildings that have been in operation, which will affect the environmental comfort of the building lobby and internal public spaces. At present, to reduce these adverse effects, it is usually adopted to add active equipment, such as setting air curtains to block air exchange or adding heating air conditioners. From the perspective of energy consumption, the infiltration of cold air into the entrance will increase the heat consumption of indoor heating equipment, which will indirectly cause considerable economic losses during the whole winter heating stage. Therefore, it is of considerable significance to explore the suitable entrance forms for improving the environmental comfort of commercial buildings and saving energy. In this paper, a commercial complex with apparent cold air infiltration problem in Hangzhou is selected as the research object to establish a model. The environmental parameters of the building entrance, including temperature, wind speed, and infiltration air volume, are obtained by Computational Fluid Dynamics (CFD) simulation, from which the heat consumption caused by the natural air infiltration in the winter and its potential economic loss is estimated as the objective metric. This study finally obtains the optimization direction of the building entrance form of the commercial complex by comparing the simulation results of other local commercial complex projects with different entrance forms. The conclusions will guide the entrance design of the same type of commercial complex in this area.

Keywords: air infiltration, commercial complex, heat consumption, CFD simulation

Procedia PDF Downloads 129
1761 Inelastic and Elastic Taping in Plantar Pressure of Runners Pronators: Clinical Trial

Authors: Liana Gomide, Juliana Rodrigues

Abstract:

The morphology of the foot defines its mode of operation and a biomechanical reform indispensable for a symmetrical distribution of plantar pressures in order not to overload some of its components in isolation. High plantar pressures at specific points in the foot may be a causal factor in several orthopedic disorders that affect the feet such as pain and stress fracture. With digital baro-podometry equipment one can observe an intensity of pressures along the entire foot and quantify some of the movements, such as a subtalar pronation present in the midfoot region. Although, they are involved in microtraumas. In clinical practice, excessive movement has been limited with the use of different taping techniques applied on the plantar arch. Thus, the objective of the present study was to analyze and compare the influence of the inelastic and elastic taping on the distribution of plantar pressure of runners pronators. This is a randomized clinical trial and blind-crossover. Twenty (20) male subjects, mean age 33 ± 7 years old, mean body mass of 71 ± 7 kg, mean height of 174 ± 6 cm, were included in the study. A data collection was carried out by a single research through barop-odometry equipment - Tekscan, model F-scan mobile. The tests were performed at three different times. In the first, an initial barop-odometric evaluation was performed, without a bandage application, with edges at a speed of 9.0 km/h. In the second and third moments, the inelastic or elastic taping was applied consecutively, according to the definition defined in the randomization. As results, it was observed that both as inelastic and elastic taping, provided significant reductions in contact pressure and peak pressure values when compared to the moment without a taping. However, an elastic taping was more effective in decreasing contact pressure (no bandage = 714 ± 201, elastic taping = 690 ± 210 and inelastic taping = 716 ± 180) and no peak pressure in the midfoot region (no bandage = 1490 ± 42, elastic taping = 1273 ± 323 and inelastic taping = 1487 ± 437). It is possible to conclude that it is an elastic taping provided by pressure in the middle region, thereby reducing the subtalar pronunciation event during the run.

Keywords: elastic taping, inelastic taping, running, subtalar pronation

Procedia PDF Downloads 145
1760 Designing an App to Solve Surveying Challenges

Authors: Ali Mohammadi

Abstract:

Forming and equipping the surveyors team for construction projects such as dams, roads, and tunnels is always one of the first challenges and hiring surveyors who are proficient in reading maps and controlling structures, purchasing appropriate surveying equipment that the employer can find Also, using methods that can save time, in the bigger the project, the more these challenges show themselves. Finding a surveyor engineer who can lead the teams and train surveyors of the collection and buy TOTAL STATION according to the company's budget and the surveyors' ability to use them and the time available to each team In the following, we will introduce a surveying app and examine how to use it, which shows how useful it can be for surveyors in projects.

Keywords: DTM CUTFILL, datatransfer, section, tunnel, traverse

Procedia PDF Downloads 73
1759 Design and Development of Motorized Placer for Balloon Uterine Stents in Gynecology

Authors: Metehan Mutlu, Meltem Elitas

Abstract:

This study aims to provide an automated method for placing the balloon uterine stents after hysteroscopy adhesiolysis. Currently, there are no automatized tools to place the balloon uterine stent; therefore, surgeons into the endometrial cavity manually fit it. However, it is very hard to pass the balloon stent through the cervical canal, which is roughly 10mm after the surgery. Our method aims to provide an effective and practical way of placing the stent, by automating the procedure through our designed device. Furthermore, our device does the required tasks fast compared to traditional methods, reduces the narcosis time, and decreases the bacterial contamination risks.

Keywords: balloon uterine stent, endometrial cavity, hysteroscopy, motorized-tool

Procedia PDF Downloads 274
1758 A Profile of Out-of-Hospital Cardiac Arrest in ‘Amang’ Rodriguez Memorial Medical Center: A Prospective Cohort Study

Authors: Donna Erika E. De Jesus

Abstract:

Introduction: Cardiac arrest occurs when abrupt cessation of cardiac function results in loss of effective circulation and complete cardiovascular collapse. For every minute of cardiac arrest without early intervention (cardiopulmonary resuscitation [CPR], defibrillation), chances of survival drop by 7-10%. It is crucial that CPR be initiated within 4-6 minutes to avoid brain death. Most out-of-hospital cardiac arrests (OHCA) occur in a residential setting where access to trained personnel and equipment is not readily available, resulting in poor victim outcomes. Methods: This is a descriptive study done from August to November 2021 using a prospective cohort design. Participants of the study include adult patients aged 18 years and above brought to the emergency room who suffered from out-of-hospital cardiac arrest. Out of the total 102 cases of OHCA, 63 participants were included in the study. Descriptive statistics were used to summarize the demographic and clinical characteristics of the patients. Results: 43 were male patients, comprising the majority at 73.02%. Hypertension was identified as the top co-morbidity, followed by diabetes mellitus, heart failure, and chronic kidney disease (CKD). Medical causes of arrest were identified in 96.83% of the cases. 90.48% of cardiac arrests occurred at home. Only 26 patients (41.27%) received pre-hospital intervention prior to ER arrival, which comprised only hands-only CPR. Twenty-three of which were performed by individuals with background knowledge of CPR. 60.32% were brought via self-conduction, the remainder by ambulances, which were noted to have no available equipment necessary to provide proper resuscitation. The average travel time from dispatch to ER arrival is 20 minutes. Conclusion: Overall survival of OHCA in our local setting remains dismal, as a return of spontaneous circulation was not achieved in any of the patients. The small number of patients having pre-hospital CPR indicates the need for emphasis on training and community education.

Keywords: out-of-hospital cardiac arrest, cardiopulmonary resuscitation, basic life support, emergency medical services

Procedia PDF Downloads 102
1757 Non-Conformance Clearance through an Intensified Mentorship towards ISO 15189 Accreditation: The Case of Jimma and Hawassa Hospital Microbiology Laboratories, Ethiopia

Authors: Dawit Assefa, Kassaye Tekie, Gebrie Alebachew, Degefu Beyene, Bikila Alemu, Naji Mohammed, Asnakech Agegnehu, Seble Tsehay, Geremew Tasew

Abstract:

Background: Implementation of a Laboratory Quality Management System (LQMS) is critical to ensure accurate, reliable, and efficient laboratory testing of antimicrobial resistance (AMR). However, limited LQMS implementation and progress toward accreditation in the AMR surveillance laboratory testing setting exist in Ethiopia. By addressing non-conformances (NCs) and working towards accreditation, microbiology laboratories can improve the quality of their services, increase staff competence, and contribute to mitigate the spread of AMR. Methods: Using standard ISO 15189 horizontal and vertical assessment checklists, certified assessors identified NCs at Hawassa and Jimma Hospital microbiology laboratories. The Ethiopian Public Health Institute AMR mentors and IDDS staff prioritized closing the NCs through the implementation of an intensified mentorship program that included ISO 15189 orientation training, resource allocation, and action plan development. Results: For the two facilities to clear their NCs, an intensified mentorship approach was adopted by providing ISO 15189 orientation training, provision of buffer reagents, controls, standards, and axillary equipment, and facilitating equipment maintenance and calibration. Method verification and competency assessment were also conducted along with the implementation of standard operating procedures and recommended corrective actions. This approach enhanced the laboratory's readiness for accreditation. After addressing their NCs, the two laboratories applied to Ethiopian Accreditation Services for ISO 15189 accreditation. Conclusions: Clearing NCs through the implementation of intensified mentorship was crucial in preparing the two laboratories for accreditation and improving quality laboratory test results. This approach can guide other microbiology laboratories’ accreditation attainment efforts.

Keywords: non-conformance clearance, intensified mentorship, accreditation, ISO 15189

Procedia PDF Downloads 74
1756 Correlates of Income Generation of Small-Scale Fish Processors in Abeokuta Metropolis, Ogun State, Nigeria

Authors: Ayodeji Motunrayo Omoare

Abstract:

Economically fish provides an important source of food and income for both men and women especially many households in the developing world and fishing has an important social and cultural position in river-rine communities. However, fish is highly susceptible to deterioration. Consequently, this study was carried out to correlate income generation of small-scale women fish processors in Abeokuta metropolis, Ogun State, Nigeria. Eighty small-scale women fish processors were randomly selected from five communities as the sample size for this study. Collected data were analyzed using both descriptive and inferential statistics. The results showed that the mean age of the respondents was 31.75 years with average household size of 4 people while 47.5% of the respondents had primary education. Most (86.3%) of the respondents were married and had spent more than 11 years in fish processing. The respondents were predominantly Yoruba tribe (91.2%). Majority (71.3%) of the respondents used traditional kiln for processing their fish while 23.7% of the respondents used hot vegetable oil to fry their fish. Also, the result revealed that respondents sourced capital from Personal Savings (48.8%), Cooperatives (27.5%), Friends and Family (17.5%) and Microfinance Banks (6.2%) for fish processing activities. The respondents generated an average income of ₦7,000.00 from roasted fish, ₦3,500.00 from dried fish, and ₦5,200.00 from fried fish daily. However, inadequate processing equipment (95.0%), non-availability of credit facility from microfinance banks (85.0%), poor electricity supply (77.5%), inadequate extension service support (70.0%), and fuel scarcity (68.7%) were major constraints to fish processing in the study area. Results of chi-square analysis showed that there was a significant relationship between personal characteristics (χ2 = 36.83, df = 9), processing methods (χ2 = 15.88, df = 3) and income generated at p < 0.05 level of significance. It can be concluded that significant relationship existed between processing methods and income generated. The study, therefore, recommends that modern processing equipment should be made available to the respondents at a subsidized price by the agro-allied companies.

Keywords: correlates, income, fish processors, women, small-scale

Procedia PDF Downloads 237
1755 Enhancing Engineering Students Educational Experience: Studying Hydrostatic Pumps Association System in Fluid Mechanics Laboratories

Authors: Alexandre Daliberto Frugoli, Pedro Jose Gabriel Ferreira, Pedro Americo Frugoli, Lucio Leonardo, Thais Cavalheri Santos

Abstract:

Laboratory classes in Engineering courses are essential for students to be able to integrate theory with practical reality, by handling equipment and observing experiments. In the researches of physical phenomena, students can learn about the complexities of science. Over the past years, universities in developing countries have been reducing the course load of engineering courses, in accordance with cutting cost agendas. Quality education is the object of study for researchers and requires educators and educational administrators able to demonstrate that the institutions are able to provide great learning opportunities at reasonable costs. Didactic test benches are indispensable equipment in educational activities related to turbo hydraulic pumps and pumping facilities study, which have a high cost and require long class time due to measurements and equipment adjustment time. In order to overcome the aforementioned obstacles, aligned with the professional objectives of an engineer, GruPEFE - UNIP (Research Group in Physics Education for Engineering - Universidade Paulista) has developed a multi-purpose stand for the discipline of fluid mechanics which allows the study of velocity and flow meters, loads losses and pump association. In this work, results obtained by the association in series and in parallel of hydraulic pumps will be presented and discussed, mainly analyzing the repeatability of experimental procedures and their agreement with the theory. For the association in series two identical pumps were used, consisting of the connection of the discharge of a pump to the suction of the next one, allowing the fluid to receive the power of all machines in the association. The characteristic curve of the set is obtained from the curves of each of the pumps, by adding the heads corresponding to the same flow rates. The same pumps were associated in parallel. In this association, the discharge piping is common to the two machines together. The characteristic curve of the set was obtained by adding to each value of H (head height), the flow rates of each pump. For the tests, the input and output pressure of each pump were measured. For each set there were three sets of measurements, varying the flow rate in range from 6.0 to 8.5 m 3 / h. For the two associations, the results showed an excellent repeatability with variations of less than 10% between sets of measurements and also a good agreement with the theory. This variation agrees with the instrumental uncertainty. Thus, the results validate the use of the fluids bench designed for didactic purposes. As a future work, a digital acquisition system is being developed, using differential sensors of extremely low pressures (2 to 2000 Pa approximately) for the microcontroller Arduino.

Keywords: engineering education, fluid mechanics, hydrostatic pumps association, multi-purpose stand

Procedia PDF Downloads 215
1754 The Efficiency of Mechanization in Weed Control in Artificial Regeneration of Oriental Beech (Fagus orientalis Lipsky.)

Authors: Tuğrul Varol, Halil Barış Özel

Abstract:

In this study which has been conducted in Akçasu Forest Range District of Devrek Forest Directorate; 3 methods (cover removal with human force, cover removal with Hitachi F20 Excavator, and cover removal with agricultural equipment mounted on a Ferguson 240S agriculture tractor) utilized in weed control efforts in regeneration of degraded oriental beech forests have been compared. In this respect, 3 methods have been compared by determining certain work hours and standard durations of unit areas (1 hectare). For this purpose, evaluating the tasks made with human and machine force from the aspects of duration, productivity and costs, it has been aimed to determine the most productive method in accordance with the actual ecological conditions of research field. Within the scope of the study, the time studies have been conducted for 3 methods used in weed control efforts. While carrying out those studies, the performed implementations have been evaluated by dividing them into business stages. Also, the actual data have been used while calculating the cost accounts. In those calculations, the latest formulas and equations which are also used in developed countries have been utilized. The variance of analysis (ANOVA) was used in order to determine whether there is any statistically significant difference among obtained results, and the Duncan test was used for grouping if there is significant difference. According to the measurements and findings carried out within the scope of this study, it has been found during living cover removal efforts in regeneration efforts in demolished oriental beech forests that the removal of weed layer in 1 hectare of field has taken 920 hours with human force, 15.1 hours with excavator and 60 hours with an equipment mounted on a tractor. On the other hand, it has been determined that the cost of removal of living cover in unit area (1 hectare) was 3220.00 TL for man power, 788.70 TL for excavator and 2227.20 TL for equipment mounted on a tractor. According to the obtained results, it has been found that the utilization of excavator in weed control effort in regeneration of degraded oriental beech regions under actual ecological conditions of research field has been found to be more productive from both of aspects of duration and costs. These determinations carried out should be repeated in weed control efforts in degraded forest fields with different ecological conditions, it is compulsory for finding the most efficient weed control method. These findings will light the way of technical staff of forestry directorate in determination of the most effective and economic weed contol method. Thus, the more actual data will be used while preparing the weed control budgets, and there will be significant contributions to national economy. Also the results of this and similar studies are very important for developing the policies for our forestry in short and long term.

Keywords: artificial regeneration, weed control, oriental beech, productivity, mechanization, man power, cost analysis

Procedia PDF Downloads 412
1753 Intelligent Recognition Tools for Industrial Automation

Authors: Amin Nazerzadeh, Afsaneh Nouri Houshyar , Azadeh Noori Hoshyar

Abstract:

With the rapid growing of information technology, the industry and manufacturing systems are becoming more automated. Therefore, achieving the highly accurate automatic systems with reliable security is becoming more critical. Biometrics that refers to identifying individual based on physiological or behavioral traits are unique identifiers provide high reliability and security in different industrial systems. As biometric cannot easily be transferred between individuals or copied, it has been receiving extensive attention. Due to the importance of security applications, this paper provides an overview on biometrics and discuss about background, types and applications of biometric as an effective tool for the industrial applications.

Keywords: Industial and manufacturing applications, intelligence and security, information technology, recognition; security technology; biometrics

Procedia PDF Downloads 149
1752 Integration Process and Analytic Interface of different Environmental Open Data Sets with Java/Oracle and R

Authors: Pavel H. Llamocca, Victoria Lopez

Abstract:

The main objective of our work is the comparative analysis of environmental data from Open Data bases, belonging to different governments. This means that you have to integrate data from various different sources. Nowadays, many governments have the intention of publishing thousands of data sets for people and organizations to use them. In this way, the quantity of applications based on Open Data is increasing. However each government has its own procedures to publish its data, and it causes a variety of formats of data sets because there are no international standards to specify the formats of the data sets from Open Data bases. Due to this variety of formats, we must build a data integration process that is able to put together all kind of formats. There are some software tools developed in order to give support to the integration process, e.g. Data Tamer, Data Wrangler. The problem with these tools is that they need data scientist interaction to take part in the integration process as a final step. In our case we don’t want to depend on a data scientist, because environmental data are usually similar and these processes can be automated by programming. The main idea of our tool is to build Hadoop procedures adapted to data sources per each government in order to achieve an automated integration. Our work focus in environment data like temperature, energy consumption, air quality, solar radiation, speeds of wind, etc. Since 2 years, the government of Madrid is publishing its Open Data bases relative to environment indicators in real time. In the same way, other governments have published Open Data sets relative to the environment (like Andalucia or Bilbao). But all of those data sets have different formats and our solution is able to integrate all of them, furthermore it allows the user to make and visualize some analysis over the real-time data. Once the integration task is done, all the data from any government has the same format and the analysis process can be initiated in a computational better way. So the tool presented in this work has two goals: 1. Integration process; and 2. Graphic and analytic interface. As a first approach, the integration process was developed using Java and Oracle and the graphic and analytic interface with Java (jsp). However, in order to open our software tool, as second approach, we also developed an implementation with R language as mature open source technology. R is a really powerful open source programming language that allows us to process and analyze a huge amount of data with high performance. There are also some R libraries for the building of a graphic interface like shiny. A performance comparison between both implementations was made and no significant differences were found. In addition, our work provides with an Official Real-Time Integrated Data Set about Environment Data in Spain to any developer in order that they can build their own applications.

Keywords: open data, R language, data integration, environmental data

Procedia PDF Downloads 306
1751 Learning the Most Common Causes of Major Industrial Accidents and Apply Best Practices to Prevent Such Accidents

Authors: Rajender Dahiya

Abstract:

Investigation outcomes of major process incidents have been consistent for decades and validate that the causes and consequences are often identical. The debate remains as we continue to experience similar process incidents even with enormous development of new tools, technologies, industry standards, codes, regulations, and learning processes? The objective of this paper is to investigate the most common causes of major industrial incidents and reveal industry challenges and best practices to prevent such incidents. The author, in his current role, performs audits and inspections of a variety of high-hazard industries in North America, including petroleum refineries, chemicals, petrochemicals, manufacturing, etc. In this paper, he shares real life scenarios, examples, and case studies from high hazards operating facilities including key challenges and best practices. This case study will provide a clear understanding of the importance of near miss incident investigation. The incident was a Safe operating limit excursion. The case describes the deficiencies in management programs, the competency of employees, and the culture of the corporation that includes hazard identification and risk assessment, maintaining the integrity of safety-critical equipment, operating discipline, learning from process safety near misses, process safety competency, process safety culture, audits, and performance measurement. Failure to identify the hazards and manage the risks of highly hazardous materials and processes is one of the primary root-causes of an incident, and failure to learn from past incidents is the leading cause of the recurrence of incidents. Several investigations of major incidents discovered that each showed several warning signs before occurring, and most importantly, all were preventable. The author will discuss why preventable incidents were not prevented and review the mutual causes of learning failures from past major incidents. The leading causes of past incidents are summarized below. Management failure to identify the hazard and/or mitigate the risk of hazardous processes or materials. This process starts early in the project stage and continues throughout the life cycle of the facility. For example, a poorly done hazard study such as HAZID, PHA, or LOPA is one of the leading causes of the failure. If this step is performed correctly, then the next potential cause is. Management failure to maintain the integrity of safety critical systems and equipment. In most of the incidents, mechanical integrity of the critical equipment was not maintained, safety barriers were either bypassed, disabled, or not maintained. The third major cause is Management failure to learn and/or apply learning from the past incidents. There were several precursors before those incidents. These precursors were either ignored altogether or not taken seriously. This paper will conclude by sharing how a well-implemented operating management system, good process safety culture, and competent leaders and staff contributed to managing the risks to prevent major incidents.

Keywords: incident investigation, risk management, loss prevention, process safety, accident prevention

Procedia PDF Downloads 50
1750 An Effort at Improving Reliability of Laboratory Data in Titrimetric Analysis for Zinc Sulphate Tablets Using Validated Spreadsheet Calculators

Authors: M. A. Okezue, K. L. Clase, S. R. Byrn

Abstract:

The requirement for maintaining data integrity in laboratory operations is critical for regulatory compliance. Automation of procedures reduces incidence of human errors. Quality control laboratories located in low-income economies may face some barriers in attempts to automate their processes. Since data from quality control tests on pharmaceutical products are used in making regulatory decisions, it is important that laboratory reports are accurate and reliable. Zinc Sulphate (ZnSO4) tablets is used in treatment of diarrhea in pediatric population, and as an adjunct therapy for COVID-19 regimen. Unfortunately, zinc content in these formulations is determined titrimetrically; a manual analytical procedure. The assay for ZnSO4 tablets involves time-consuming steps that contain mathematical formulae prone to calculation errors. To achieve consistency, save costs, and improve data integrity, validated spreadsheets were developed to simplify the two critical steps in the analysis of ZnSO4 tablets: standardization of 0.1M Sodium Edetate (EDTA) solution, and the complexometric titration assay procedure. The assay method in the United States Pharmacopoeia was used to create a process flow for ZnSO4 tablets. For each step in the process, different formulae were input into two spreadsheets to automate calculations. Further checks were created within the automated system to ensure validity of replicate analysis in titrimetric procedures. Validations were conducted using five data sets of manually computed assay results. The acceptance criteria set for the protocol were met. Significant p-values (p < 0.05, α = 0.05, at 95% Confidence Interval) were obtained from students’ t-test evaluation of the mean values for manual-calculated and spreadsheet results at all levels of the analysis flow. Right-first-time analysis and principles of data integrity were enhanced by use of the validated spreadsheet calculators in titrimetric evaluations of ZnSO4 tablets. Human errors were minimized in calculations when procedures were automated in quality control laboratories. The assay procedure for the formulation was achieved in a time-efficient manner with greater level of accuracy. This project is expected to promote cost savings for laboratory business models.

Keywords: data integrity, spreadsheets, titrimetry, validation, zinc sulphate tablets

Procedia PDF Downloads 165
1749 An Empirical Study of the Moderation Effects of Commitment, Trust, and Relationship Value in the Relation of Goods and Services Related to Business to Business Brand Images on Customer Loyalty

Authors: Jorge Luis Morales Romero, Enrique Murillo Othón

Abstract:

Business to business (B2B) relationships generally go beyond a purely profit-based result, with firms seeking to maintain a relationship for many years because a breakup or getting a new supplier can be very costly. Therefore, identifying the factors which determine a successful relationship in the long term is of great interest to companies. That is why their reputation and the brand image that customers have of them are among the main factors that can achieve a successful relationship; Because of the positive effect which is driven by the client’s loyalty. Additionally, the perception that a customer may have about a brand is different when it is related to goods or to services. Thereby, they create in their minds their own brand image of it based on the past experiences they have had; Thus, a positive relationship is established between goods-related brand image, service-related brand image, and customer loyalty. The present investigation examines the boundary conditions of said relationship by testing the moderating effects of trust, commitment, and relationship value in a B2B environment. All the variables were tested independently as moderators for service-related brand image/loyalty and for goods-related brand image/loyalty, as they are assumed to be separate variables. Survey data was collected through interviews with customers that have both a product-buying relationship and a service relationship with a global B2B brand of healthcare equipment operating in the Mexican healthcare market. Interviewed respondents were either the user or the purchasing manager and/or the responsible for the equipment maintenance for the customer organization. Hence, they were appropriate informants regarding the B2B relationship with this healthcare brand. The moderation models were estimated using the PROCESS macro for the Statistical Package for the Social Sciences Software (SPSS). Results show statistical evidence that both Relationship Value and Trust are significant moderators for the service-related brand image/loyalty relation but not significant for the goods-related brand/loyalty relation. On the other hand, Commitment results in a significant moderator for the goods-related brand/loyalty relation but is not significant for the service-related brand image/loyalty relation.

Keywords: commitment, trust, relationship value, loyalty, B2B, moderator

Procedia PDF Downloads 87