Search results for: optimized asset allocation
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 2588

Search results for: optimized asset allocation

488 Using Google Distance Matrix Application Programming Interface to Reveal and Handle Urban Road Congestion Hot Spots: A Case Study from Budapest

Authors: Peter Baji

Abstract:

In recent years, a growing body of literature emphasizes the increasingly negative impacts of urban road congestion in the everyday life of citizens. Although there are different responses from the public sector to decrease traffic congestion in urban regions, the most effective public intervention is using congestion charges. Because travel is an economic asset, its consumption can be controlled by extra taxes or prices effectively, but this demand-side intervention is often unpopular. Measuring traffic flows with the help of different methods has a long history in transport sciences, but until recently, there was not enough sufficient data for evaluating road traffic flow patterns on the scale of an entire road system of a larger urban area. European cities (e.g., London, Stockholm, Milan), in which congestion charges have already been introduced, designated a particular zone in their downtown for paying, but it protects only the users and inhabitants of the CBD (Central Business District) area. Through the use of Google Maps data as a resource for revealing urban road traffic flow patterns, this paper aims to provide a solution for a fairer and smarter congestion pricing method in cities. The case study area of the research contains three bordering districts of Budapest which are linked by one main road. The first district (5th) is the original downtown that is affected by the congestion charge plans of the city. The second district (13th) lies in the transition zone, and it has recently been transformed into a new CBD containing the biggest office zone in Budapest. The third district (4th) is a mainly residential type of area on the outskirts of the city. The raw data of the research was collected with the help of Google’s Distance Matrix API (Application Programming Interface) which provides future estimated traffic data via travel times between freely fixed coordinate pairs. From the difference of free flow and congested travel time data, the daily congestion patterns and hot spots are detectable in all measured roads within the area. The results suggest that the distribution of congestion peak times and hot spots are uneven in the examined area; however, there are frequently congested areas which lie outside the downtown and their inhabitants also need some protection. The conclusion of this case study is that cities can develop a real-time and place-based congestion charge system that forces car users to avoid frequently congested roads by changing their routes or travel modes. This would be a fairer solution for decreasing the negative environmental effects of the urban road transportation instead of protecting a very limited downtown area.

Keywords: Budapest, congestion charge, distance matrix API, application programming interface, pilot study

Procedia PDF Downloads 189
487 Localized Detection of ᴅ-Serine by Using an Enzymatic Amperometric Biosensor and Scanning Electrochemical Microscopy

Authors: David Polcari, Samuel C. Perry, Loredano Pollegioni, Matthias Geissler, Janine Mauzeroll

Abstract:

ᴅ-serine acts as an endogenous co-agonist for N-methyl-ᴅ-aspartate receptors in neuronal synapses. This makes it a key component in the development and function of a healthy brain, especially given its role in several neurodegenerative diseases such as Alzheimer’s disease and dementia. Despite such clear research motivations, the primary site and mechanism of ᴅ-serine release is still currently unclear. For this reason, we are developing a biosensor for the detection of ᴅ-serine utilizing a microelectrode in combination with a ᴅ-amino acid oxidase enzyme, which produces stoichiometric quantities of hydrogen peroxide in response to ᴅ-serine. For the fabrication of a biosensor with good selectivity, we use a permselective poly(meta-phenylenediamine) film to ensure only the target molecule is reacted, according to the size exclusion principle. In this work, we investigated the effect of the electrodeposition conditions used on the biosensor’s response time and selectivity. Careful optimization of the fabrication process allowed for enhanced biosensor response time. This allowed for the real time sensing of ᴅ-serine in a bulk solution, and also provided in means to map the efflux of ᴅ-serine in real time. This was done using scanning electrochemical microscopy (SECM) with the optimized biosensor to measure localized release of ᴅ-serine from an agar filled glass capillary sealed in an epoxy puck, which acted as a model system. The SECM area scan simultaneously provided information regarding the rate of ᴅ-serine flux from the model substrate, as well as the size of the substrate itself. This SECM methodology, which provides high spatial and temporal resolution, could be useful to investigate the primary site and mechanism of ᴅ-serine release in other biological samples.

Keywords: ᴅ-serine, enzymatic biosensor, microelectrode, scanning electrochemical microscopy

Procedia PDF Downloads 221
486 Cascaded Transcritical/Supercritical CO2 Cycles and Organic Rankine Cycles to Recover Low-Temperature Waste Heat and LNG Cold Energy Simultaneously

Authors: Haoshui Yu, Donghoi Kim, Truls Gundersen

Abstract:

Low-temperature waste heat is abundant in the process industries, and large amounts of Liquefied Natural Gas (LNG) cold energy are discarded without being recovered properly in LNG terminals. Power generation is an effective way to utilize low-temperature waste heat and LNG cold energy simultaneously. Organic Rankine Cycles (ORCs) and CO2 power cycles are promising technologies to convert low-temperature waste heat and LNG cold energy into electricity. If waste heat and LNG cold energy are utilized simultaneously in one system, the performance may outperform separate systems utilizing low-temperature waste heat and LNG cold energy, respectively. Low-temperature waste heat acts as the heat source and LNG regasification acts as the heat sink in the combined system. Due to the large temperature difference between the heat source and the heat sink, cascaded power cycle configurations are proposed in this paper. Cascaded power cycles can improve the energy efficiency of the system considerably. The cycle operating at a higher temperature to recover waste heat is called top cycle and the cycle operating at a lower temperature to utilize LNG cold energy is called bottom cycle in this study. The top cycle condensation heat is used as the heat source in the bottom cycle. The top cycle can be an ORC, transcritical CO2 (tCO2) cycle or supercritical CO2 (sCO2) cycle, while the bottom cycle only can be an ORC due to the low-temperature range of the bottom cycle. However, the thermodynamic path of the tCO2 cycle and sCO2 cycle are different from that of an ORC. The tCO2 cycle and the sCO2 cycle perform better than an ORC for sensible waste heat recovery due to a better temperature match with the waste heat source. Different combinations of the tCO2 cycle, sCO2 cycle and ORC are compared to screen the best configurations of the cascaded power cycles. The influence of the working fluid and the operating conditions are also investigated in this study. Each configuration is modeled and optimized in Aspen HYSYS. The results show that cascaded tCO2/ORC performs better compared with cascaded ORC/ORC and cascaded sCO2/ORC for the case study.

Keywords: LNG cold energy, low-temperature waste heat, organic Rankine cycle, supercritical CO₂ cycle, transcritical CO₂ cycle

Procedia PDF Downloads 249
485 Development of Surface Modification Technology for Control Element Drive Mechanism Nozzle and Fatigue Enhancement of Ni-Based Alloys

Authors: Auezhan Amanov, Inho Cho, Young-Sik Pyun

Abstract:

Control element drive mechanism (CEDM) nozzle is manufactured as welded on the reactor vessel and currently uses Alloy 690 material. The top of the reactor is equipped with about 100 CEDM nozzles with an internal diameter of about 70 mm. Relatively large Inlet/Outlet nozzles are equipped with two outlet nozzles and four inlet nozzles on the reactor wall. The inner diameter of the nozzle is vulnerable to stress corrosion cracking (SCC), and in order to solve this problem, an ultrasonic nanocrystal surface modification (UNSM) treatment is performed on the inner diameter of the nozzle and the weld surface. The ultimate goal is to improve the service life of parts by applying compressive residual stress and suppressing primary water stress corrosion cracking (PWSCC). The main purpose is to design and fabricate a UNSM treatment device for the internal diameter processing of CEDM nozzles and inlet/outlet nozzles. In order to develop the system, the basic technology such as the development of UNSM tooling is developed and the mechanical properties and fatigue performance of before and after UNSM treatment of reactor nozzle material made of Ni-based alloys using the specimen are compared and evaluated. The inner diameter of the nozzle was treated by a newly developed UNSM treatment under the optimized treatment parameters. It was found that the mechanical properties and fatigue performance of nozzle were improved in comparison with the untreated nozzle, which may be attributed to the increase in hardness, induced compressive residual stress.

Keywords: control element drive mechanism nozzle, fatigue, Ni-based alloy, ultrasonic nanocrystal surface modification, UNSM

Procedia PDF Downloads 106
484 Metabolic Profiling of Populus trichocarpa Family 1 UDP-Glycosyltransferases

Authors: Patricia M. B. Saint-Vincent, Anna Furches, Stephanie Galanie, Erica Teixeira Prates, Piet Jones, Nancy Engle, David Kainer, Wellington Muchero, Daniel Jacobson, Timothy J. Tschaplinski

Abstract:

Uridine diphosphate-glycosyltransferases (UGTs) are enzymes that catalyze sugar transfer to a variety of plant metabolites. UGT substrates, which include plant secondary metabolites involved in lignification, demonstrate new activities and incorporation when glycosylated. Knowledge of UGT function, substrate specificity, and enzyme products is important for plant engineering efforts, especially related to increasing plant biomass through lignification. UGTs in Populus trichocarpa, a biofuel feedstock, and model woody plant, were selected from a pool of gene candidates using rapid prioritization strategies. A functional genomics workflow, consisting of a metabolite genome-wide association study (mGWAS), expression of synthetic codon-optimized genes, and high-throughput biochemical assays with mass spectrometry-based analysis, was developed for determining the substrates and products of previously-uncharacterized enzymes. A total of 40 UGTs from P. trichocarpa were profiled, and the biochemical assay results were compared to predicted mGWAS connections. Assay results confirmed seven of 11 leaf mGWAS associations and demonstrated varying levels of substrate specificity among candidate UGTs. P. trichocarpa UGT substrate processing confirms the role of these newly-characterized enzymes in lignan, flavonoid, and phytohormone metabolism, with potential implications for cell wall biosynthesis, nitrogen uptake, and biotic and abiotic stress responses.

Keywords: Populus, metabolite-gene associations, GWAS, bio feedstocks, glycosyltransferase

Procedia PDF Downloads 106
483 Analysing the Interactive Effects of Factors Influencing Sand Production on Drawdown Time in High Viscosity Reservoirs

Authors: Gerald Gwamba, Bo Zhou, Yajun Song, Dong Changyin

Abstract:

The challenges that sand production presents to the oil and gas industry, particularly while working in poorly consolidated reservoirs, cannot be overstated. From restricting production to blocking production tubing, sand production increases the costs associated with production as it elevates the cost of servicing production equipment over time. Production in reservoirs that present with high viscosities, flow rate, cementation, clay content as well as fine sand contents is even more complex and challenging. As opposed to the one-factor at a-time testing, investigating the interactive effects arising from a combination of several factors offers increased reliability of results as well as representation of actual field conditions. It is thus paramount to investigate the conditions leading to the onset of sanding during production to ensure the future sustainability of hydrocarbon production operations under viscous conditions. We adopt the Design of Experiments (DOE) to analyse, using Taguchi factorial designs, the most significant interactive effects of sanding. We propose an optimized regression model to predict the drawdown time at sand production. The results obtained underscore that reservoirs characterized by varying (high and low) levels of viscosity, flow rate, cementation, clay, and fine sand content have a resulting impact on sand production. The only significant interactive effect recorded arises from the interaction between BD (fine sand content and flow rate), while the main effects included fluid viscosity and cementation, with percentage significances recorded as 31.3%, 37.76%, and 30.94%, respectively. The drawdown time model presented could be useful for predicting the time to reach the maximum drawdown pressure under viscous conditions during the onset of sand production.

Keywords: factorial designs, DOE optimization, sand production prediction, drawdown time, regression model

Procedia PDF Downloads 141
482 Qusai-Solid-State Electrochromic Device Based on PolyMethyl Methacrylate (PMMA)/Succinonitrile Gel Polymer Electrolyte

Authors: Jen-Yuan Wang, Min-Chuan Wang, Der-Jun Jan

Abstract:

Polymer electrolytes can be classified into four major categories, solid polymer electrolytes (SPEs), gel polymer electrolytes (GPEs), polyelectrolytes and composite polymer electrolytes. SPEs suffer from low ionic conductivity at room temperature. The main problems for GPEs are the poor thermal stability and mechanical properties. In this study, a GPE containing PMMA and succinonitrile is prepared to solve the problems mentioned above, and applied to the assembly of a quasi-solid-state electrochromic device (ECD). In the polymer electrolyte, poly(methyl methacrylate) (PMMA) is the polymer matrix and propylene carbonate (PC) is used as the plasticizer. To enhance the mechanical properties of this GPE, succinonitrile (SN) is introduced as the additive. For the electrochromic materials, tungsten oxide (WO3) is used as the cathodic coloring film, which is fabricated by pulsed dc magnetron reactive sputtering. For the anodic coloring material, Prussian blue nanoparticles (PBNPs) are synthesized and coated on the transparent Sn-doped indium oxide (ITO) glass. The thickness of ITO, WO3 and PB film is 110, 170 and 200 nm, respectively. The size of the ECD is 5×5 cm2. The effect of the introduction of SN into the GPEs is discussed by observing the electrochromic behaviors of the WO3-PB ECD. Besides, the composition ratio of PC to SN is also investigated by measuring the ionic conductivity. The optimized ratio of PC to SN is 4:1, and the ionic conductivity under this condition is 6.34x10-5 S∙cm-1, which is higher than that of PMMA/PC (1.35x10-6 S∙cm-1) and PMMA/EC/PC (4.52x10-6 S∙cm-1). This quasi-solid-state ECD fabricated with the PMMA/SN based GPE shows an optical contrast of ca. 53% at 690 nm. The optical transmittance of the ECD can be reversibly modulated from 72% (bleached) to 19% (darkened), by applying potentials of 1.5 and -2.2 V, respectively. During the durability test, the optical contrast of this ECD remains 44.5% after 2400 cycles, which is 83% of the original one.

Keywords: electrochromism, tungsten oxide, prussian blue, poly(methyl methacrylate), succinonitrile

Procedia PDF Downloads 286
481 Passive Vibration Isolation Analysis and Optimization for Mechanical Systems

Authors: Ozan Yavuz Baytemir, Ender Cigeroglu, Gokhan Osman Ozgen

Abstract:

Vibration is an important issue in the design of various components of aerospace, marine and vehicular applications. In order not to lose the components’ function and operational performance, vibration isolation design involving the optimum isolator properties selection and isolator positioning processes appear to be a critical study. Knowing the growing need for the vibration isolation system design, this paper aims to present two types of software capable of implementing modal analysis, response analysis for both random and harmonic types of excitations, static deflection analysis, Monte Carlo simulations in addition to study of parameter and location optimization for different types of isolation problem scenarios. Investigating the literature, there is no such study developing a software-based tool that is capable of implementing all those analysis, simulation and optimization studies in one platform simultaneously. In this paper, the theoretical system model is generated for a 6-DOF rigid body. The vibration isolation system of any mechanical structure is able to be optimized using hybrid method involving both global search and gradient-based methods. Defining the optimization design variables, different types of optimization scenarios are listed in detail. Being aware of the need for a user friendly vibration isolation problem solver, two types of graphical user interfaces (GUIs) are prepared and verified using a commercial finite element analysis program, Ansys Workbench 14.0. Using the analysis and optimization capabilities of those GUIs, a real application used in an air-platform is also presented as a case study at the end of the paper.

Keywords: hybrid optimization, Monte Carlo simulation, multi-degree-of-freedom system, parameter optimization, location optimization, passive vibration isolation analysis

Procedia PDF Downloads 555
480 Exploring the Role of Hydrogen to Achieve the Italian Decarbonization Targets using an OpenScience Energy System Optimization Model

Authors: Alessandro Balbo, Gianvito Colucci, Matteo Nicoli, Laura Savoldi

Abstract:

Hydrogen is expected to become an undisputed player in the ecological transition throughout the next decades. The decarbonization potential offered by this energy vector provides various opportunities for the so-called “hard-to-abate” sectors, including industrial production of iron and steel, glass, refineries and the heavy-duty transport. In this regard, Italy, in the framework of decarbonization plans for the whole European Union, has been considering a wider use of hydrogen to provide an alternative to fossil fuels in hard-to-abate sectors. This work aims to assess and compare different options concerning the pathway to be followed in the development of the future Italian energy system in order to meet decarbonization targets as established by the Paris Agreement and by the European Green Deal, and to infer a techno-economic analysis of the required asset alternatives to be used in that perspective. To accomplish this objective, the Energy System Optimization Model TEMOA-Italy is used, based on the open-source platform TEMOA and developed at PoliTo as a tool to be used for technology assessment and energy scenario analysis. The adopted assessment strategy includes two different scenarios to be compared with a business-as-usual one, which considers the application of current policies in a time horizon up to 2050. The studied scenarios are based on the up-to-date hydrogen-related targets and planned investments included in the National Hydrogen Strategy and in the Italian National Recovery and Resilience Plan, with the purpose of providing a critical assessment of what they propose. One scenario imposes decarbonization objectives for the years 2030, 2040 and 2050, without any other specific target. The second one (inspired to the national objectives on the development of the sector) promotes the deployment of the hydrogen value-chain. These scenarios provide feedback about the applications hydrogen could have in the Italian energy system, including transport, industry and synfuels production. Furthermore, the decarbonization scenario where hydrogen production is not imposed, will make use of this energy vector as well, showing the necessity of its exploitation in order to meet pledged targets by 2050. The distance of the planned policies from the optimal conditions for the achievement of Italian objectives is be clarified, revealing possible improvements of various steps of the decarbonization pathway, which seems to have as a fundamental element Carbon Capture and Utilization technologies for its accomplishment. In line with the European Commission open science guidelines, the transparency and the robustness of the presented results is ensured by the adoption of the open-source open-data model such as the TEMOA-Italy.

Keywords: decarbonization, energy system optimization models, hydrogen, open-source modeling, TEMOA

Procedia PDF Downloads 63
479 Grey Relational Analysis Coupled with Taguchi Method for Process Parameter Optimization of Friction Stir Welding on 6061 AA

Authors: Eyob Messele Sefene, Atinkut Atinafu Yilma

Abstract:

The highest strength-to-weight ratio criterion has fascinated increasing curiosity in virtually all areas where weight reduction is indispensable. One of the recent advances in manufacturing to achieve this intention endears friction stir welding (FSW). The process is widely used for joining similar and dissimilar non-ferrous materials. In FSW, the mechanical properties of the weld joints are impelled by property-selected process parameters. This paper presents verdicts of optimum process parameters in attempting to attain enhanced mechanical properties of the weld joint. The experiment was conducted on a 5 mm 6061 aluminum alloy sheet. A butt joint configuration was employed. Process parameters, rotational speed, traverse speed or feed rate, axial force, dwell time, tool material and tool profiles were utilized. Process parameters were also optimized, making use of a mixed L18 orthogonal array and the Grey relation analysis method with larger is better quality characteristics. The mechanical properties of the weld joint are examined through the tensile test, hardness test and liquid penetrant test at ambient temperature. ANOVA was conducted in order to investigate the significant process parameters. This research shows that dwell time, rotational speed, tool shape, and traverse speed have become significant, with a joint efficiency of about 82.58%. Nine confirmatory tests are conducted, and the results indicate that the average values of the grey relational grade fall within the 99% confidence interval. Hence the experiment is proven reliable.

Keywords: friction stir welding, optimization, 6061 AA, Taguchi

Procedia PDF Downloads 84
478 Hyper Parameter Optimization of Deep Convolutional Neural Networks for Pavement Distress Classification

Authors: Oumaima Khlifati, Khadija Baba

Abstract:

Pavement distress is the main factor responsible for the deterioration of road structure durability, damage vehicles, and driver comfort. Transportation agencies spend a high proportion of their funds on pavement monitoring and maintenance. The auscultation of pavement distress was based on the manual survey, which was extremely time consuming, labor intensive, and required domain expertise. Therefore, the automatic distress detection is needed to reduce the cost of manual inspection and avoid more serious damage by implementing the appropriate remediation actions at the right time. Inspired by recent deep learning applications, this paper proposes an algorithm for automatic road distress detection and classification using on the Deep Convolutional Neural Network (DCNN). In this study, the types of pavement distress are classified as transverse or longitudinal cracking, alligator, pothole, and intact pavement. The dataset used in this work is composed of public asphalt pavement images. In order to learn the structure of the different type of distress, the DCNN models are trained and tested as a multi-label classification task. In addition, to get the highest accuracy for our model, we adjust the structural optimization hyper parameters such as the number of convolutions and max pooling, filers, size of filters, loss functions, activation functions, and optimizer and fine-tuning hyper parameters that conclude batch size and learning rate. The optimization of the model is executed by checking all feasible combinations and selecting the best performing one. The model, after being optimized, performance metrics is calculated, which describe the training and validation accuracies, precision, recall, and F1 score.

Keywords: distress pavement, hyperparameters, automatic classification, deep learning

Procedia PDF Downloads 77
477 Optimization of Ultrasound Assisted Extraction of Polysaccharides from Plant Waste Materials: Selected Model Material is Hazelnut Skin

Authors: T. Yılmaz, Ş. Tavman

Abstract:

In this study, optimization of ultrasound assisted extraction (UAE) of hemicellulose based polysaccharides from plant waste material has been studied. Selected material is hazelnut skin. Extraction variables for the operation are extraction time, amplitude and application temperature. Optimum conditions have been evaluated depending on responses such as amount of wet crude polysaccharide, total carbohydrate content and dried sample. Pretreated hazelnut skin powders were used for the experiments. 10 grams of samples were suspended in 100 ml water in a jacketed vessel with additional magnetic stirring. Mixture was sonicated by immersing ultrasonic probe processor. After the extraction procedures, ethanol soluble and insoluble sides were separated for further examinations. The obtained experimental data were analyzed by analysis of variance (ANOVA). Second order polynomial models were developed using multiple regression analysis. The individual and interactive effects of applied variables were evaluated by Box Behnken Design. The models developed from the experimental design were predictive and good fit with the experimental data with high correlation coefficient value (R2 more than 0.95). Extracted polysaccharides from hazelnut skin are assumed to be pectic polysaccharides according to the literature survey of Fourier Transform Spectrometry (FTIR) analysis results. No more change can be observed between spectrums of different sonication times. Application of UAE at optimized condition has an important effect on extraction of hemicellulose from plant material by satisfying partial hydrolysis to break the bounds with other components in plant cell wall material. This effect can be summarized by varied intensity of microjets and microstreaming at varied sonication conditions.

Keywords: hazelnut skin, optimization, polysaccharide, ultrasound assisted extraction

Procedia PDF Downloads 323
476 Development of Immuno-Modulators: Application of Molecular Dynamics Simulation

Authors: Ruqaiya Khalil, Saman Usmani, Zaheer Ul-Haq

Abstract:

The accurate characterization of ligand binding affinity is indispensable for designing molecules with optimized binding affinity. Computational tools help in many directions to predict quantitative correlations between protein-ligand structure and their binding affinities. Molecular dynamics (MD) simulation is a modern state-of-the-art technique to evaluate the underlying basis of ligand-protein interactions by characterizing dynamic and energetic properties during the event. Autoimmune diseases arise from an abnormal immune response of the body against own tissues. The current regimen for the described condition is limited to immune-modulators having compromised pharmacodynamics and pharmacokinetics profiles. One of the key player mediating immunity and tolerance, thus invoking autoimmunity is Interleukin-2; a cytokine influencing the growth of T cells. Molecular dynamics simulation techniques are applied to seek insight into the inhibitory mechanisms of newly synthesized compounds that manifested immunosuppressant potentials during in silico pipeline. In addition to estimation of free energies associated with ligand binding, MD simulation yielded us a great deal of information about ligand-macromolecule interactions to evaluate the pattern of interactions and the molecular basis of inhibition. The present study is a continuum of our efforts to identify interleukin-2 inhibitors of both natural and synthetic origin. Herein, we report molecular dynamics simulation studies of Interluekin-2 complexed with different antagonists previously reported by our group. The study of protein-ligand dynamics enabled us to gain a better understanding of the contribution of different active site residues in ligand binding. The results of the study will be used as the guide to rationalize the fragment based synthesis of drug-like interleukin-2 inhibitors as immune-modulators.

Keywords: immuno-modulators, MD simulation, protein-ligand interaction, structure-based drug design

Procedia PDF Downloads 249
475 SME Internationalisation and Its Financing: An Exploratory Study That Analyses Government Support and Funding Mechanisms for Irish and Scottish International SMEs

Authors: L. Spencer, S. O’ Donohoe

Abstract:

Much of the research to date on internationalisation relates to large firms with much less known about how small and medium-sized enterprises (SMEs) engage in internationalisation. Given the crucial role of SMEs in contributing to economic growth, there is now an emphasis on the need for SMEs internationalise. Yet little is known about how SMEs undertake and finance such expansion and whether or not internationalisation actually hinders or helps them in securing finance. The purpose of this research is to explore the internationalisation process for SMEs, the sources of funding used in financing this expansion and support received from the state agencies in assisting their overseas expansion. A conceptual framework has been devised which marries the two strands of literature together (internationalisation and financing the firm). The exploratory nature of this research dictates that the most appropriate methodology was to use semi-structured interviews with SME owners; bank representatives and support agencies. In essence, a triangulated approach to the research problem facilitates assessment of the perceptions and experiences from firms, the state and the financial institutions. Our sample is drawn from SMEs operating in Ireland and Scotland, two small but very open economies where SMEs are the dominant form of organisation. The sample includes a range of industry sectors. Key findings to date suggest some SMEs are born global; others are born again global whilst a significant cohort can be classed as traditional internationalisers. Unsurprisingly there is a strong industry effect with firms in the high tech sector more likely to be faster internationalisers in contrast to those in the traditional manufacturing sectors. Owner manager’s own funds are deemed key to financing initial internationalisation lending support for the financial growth life cycle model albeit more important for the faster internationalisers in contrast to the slower cohort who are more likely to deploy external sources especially bank finance. Retained earnings remain the predominant source of on-going financing for internationalising firms but trade credit is often used and invoice discounting is utilised quite frequently. In terms of lending, asset based lending backed by personal guarantees appears paramount for securing bank finance. Whilst the lack of diversified sources of funding for internationalising SMEs was found in both jurisdictions there appears no evidence to suggest that internationalisation impedes firms in securing finance. Finally state supports were cited as important to the internationalisation process, in particular those provided by Enterprise Ireland were deemed very valuable. Considering the paucity of studies to date on SME internationalisation and in particular the funding mechanisms deployed by them; this study seeks to contribute to the body of knowledge in both the international business and finance disciplines.

Keywords: funding, government support, international pathways, modes of entry

Procedia PDF Downloads 237
474 A Structure-Based Approach for Adaptable Building System

Authors: Alireza Taghdiri, Sara Ghanbarzade Ghomi

Abstract:

Existing buildings are permanently subjected to change, continuously renovated and repaired in their long service life. Old buildings are destroyed and their material and components are recycled or reused for constructing new ones. In this process, importance of sustainability principles for building construction is obviously known and great significance must be attached to consumption of resources, resulting effects on the environment and economic costs. Utilization strategies for extending buildings service life and delay in destroying have positive effect on environment protection. In addition, simpler alterability or expandability of buildings’ structures and reducing energy and natural resources consumption have benefits for users, producers and environment. To solve these problems, by applying theories of open building, structural components of some conventional building systems have been analyzed and then, a new geometry adaptive building system is developed which can transform and support different imposed loads. In order to achieve this goal, various research methods and tools such as professional and scientific literatures review, comparative analysis, case study and computer simulation were applied and data interpretation was implemented using descriptive statistics and logical arguments. Therefore, hypothesis and proposed strategies were evaluated and an adaptable and reusable 2-dimensional building system was presented which can respond appropriately to dwellers and end-users needs and provide reusability of structural components of building system in new construction or function. Investigations showed that this incremental building system can be successfully applied in achieving the architectural design objectives and by small modifications on components and joints, it is easy to obtain different and adaptable load-optimized component alternatives for flexible spaces.

Keywords: adaptability, durability, open building, service life, structural building system

Procedia PDF Downloads 568
473 Digital Twins in the Built Environment: A Systematic Literature Review

Authors: Bagireanu Astrid, Bros-Williamson Julio, Duncheva Mila, Currie John

Abstract:

Digital Twins (DT) are an innovative concept of cyber-physical integration of data between an asset and its virtual replica. They have originated in established industries such as manufacturing and aviation and have garnered increasing attention as a potentially transformative technology within the built environment. With the potential to support decision-making, real-time simulations, forecasting abilities and managing operations, DT do not fall under a singular scope. This makes defining and leveraging the potential uses of DT a potential missed opportunity. Despite its recognised potential in established industries, literature on DT in the built environment remains limited. Inadequate attention has been given to the implementation of DT in construction projects, as opposed to its operational stage applications. Additionally, the absence of a standardised definition has resulted in inconsistent interpretations of DT in both industry and academia. There is a need to consolidate research to foster a unified understanding of the DT. Such consolidation is indispensable to ensure that future research is undertaken with a solid foundation. This paper aims to present a comprehensive systematic literature review on the role of DT in the built environment. To accomplish this objective, a review and thematic analysis was conducted, encompassing relevant papers from the last five years. The identified papers are categorised based on their specific areas of focus, and the content of these papers was translated into a through classification of DT. In characterising DT and the associated data processes identified, this systematic literature review has identified 6 DT opportunities specifically relevant to the built environment: Facilitating collaborative procurement methods, Supporting net-zero and decarbonization goals, Supporting Modern Methods of Construction (MMC) and off-site manufacturing (OSM), Providing increased transparency and stakeholders collaboration, Supporting complex decision making (real-time simulations and forecasting abilities) and Seamless integration with Internet of Things (IoT), data analytics and other DT. Finally, a discussion of each area of research is provided. A table of definitions of DT across the reviewed literature is provided, seeking to delineate the current state of DT implementation in the built environment context. Gaps in knowledge are identified, as well as research challenges and opportunities for further advancements in the implementation of DT within the built environment. This paper critically assesses the existing literature to identify the potential of DT applications, aiming to harness the transformative capabilities of data in the built environment. By fostering a unified comprehension of DT, this paper contributes to advancing the effective adoption and utilisation of this technology, accelerating progress towards the realisation of smart cities, decarbonisation, and other envisioned roles for DT in the construction domain.

Keywords: built environment, design, digital twins, literature review

Procedia PDF Downloads 67
472 Technological Development of a Biostimulant Bioproduct for Fruit Seedlings: An Engineering Overview

Authors: Andres Diaz Garcia

Abstract:

The successful technological development of any bioproduct, including those of the biostimulant type, requires to adequately completion of a series of stages allied to different disciplines that are related to microbiological, engineering, pharmaceutical chemistry, legal and market components, among others. Engineering as a discipline has a key contribution in different aspects of fermentation processes such as the design and optimization of culture media, the standardization of operating conditions within the bioreactor and the scaling of the production process of the active ingredient that it will be used in unit operations downstream. However, all aspects mentioned must take into account many biological factors of the microorganism such as the growth rate, the level of assimilation to various organic and inorganic sources and the mechanisms of action associated with its biological activity. This paper focuses on the practical experience within the Colombian Corporation for Agricultural Research (Agrosavia), which led to the development of a biostimulant bioproduct based on native rhizobacteria Bacillus amyloliquefaciens, oriented mainly to plant growth promotion in cape gooseberry nurseries and fruit crops in Colombia, and the challenges that were overcome from the expertise in the area of engineering. Through the application of strategies and engineering tools, a culture medium was optimized to obtain concentrations higher than 1E09 CFU (colony form units)/ml in liquid fermentation, the process of biomass production was standardized and a scale-up strategy was generated based on geometric (H/D of bioreactor relationships), and operational criteria based on a minimum dissolved oxygen concentration and that took into account the differences in the capacity of control of the process in the laboratory and pilot scales. Currently, the bioproduct obtained through this technological process is in stages of registration in Colombia for cape gooseberry fruits for export.

Keywords: biochemical engineering, liquid fermentation, plant growth promoting, scale-up process

Procedia PDF Downloads 107
471 Combustion Characteristics of Wet Woody Biomass in a Grate Furnace: Including Measurements within the Bed

Authors: Narges Razmjoo, Hamid Sefidari, Michael Strand

Abstract:

Biomass combustion is a growing technique for heat and power production due to the increasing stringent regulations with CO2 emissions. Grate-fired systems have been regarded as a common and popular combustion technology for burning woody biomass. However, some grate furnaces are not well optimized and may emit significant amount of unwanted compounds such as dust, NOx, CO, and unburned gaseous components. The combustion characteristics inside the fuel bed are of practical interest, as they are directly related to the release of volatiles and affect the stability and the efficiency of the fuel bed combustion. Although numerous studies have been presented on the grate firing of biomass, to the author’s knowledge, none of them have conducted a detailed experimental study within the fuel bed. It is difficult to conduct measurements of temperature and gas species inside the burning bed of the fuel in full-scale boilers. Results from such inside bed measurements can also be applied by the numerical experts for modeling the fuel bed combustion. The current work presents an experimental investigation into the combustion behavior of wet woody biomass (53 %) in a 4 MW reciprocating grate boiler, by focusing on the gas species distribution along the height of the fuel bed. The local concentrations of gases (CO, CO2, CH4, NO, and O2) inside the fuel bed were measured through a glass port situated on the side wall of the furnace. The measurements were carried out at five different heights of the fuel bed, by means of a bent stainless steel probe containing a type-k thermocouple. The sample gas extracted from the fuel bed, through the probe, was filtered and dried and then was analyzed using two infrared spectrometers. Temperatures of about 200-1100 °C were measured close to the grate, indicating that char combustion is occurring at the bottom of the fuel bed and propagates upward. The CO and CO2 concentration varied in the range of 15-35 vol % and 3-16 vol %, respectively, and NO concentration varied between 10-140 ppm. The profile of the gas concentrations distribution along the bed height provided a good overview of the combustion sub-processes in the fuel bed.

Keywords: experimental, fuel bed, grate firing, wood combustion

Procedia PDF Downloads 317
470 Ability of Gastric Enzyme Extract of Adult Camel to Clot Bovine Milk

Authors: Boudjenah-Haroun Saliha, Isselnane Souad, Nouani Abdelwahab, Baaissa Babelhadj, Mati Abderrahmane

Abstract:

Algeria is experiencing significant development of the dairy sector, where consumption of milk and milk products increased by 2.7 million tons in 2008 to 4,400,000 tons in 2013, and cheese production has reached 1640 tons in the year 2014 with average consumption of 0.7 kg/person/year. Although rennet is still the most used coagulating enzyme in cheese, its production has been growing worldwide shortage. This shortage is primarily due to a growing increase in the production and consumption of cheese, and the inability to increase in parallel the production of rennet. This shortage has caused very large fluctuations in its price). To overcome these obstacles, much research has been undertaken to find effective and competitive substitutes used industrially. For this, the selection of a local production of rennet substitute is desirable. It would allow a permanent supply with limited dependence on imports and price fluctuations. Investigations conducted by our research team showed that extracts coagulants from the stomachs of older camels are characterized by a coagulating power than those from younger camels. The objective of this work is to study the possibility of substituting commercial rennet coagulant by gastric enzymes from adult camels for coagulation bovine milk. Excerpts from the raw camel coagulants obtained are characterized through their teneures proteins and clotting and proteolytic activities. Milk clotting conditions by the action of these extracts were optimized. Milk clotting time all treated with enzyme preparations and under different conditions was calculated. Bovine rennet has been used for comparison. The results show that crude extracts from gastric adult camel can be good substituting bovine rennet.

Keywords: Algeria, camel, cheese, coagulation, gastric extracts, milk

Procedia PDF Downloads 426
469 Data Centers’ Temperature Profile Simulation Optimized by Finite Elements and Discretization Methods

Authors: José Alberto García Fernández, Zhimin Du, Xinqiao Jin

Abstract:

Nowadays, data center industry faces strong challenges for increasing the speed and data processing capacities while at the same time is trying to keep their devices a suitable working temperature without penalizing that capacity. Consequently, the cooling systems of this kind of facilities use a large amount of energy to dissipate the heat generated inside the servers, and developing new cooling techniques or perfecting those already existing would be a great advance in this type of industry. The installation of a temperature sensor matrix distributed in the structure of each server would provide the necessary information for collecting the required data for obtaining a temperature profile instantly inside them. However, the number of temperature probes required to obtain the temperature profiles with sufficient accuracy is very high and expensive. Therefore, other less intrusive techniques are employed where each point that characterizes the server temperature profile is obtained by solving differential equations through simulation methods, simplifying data collection techniques but increasing the time to obtain results. In order to reduce these calculation times, complicated and slow computational fluid dynamics simulations are replaced by simpler and faster finite element method simulations which solve the Burgers‘ equations by backward, forward and central discretization techniques after simplifying the energy and enthalpy conservation differential equations. The discretization methods employed for solving the first and second order derivatives of the obtained Burgers‘ equation after these simplifications are the key for obtaining results with greater or lesser accuracy regardless of the characteristic truncation error.

Keywords: Burgers' equations, CFD simulation, data center, discretization methods, FEM simulation, temperature profile

Procedia PDF Downloads 157
468 Redefining Success Beyond Borders: A Deep Dive into Effective Methods to Boost Morale Among Virtual Workers for Exponential Project Performance

Authors: Florence Ibeh, David Oyewmi Oyekunle, David Boohene

Abstract:

The continuous advancement of information technology has completely transformed how businesses and organizations operate on a global scale. The widespread availability of virtual communication tools enables individuals to opt for remote work. While remote employment offers various benefits, such as facilitating corporate growth and enhancing customer support, it also presents distinct challenges. Therefore, investigating the intricacies of virtual team morale is crucial for ensuring the achievement of project objectives. For this study, content analysis of pre-existing secondary data was employed to examine the phenomenon. Essential elements vital for improving the success of projects within virtual teams were identified. These factors include technology adoption, creating a distraction-free work environment, effective leadership, trust-building, clear communication channels, well-defined task allocation, active team participation, and motivation. Furthermore, the study established a substantial correlation between morale levels and the participation and productivity of virtual team members. Higher levels of morale were associated with optimal performance among virtual teams. The study determined that the key factors for enhancing project performance in virtual teams are the adoption of technology, a focused environment, effective leadership, trust, communication, well-defined tasks, collaborative teamwork, and motivation. Additionally, the study discovered that modifying the optimal strategies employed by in-office teams can enhance the diminished morale prevalent in remote teams to sustain a high level of team morale for virtual teams. The findings of this study are highly significant in the dynamic field of project management. Currently, there is limited information regarding strategies that address challenges arising from external factors in virtual teams, such as ambient noise and disruptions caused by family members. The findings underscore the significance of selecting appropriate communication technologies, delineating distinct roles and responsibilities for virtual team members, and nurturing a culture of accountability and trust. Promoting seamless collaboration and instilling motivation among virtual team members are deemed highly effective in augmenting employee engagement and performance within virtual team setting.

Keywords: virtual teams, morale, project performance, distract-free environment, technology adaptation

Procedia PDF Downloads 80
467 Predicting the Impact of Scope Changes on Project Cost and Schedule Using Machine Learning Techniques

Authors: Soheila Sadeghi

Abstract:

In the dynamic landscape of project management, scope changes are an inevitable reality that can significantly impact project performance. These changes, whether initiated by stakeholders, external factors, or internal project dynamics, can lead to cost overruns and schedule delays. Accurately predicting the consequences of these changes is crucial for effective project control and informed decision-making. This study aims to develop predictive models to estimate the impact of scope changes on project cost and schedule using machine learning techniques. The research utilizes a comprehensive dataset containing detailed information on project tasks, including the Work Breakdown Structure (WBS), task type, productivity rate, estimated cost, actual cost, duration, task dependencies, scope change magnitude, and scope change timing. Multiple machine learning models are developed and evaluated to predict the impact of scope changes on project cost and schedule. These models include Linear Regression, Decision Tree, Ridge Regression, Random Forest, Gradient Boosting, and XGBoost. The dataset is split into training and testing sets, and the models are trained using the preprocessed data. Cross-validation techniques are employed to assess the robustness and generalization ability of the models. The performance of the models is evaluated using metrics such as Mean Squared Error (MSE) and R-squared. Residual plots are generated to assess the goodness of fit and identify any patterns or outliers. Hyperparameter tuning is performed to optimize the XGBoost model and improve its predictive accuracy. The feature importance analysis reveals the relative significance of different project attributes in predicting the impact on cost and schedule. Key factors such as productivity rate, scope change magnitude, task dependencies, estimated cost, actual cost, duration, and specific WBS elements are identified as influential predictors. The study highlights the importance of considering both cost and schedule implications when managing scope changes. The developed predictive models provide project managers with a data-driven tool to proactively assess the potential impact of scope changes on project cost and schedule. By leveraging these insights, project managers can make informed decisions, optimize resource allocation, and develop effective mitigation strategies. The findings of this research contribute to improved project planning, risk management, and overall project success.

Keywords: cost impact, machine learning, predictive modeling, schedule impact, scope changes

Procedia PDF Downloads 24
466 Adaptability of Steel-Framed Industrialized Building System

Authors: Alireza Taghdiri, Sara Ghanbarzade Ghomi

Abstract:

Existing buildings are permanently subjected to change, continuously renovated and repaired in their long service life. Old buildings are destroyed and their material and components are recycled or reused for constructing new ones. In this process, importance of sustainability principles for building construction is obviously known and great significance must be attached to consumption of resources, resulting effects on the environment and economic costs. Utilization strategies for extending buildings service life and delay in destroying have positive effect on environment protection. In addition, simpler alterability or expandability of buildings’ structures and reducing energy and natural resources consumption have benefits for users, producers and environment. To solve these problems, by applying theories of open building, structural components of some conventional building systems have been analyzed and then, a new geometry adaptive building system is developed which can transform and support different imposed loads. In order to achieve this goal, various research methods and tools such as professional and scientific literatures review, comparative analysis, case study and computer simulation were applied and data interpretation was implemented using descriptive statistics and logical arguments. Therefore, hypothesis and proposed strategies were evaluated and an adaptable and reusable 2-dimensional building system was presented which can respond appropriately to dwellers and end-users needs and provide reusability of structural components of building system in new construction or function. Investigations showed that this incremental building system can be successfully applied in achieving the architectural design objectives and by small modifications on components and joints, it is easy to obtain different and adaptable load-optimized component alternatives for flexible spaces.

Keywords: adaptability, durability, open building, service life, structural building system

Procedia PDF Downloads 352
465 An Empirical Analysis on the Evolution Characteristics and Textual Content of Campus Football Policy in China

Authors: Shangjun Zou, Zhiyuan Wang, Songhui You

Abstract:

Introduction In recent years, the Chinese government has issued several policies to promote the institutional reform and innovation of the development of campus football, but many problems have been exposed in the process of policy implementation. Therefore, this paper attempts to conduct an empirical analysis of the campus football policy texts to reveal the dynamic development of the microsystem in the process of policy evolution. Methods The selected policy contents are coded by constructing a two-dimensional analysis framework of campus football policy tool-policy objective. Specifically, the X dimension consists of three oriented policy tools: environment, supply and demand, while the Y dimension is divided into six aspects of policy objectives, including institution, competition, player teaching, coach training, resource guarantee and popularization. And the distribution differences of textual analysis units on X and Y dimensions are tested by using SPSS22.0 so as to evaluate the characteristics and development trend of campus football policy on respective subjects. Results 1) In the policy evolution process of campus football stepping into the 2.0 Era, there were no significant differences in the frequency distribution of policy tools(p=0.582) and policy objectives(p=0.603). The collaborative governance of multiple participants has become the primary trend, and the guiding role of Chinese Football Association has gradually become prominent. 2) There were significant differences in the distribution of policy tools before the evolution at a 95% confidence level(p=0.041). With environmental tools always maintaining the dominant position, the overall synergy of policy tools increased slightly. 3) There were significant differences in the distribution of policy objectives after the evolution at a 90% confidence level(p=0.069). The competition system of policy objective has not received enough attention while the construction of institution and resource guarantee system has been strengthened. Conclusion The upgraded version of campus football should adhere to the education concept of health first, promote the coordinated development of youth cultural learning and football skills, and strive to achieve more solid popularization, more scientific institution, more comprehensive resource guarantee and adequate integration. At the same time, it is necessary to strengthen the collaborative allocation of policy tools and reasonable planning of policy objectives so as to promote the high quality and sustainable development of campus football in the New Era. Endnote The policy texts selected in this paper are “Implementation Opinions on Accelerating the Development of Youth Campus Football” and “Action Plans for the Construction of Eight Systems of National Youth Campus Football”, which were promulgated on August 13, 2015 and September 25, 2020 respectively.

Keywords: campus football, content analysis, evolution characteristics, policy objective, policy tool

Procedia PDF Downloads 180
464 Measurement of Influence of the COVID-19 Pandemic on Efficiency of Japan’s Railway Companies

Authors: Hideaki Endo, Mika Goto

Abstract:

The global outbreak of the COVID-19 pandemic has seriously affected railway businesses. The number of railway passengers decreased due to the decline in the number of commuters and business travelers to avoid crowded trains and a sharp drop in inbound tourists visiting Japan. This has affected not only railway businesses but also related businesses, including hotels, leisure businesses, and retail businesses at station buildings. In 2021, the companies were divided into profitable and loss-making companies. This division suggests that railway companies, particularly loss-making companies, needed to decrease operational inefficiency. To measure the impact of COVID-19 and discuss the sustainable management strategies of railway companies, we examine the cost inefficiency of Japanese listed railway companies by applying stochastic frontier analysis (SFA) to their operational and financial data. First, we employ the stochastic frontier cost function approach to measure inefficiency. The cost frontier function is formulated as a Cobb–Douglas type, and we estimated parameters and variables for inefficiency. This study uses panel data comprising 26 Japanese-listed railway companies from 2005 to 2020. This period includes several events deteriorating the business environment, such as the financial crisis from 2007 to 2008 and the Great East Japan Earthquake of 2011, and we compare those impacts with those of the COVID-19 pandemic after 2020. Second, we identify the characteristics of the best-practice railway companies and examine the drivers of cost inefficiencies. Third, we analyze the factors influencing cost inefficiency by comparing the profiles of the top 10 railway companies and others before and during the pandemic. Finally, we examine the relationship between cost inefficiency and the implementation of efficiency measures for each railway company. We obtained the following four findings. First, most Japanese railway companies showed the lowest cost inefficiency (most efficient) in 2014 and the highest in 2020 (least efficient) during the COVID-19 pandemic. The second worst occurred in 2009 when it was affected by the financial crisis. However, we did not observe a significant impact of the 2011 Great East Japan Earthquake. This is because no railway company was influenced by the earthquake in this operating area, except for JR-EAST. Second, the best-practice railway companies are KEIO and TOKYU. The main reason for their good performance is that both operate in and near the Tokyo metropolitan area, which is densely populated. Third, we found that non-best-practice companies had a larger decrease in passenger kilometers than best-practice companies. This indicates that passengers made fewer long-distance trips because they refrained from inter-prefectural travel during the pandemic. Finally, we found that companies that implement more efficiency improvement measures had higher cost efficiency and they effectively used their customer databases through proactive DX investments in marketing and asset management.

Keywords: COVID-19 pandemic, stochastic frontier analysis, railway sector, cost efficiency

Procedia PDF Downloads 63
463 Urban Waste Management for Health and Well-Being in Lagos, Nigeria

Authors: Bolawole F. Ogunbodede, Mokolade Johnson, Adetunji Adejumo

Abstract:

High population growth rate, reactive infrastructure provision, inability of physical planning to cope with developmental pace are responsible for waste water crisis in the Lagos Metropolis. Septic tank is still the most prevalent waste-water holding system. Unfortunately, there is a dearth of septage treatment infrastructure. Public waste-water treatment system statistics relative to the 23 million people in Lagos State is worrisome. 1.85 billion Cubic meters of wastewater is generated on daily basis and only 5% of the 26 million population is connected to public sewerage system. This is compounded by inadequate budgetary allocation and erratic power supply in the last two decades. This paper explored community participatory waste-water management alternative at Oworonshoki Municipality in Lagos. The study is underpinned by decentralized Waste-water Management systems in built-up areas. The initiative accommodates 5 step waste-water issue including generation, storage, collection, processing and disposal through participatory decision making in two Oworonshoki Community Development Association (CDA) areas. Drone assisted mapping highlighted building footage. Structured interviews and focused group discussion of land lord associations in the CDA areas provided collaborator platform for decision-making. Water stagnation in primary open drainage channels and natural retention ponds in framing wetlands is traceable to frequent of climate change induced tidal influences in recent decades. Rise in water table resulting in septic-tank leakage and water pollution is reported to be responsible for the increase in the water born infirmities documented in primary health centers. This is in addition to unhealthy dumping of solid wastes in the drainage channels. The effect of uncontrolled disposal system renders surface waters and underground water systems unsafe for human and recreational use; destroys biotic life; and poisons the fragile sand barrier-lagoon urban ecosystems. Cluster decentralized system was conceptualized to service 255 households. Stakeholders agreed on public-private partnership initiative for efficient wastewater service delivery.

Keywords: health, infrastructure, management, septage, well-being

Procedia PDF Downloads 162
462 How to Assess the Attractiveness of Business Location According to the Mainstream Concepts of Comparative Advantages

Authors: Philippe Gugler

Abstract:

Goal of the study: The concept of competitiveness has been addressed by economic theorists and policymakers for several hundreds of years, with both groups trying to understand the drivers of economic prosperity and social welfare. The goal of this contribution is to address the major useful theoretical contributions that permit to identify the main drivers of a territory’s competitiveness. We first present the major contributions found in the classical and neo-classical theories. Then, we concentrate on two majors schools providing significant thoughts on the competitiveness of locations: the Economic Geography (EG) School and the International Business (IB) School. Methodology: The study is based on a literature review of the classical and neo-classical theories, on the economic geography theories and on the international business theories. This literature review establishes links between these theoretical mainstreams. This work is based on the academic framework establishing a meaningful literature review aimed to respond to our research question and to develop further research in this field. Results: The classical and neo-classical pioneering theories provide initial insights that territories are different and that these differences explain the discrepancies in their levels of prosperity and standards of living. These theories emphasized different factors impacting the level and the growth of productivity in a given area and therefore the degree of their competitiveness. However, these theories are not sufficient to more precisely identify the drivers and enablers of location competitiveness and to explain, in particular, the factors that drive the creation of economic activities, the expansion of economic activities, the creation of new firms and the attraction of foreign firms. Prosperity is due to economic activities created by firms. Therefore, we need more theoretical insights to scrutinize the competitive advantages of territories or, in other words, their ability to offer the best conditions that enable economic agents to achieve higher rates of productivity in open markets. Two major theories provide, to a large extent, the needed insights: the economic geography theory and the international business theory. The economic geography studies scrutinized in this study from Marshall to Porter, aim to explain the drivers of the concentration of specific industries and activities in specific locations. These activity agglomerations may be due to the creation of new enterprises, the expansion of existing firms, and the attraction of firms located elsewhere. Regarding this last possibility, the international business (IB) theories focus on the comparative advantages of locations as far as multinational enterprises (MNEs) strategies are concerned. According to international business theory, the comparative advantages of a location serves firms not only by exploiting their ownership advantages (mostly as far as market seeking, resource seeking and efficiency seeking investments are concerned) but also by augmenting and/or creating new ownership advantages (strategic asset seeking investments). The impact of a location on the competitiveness of firms is considered from both sides: the MNE’s home country and the MNE’s host country.

Keywords: competitiveness, economic geography, international business, attractiveness of businesses

Procedia PDF Downloads 137
461 Technology Road Mapping in the Fourth Industrial Revolution: A Comprehensive Analysis and Strategic Framework

Authors: Abdul Rahman Hamdan

Abstract:

The Fourth Industrial Revolution (4IR) has brought unprecedented technological advancements that have disrupted many industries worldwide. In keeping up with the technological advances and rapid disruption by the introduction of many technological advancements brought forth by the 4IR, the use of technology road mapping has emerged as one of the critical tools for organizations to leverage. Technology road mapping can be used by many companies to guide them to become more adaptable and anticipate future transformation and innovation, and avoid being redundant or irrelevant due to the rapid changes in technological advancement. This research paper provides a comprehensive analysis of technology road mapping within the context of the 4IR. The objectives of the paper are to provide companies with practical insights and a strategic framework of technology road mapping for them to navigate the fast-changing nature of the 4IR. This study also contributes to the understanding and practice of technology road mapping in the 4IR and, at the same time, provides organizations with the necessary tools and critical insight to navigate the 4IR transformation by leveraging technology road mapping. Based on the literature review and case studies, the study analyses key principles, methodologies, and best practices in technology road mapping and integrates them with the unique characteristics and challenges of the 4IR. The research paper gives the background of the fourth industrial revolution. It explores the disruptive potential of technologies in the 4IR and the critical need for technology road mapping that consists of strategic planning and foresight to remain competitive and relevant in the 4IR era. It also highlights the importance of technology road mapping as an organisation’s proactive approach to align the organisation’s objectives and resources to their technology and product development in meeting the fast-evolving technological 4IR landscape. The paper also includes the theoretical foundations of technology road mapping and examines various methodological approaches, and identifies external stakeholders in the process, such as external experts, stakeholders, collaborative platforms, and cross-functional teams to ensure an integrated and robust technological roadmap for the organisation. Moreover, this study presents a comprehensive framework for technology road mapping in the 4IR by incorporating key elements and processes such as technology assessment, competitive intelligence, risk analysis, and resource allocation. It provides a framework for implementing technology road mapping from strategic planning, goal setting, and technology scanning to road mapping visualisation, implementation planning, monitoring, and evaluation. In addition, the study also addresses the challenges and limitations related to technology roadmapping in 4IR, including the gap analysis. In conclusion of the study, the study will propose a set of practical recommendations for organizations that intend to leverage technology road mapping as a strategic tool in the 4IR in driving innovation and becoming competitive in the current and future ecosystem.

Keywords: technology management, technology road mapping, technology transfer, technology planning

Procedia PDF Downloads 61
460 Mechanisms Underlying the Effects of School-Based Internet Intervention for Alcohol Drinking Behaviours among Chinese Adolescent

Authors: Keith T. S. Tung, Frederick K. Ho, Rosa S. Wong, Camilla K. M. Lo, Wilfred H. S. Wong, C. B. Chow, Patrick Ip

Abstract:

Objectives: Underage drinking is an important public health problem both locally and globally. Conventional prevention/intervention relies on unidirectional knowledge transfer such as mail leaflets or health talks which showed mixed results in changing the target behaviour. Previously, we conducted a school internet-based intervention which was found to be effective in reducing alcohol use among adolescents, yet the underlying mechanisms have not been properly investigated. This study, therefore, examined the mechanisms that explain how the intervention produced a change in alcohol drinking behaviours among Chinese adolescent as observed in our previous clustered randomised controlled trial (RCT) study. Methods: This is a cluster randomised controlled trial with parallel group design. Participating schools were randomised to the Internet intervention or the conventional health education group (control) with a 1:1 allocation ratio. Secondary 1–3 students of the participating schools were enrolled in this study. The Internet intervention was a web-based quiz game competition, in which participating students would answer 1,000 alcohol-related multiple-choice quiz questions. Conventional health education group received a promotional package on equivalent alcohol-related knowledge. The participants’ alcohol-related attitude, knowledge, and perceived behavioural control were self-reported before the intervention (baseline) and one month and three months after the intervention. Results: Our RCT results showed that participants in the Internet group were less likely to drink (risk ratio [RR] 0.79, p < 0.01) as well as in lesser amount (β -0.06, p < 0.05) compared to those in the control group at both post-intervention follow-ups. Within the intervention group, regression analyses showed that high quiz scorer had greater improvement in alcohol-related knowledge (β 0.28, p < 0.01) and attitude (β -0.26, p < 0.01) at 1 month after intervention, which in turn increased their perceived behavioural control against alcohol use (β 0.10 and -0.26, both p < 0.01). Attitude, compared to knowledge, was found to be a stronger contributor to the intervention effect on perceived behavioural control. Conclusions: Our internet-based intervention has demonstrated effectiveness in reducing the risk of underage drinking when compared with conventional health education. Our study results further showed an attitude to be a more important factor than knowledge in changing health-related behaviour. This has an important implication for future prevention/intervention on an underage drinking problem.

Keywords: adolescents, internet-based intervention, randomized controlled trial, underage drinking

Procedia PDF Downloads 145
459 Using Recycled Wastes (Glass Powder) as Partially Replacement for Cement

Authors: Passant Youssef, Ahmed El-Tair, Amr El-Nemr

Abstract:

Lately, with the environmental changes, enthusiasts trigger to stop the contamination of environment. Thus, various efforts were exerted for innovating environmental friendly concrete to sustain as a ‘Green Building’ material. Green building materials consider the cement industry as one of the most sources of air pollutant with high rate of carbon dioxide (CO₂) emissions. Several methods were developed to extensively reduce the influence of cement industry on environment. These methods such as using supplementary cementitious material or improving the cement manufacturing process are still under investigation. However, with the presence of recycled wastes from construction and finishing materials, the use of supplementary cementitious materials seems to provide an economic solution. Furthermore, it improves the mechanical properties of cement paste, in addition to; it modulates the workability and durability of concrete. In this paper, the glass powder was considered to be used as partial replacement of cement. This study provided the mechanical influence for using the glass powder as partial replacement of cement. In addition, it examines the microstructure of cement mortar using scanning electron microscope and X-ray diffraction. The cement in concrete is replaced by waste glass powder in steps of 5%, 10%, 15%, 20% and 25% by weight of cement and its effects on compressive and flexure strength were determined after 7 and 28 days. It was found that the 5% glass powder replacement increased the 7 days compressive strength by 20.5%, however, there was no increase in compressive strength after 28 days; which means that the glass powder did not react in the cement mortar due to its amorphous nature on the long run, and it can act as fine aggregate better that cement replacement. As well as, the 5% and 10% glass powder replacement increased the 28 days flexural strength by 46.9%. SEM micrographs showed very dense matrix for the optimum specimen compared to control specimen as well; some glass particles were clearly observed. High counts of silica were optimized from XRD while amorphous materials such as calcium silicate cannot be directly detected.

Keywords: supplementary materials, glass powder, concrete, cementitious materials

Procedia PDF Downloads 203