Search results for: wireless channel modelling
355 Green Procedure for Energy and Emission Balancing of Alternative Scenario Improvements for Cogeneration System: A Case of Hardwood Lumber Manufacturing Process
Authors: Aldona Kluczek
Abstract:
Energy efficient process have become a pressing research field in manufacturing. The arguments for having an effective industrial energy efficiency processes are interacted with factors: economic and environmental impact, and energy security. Improvements in energy efficiency are most often achieved by implementation of more efficient technology or manufacturing process. Current processes of electricity production represents the biggest consumption of energy and the greatest amount of emissions to the environment. The goal of this study is to improve the potential energy-savings and reduce greenhouse emissions related to improvement scenarios for the treatment of hardwood lumber produced by an industrial plant operating in the U.S. through the application of green balancing procedure, in order to find the preferable efficient technology. The green procedure for energy is based on analysis of energy efficiency data. Three alternative scenarios of the cogeneration systems plant (CHP) construction are considered: generation of fresh steam, the purchase of a new boiler with the operating pressure 300 pounds per square inch gauge (PSIG), an installation of a new boiler with a 600 PSIG pressure. In this paper, the application of a bottom-down modelling for energy flow to devise a streamlined Energy and Emission Flow Analyze method for the technology of producing electricity is illustrated. It will identify efficiency or technology of a given process to be reached, through the effective use of energy, or energy management. Results have shown that the third scenario seem to be the efficient alternative scenario considered from the environmental and economic concerns for treating hardwood lumber. The energy conservation evaluation options could save an estimated 6,215.78 MMBtu/yr in each year, which represents 9.5% of the total annual energy usage. The total annual potential cost savings from all recommendations is $143,523/yr, which represents 30.1% of the total annual energy costs. Estimation have presented that energy cost savings are possible up to 43% (US$ 143,337.85), representing 18.6% of the total annual energy costs.Keywords: alternative scenario improvements, cogeneration system, energy and emission flow analyze, energy balancing, green procedure, hardwood lumber manufacturing process
Procedia PDF Downloads 208354 Momentum in the Stock Exchange of Thailand
Authors: Mussa Hussaini, Supasith Chonglerttham
Abstract:
Stocks are usually classified according to their characteristics which are unique enough such that the performance of each category can be differentiated from another. The reasons behind such classifications in the financial market are sometimes financial innovation or it can also be because of finding a premium in a group of stocks with similar features. One of the major classifications in stocks market is called momentum strategy. Based on this strategy stocks are classified according to their past performances into past winners and past losers. Momentum in a stock market refers to the idea that stocks will keep moving in the same direction. In other word, stocks with rising prices (past winners stocks) will continue to rise and those stocks with falling prices (past losers stocks) will continue to fall. The performance of this classification has been well documented in numerous studies in different countries. These studies suggest that past winners tend to outperform past losers in the future. However, academic research in this direction has been limited in countries such as Thailand and to the best of our knowledge, there has been no such study in Thailand after the financial crisis of 1997. The significance of this study stems from the fact that Thailand is an open market and has been encouraging foreign investments as one of the means to enhance employment, promote economic development, and technology transfer and the main equity market in Thailand, the Stock Exchange of Thailand is a crucial channel for Foreign Investment inflow into the country. The equity market size in Thailand increased from $1.72 billion in 1984 to $133.66 billion in 1993, an increase of over 77 times within a decade. The main contribution of this paper is evidence for size category in the context of the equity market in Thailand. Almost all previous studies have focused solely on large stocks or indices. This paper extends the scope beyond large stocks and indices by including small and tiny stocks as well. Further, since there is a distinct absence of detailed academic research on momentum strategy in the Stock Exchange of Thailand after the crisis, this paper also contributes to the extension of existing literature of the study. This research is also of significance for those researchers who would like to compare the performance of this strategy in different countries and markets. In the Stock Exchange of Thailand, we examined the performance of momentum strategy from 2010 to 2014. Returns on portfolios are calculated on monthly basis. Our results on momentum strategy confirm that there is positive momentum profit in large size stocks whereas there is negative momentum profit in small size stocks during the period of 2010 to 2014. Furthermore, the equal weighted average of momentum profit of both small and large size category do not provide any indication of overall momentum profit.Keywords: momentum strategy, past loser, past winner, stock exchange of Thailand
Procedia PDF Downloads 317353 Cleaning of Polycyclic Aromatic Hydrocarbons (PAH) Obtained from Ferroalloys Plant
Authors: Stefan Andersson, Balram Panjwani, Bernd Wittgens, Jan Erik Olsen
Abstract:
Polycyclic Aromatic hydrocarbons are organic compounds consisting of only hydrogen and carbon aromatic rings. PAH are neutral, non-polar molecules that are produced due to incomplete combustion of organic matter. These compounds are carcinogenic and interact with biological nucleophiles to inhibit the normal metabolic functions of the cells. Norways, the most important sources of PAH pollution is considered to be aluminum plants, the metallurgical industry, offshore oil activity, transport, and wood burning. Stricter governmental regulations regarding emissions to the outer and internal environment combined with increased awareness of the potential health effects have motivated Norwegian metal industries to increase their efforts to reduce emissions considerably. One of the objective of the ongoing industry and Norwegian research council supported "SCORE" project is to reduce potential PAH emissions from an off gas stream of a ferroalloy furnace through controlled combustion. In a dedicated combustion chamber. The sizing and configuration of the combustion chamber depends on the combined properties of the bulk gas stream and the properties of the PAH itself. In order to achieve efficient and complete combustion the residence time and minimum temperature need to be optimized. For this design approach reliable kinetic data of the individual PAH-species and/or groups thereof are necessary. However, kinetic data on the combustion of PAH are difficult to obtain and there is only a limited number of studies. The paper presents an evaluation of the kinetic data for some of the PAH obtained from literature. In the present study, the oxidation is modelled for pure PAH and also for PAH mixed with process gas. Using a perfectly stirred reactor modelling approach the oxidation is modelled including advanced reaction kinetics to study influence of residence time and temperature on the conversion of PAH to CO2 and water. A Chemical Reactor Network (CRN) approach is developed to understand the oxidation of PAH inside the combustion chamber. Chemical reactor network modeling has been found to be a valuable tool in the evaluation of oxidation behavior of PAH under various conditions.Keywords: PAH, PSR, energy recovery, ferro alloy furnace
Procedia PDF Downloads 273352 Acoustic Energy Harvesting Using Polyvinylidene Fluoride (PVDF) and PVDF-ZnO Piezoelectric Polymer
Authors: S. M. Giripunje, Mohit Kumar
Abstract:
Acoustic energy that exists in our everyday life and environment have been overlooked as a green energy that can be extracted, generated, and consumed without any significant negative impact to the environment. The harvested energy can be used to enable new technology like wireless sensor networks. Technological developments in the realization of truly autonomous MEMS devices and energy storage systems have made acoustic energy harvesting (AEH) an increasingly viable technology. AEH is the process of converting high and continuous acoustic waves from the environment into electrical energy by using an acoustic transducer or resonator. AEH is not popular as other types of energy harvesting methods since sound waves have lower energy density and such energy can only be harvested in very noisy environment. However, the energy requirements for certain applications are also correspondingly low and also there is a necessity to observe the noise to reduce noise pollution. So the ability to reclaim acoustic energy and store it in a usable electrical form enables a novel means of supplying power to relatively low power devices. A quarter-wavelength straight-tube acoustic resonator as an acoustic energy harvester is introduced with polyvinylidene fluoride (PVDF) and PVDF doped with ZnO nanoparticles, piezoelectric cantilever beams placed inside the resonator. When the resonator is excited by an incident acoustic wave at its first acoustic eigen frequency, an amplified acoustic resonant standing wave is developed inside the resonator. The acoustic pressure gradient of the amplified standing wave then drives the vibration motion of the PVDF piezoelectric beams, generating electricity due to the direct piezoelectric effect. In order to maximize the amount of the harvested energy, each PVDF and PVDF-ZnO piezoelectric beam has been designed to have the same structural eigen frequency as the acoustic eigen frequency of the resonator. With a single PVDF beam placed inside the resonator, the harvested voltage and power become the maximum near the resonator tube open inlet where the largest acoustic pressure gradient vibrates the PVDF beam. As the beam is moved to the resonator tube closed end, the voltage and power gradually decrease due to the decreased acoustic pressure gradient. Multiple piezoelectric beams PVDF and PVDF-ZnO have been placed inside the resonator with two different configurations: the aligned and zigzag configurations. With the zigzag configuration which has the more open path for acoustic air particle motions, the significant increases in the harvested voltage and power have been observed. Due to the interruption of acoustic air particle motion caused by the beams, it is found that placing PVDF beams near the closed tube end is not beneficial. The total output voltage of the piezoelectric beams increases linearly as the incident sound pressure increases. This study therefore reveals that the proposed technique used to harvest sound wave energy has great potential of converting free energy into useful energy.Keywords: acoustic energy, acoustic resonator, energy harvester, eigenfrequency, polyvinylidene fluoride (PVDF)
Procedia PDF Downloads 385351 Nonlinear Modelling of Sloshing Waves and Solitary Waves in Shallow Basins
Authors: Mohammad R. Jalali, Mohammad M. Jalali
Abstract:
The earliest theories of sloshing waves and solitary waves based on potential theory idealisations and irrotational flow have been extended to be applicable to more realistic domains. To this end, the computational fluid dynamics (CFD) methods are widely used. Three-dimensional CFD methods such as Navier-Stokes solvers with volume of fluid treatment of the free surface and Navier-Stokes solvers with mappings of the free surface inherently impose high computational expense; therefore, considerable effort has gone into developing depth-averaged approaches. Examples of such approaches include Green–Naghdi (GN) equations. In Cartesian system, GN velocity profile depends on horizontal directions, x-direction and y-direction. The effect of vertical direction (z-direction) is also taken into consideration by applying weighting function in approximation. GN theory considers the effect of vertical acceleration and the consequent non-hydrostatic pressure. Moreover, in GN theory, the flow is rotational. The present study illustrates the application of GN equations to propagation of sloshing waves and solitary waves. For this purpose, GN equations solver is verified for the benchmark tests of Gaussian hump sloshing and solitary wave propagation in shallow basins. Analysis of the free surface sloshing of even harmonic components of an initial Gaussian hump demonstrates that the GN model gives predictions in satisfactory agreement with the linear analytical solutions. Discrepancies between the GN predictions and the linear analytical solutions arise from the effect of wave nonlinearities arising from the wave amplitude itself and wave-wave interactions. Numerically predicted solitary wave propagation indicates that the GN model produces simulations in good agreement with the analytical solution of the linearised wave theory. Comparison between the GN model numerical prediction and the result from perturbation analysis confirms that nonlinear interaction between solitary wave and a solid wall is satisfactorilly modelled. Moreover, solitary wave propagation at an angle to the x-axis and the interaction of solitary waves with each other are conducted to validate the developed model.Keywords: Green–Naghdi equations, nonlinearity, numerical prediction, sloshing waves, solitary waves
Procedia PDF Downloads 285350 Evaluation of Washing Performance of Household Wastewater Purified by Advanced Oxidation Process
Authors: Nazlı Çetindağ, Pelin Yılmaz Çetiner, Metin Mert İlgün, Emine Birci, Gizemnur Yıldız Uysal, Özcan Hatipoğlu, Ehsan Tuzcuoğlu, Gökhan Sır
Abstract:
Stressing the importance of water conservation, emphasizing the need for efficient management of household water, and underlining the significance of alternative solutions are important. In this context, advanced solutions based on technologies such as the advanced oxidation process have emerged as promising methods for treating household wastewater. Evaluating household water usage holds critical importance for the sustainability of water resources. Researchers and experts are examining various technological approaches to effectively treat and reclaim water for reuse. In this framework, the advanced oxidation process has proven to be an effective method for the removal of various organic and inorganic pollutants in the treatment of household wastewater. In this study, performance will be evaluated by comparing it with the reference case. This international criterion simulates the washing of home textile products, determining various performance parameters. The specially designed stain strips, including sebum, carbon black, blood, cocoa, and red wine, used in experiments, represent various household stains. These stain types were carefully selected to represent challenging stain scenarios, ensuring a realistic assessment of washing performance. Experiments conducted under different temperatures and program conditions successfully demonstrate the practical applicability of the advanced oxidation process for treating household wastewater. It is important to note that both adherence to standards and the use of real-life stain types contribute to the broad applicability of the findings. In conclusion, this study strongly supports the effectiveness of treating household wastewater with the advanced oxidation process in terms of washing performance under both standard and practical application conditions. The study underlines the importance of alternative solutions for sustainable water resource management and highlights the potential of the advanced oxidation process in the treatment of household water, contributing significantly to optimizing water usage and developing sustainable water management solutions.Keywords: advanced oxidation process, household water usage, household appliance waste water, modelling, water reuse
Procedia PDF Downloads 64349 Generalized Synchronization in Systems with a Complex Topology of Attractor
Authors: Olga I. Moskalenko, Vladislav A. Khanadeev, Anastasya D. Koloskova, Alexey A. Koronovskii, Anatoly A. Pivovarov
Abstract:
Generalized synchronization is one of the most intricate phenomena in nonlinear science. It can be observed both in systems with a unidirectional and mutual type of coupling including the complex networks. Such a phenomenon has a number of practical applications, for example, for the secure information transmission through the communication channel with a high level of noise. Known methods for the secure information transmission needs in the increase of the privacy of data transmission that arises a question about the observation of such phenomenon in systems with a complex topology of chaotic attractor possessing two or more positive Lyapunov exponents. The present report is devoted to the study of such phenomenon in two unidirectionally and mutually coupled dynamical systems being in chaotic (with one positive Lyapunov exponent) and hyperchaotic (with two or more positive Lyapunov exponents) regimes, respectively. As the systems under study, we have used two mutually coupled modified Lorenz oscillators and two unidirectionally coupled time-delayed generators. We have shown that in both cases the generalized synchronization regime can be detected by means of the calculation of Lyapunov exponents and phase tube approach whereas due to the complex topology of attractor the nearest neighbor method is misleading. Moreover, the auxiliary system approaches being the standard method for the synchronous regime observation, for the mutual type of coupling results in incorrect results. To calculate the Lyapunov exponents in time-delayed systems we have proposed an approach based on the modification of Gram-Schmidt orthogonalization procedure in the context of the time-delayed system. We have studied in detail the mechanisms resulting in the generalized synchronization regime onset paying a great attention to the field where one positive Lyapunov exponent has already been become negative whereas the second one is a positive yet. We have found the intermittency here and studied its characteristics. To detect the laminar phase lengths the method based on a calculation of local Lyapunov exponents has been proposed. The efficiency of the method has been verified using the example of two unidirectionally coupled Rössler systems being in the band chaos regime. We have revealed the main characteristics of intermittency, i.e. the distribution of the laminar phase lengths and dependence of the mean length of the laminar phases on the criticality parameter, for all systems studied in the report. This work has been supported by the Russian President's Council grant for the state support of young Russian scientists (project MK-531.2018.2).Keywords: complex topology of attractor, generalized synchronization, hyperchaos, Lyapunov exponents
Procedia PDF Downloads 275348 Application of Shore Protective Structures in Optimum Land Using of Defense Sites Located in Coastal Cities
Authors: Mir Ahmad Lashteh Neshaei, Hamed Afsoos Biria, Ata Ghabraei, Mir Abdolhamid Mehrdad
Abstract:
Awareness of effective land using issues in coastal area including protection of natural ecosystems and coastal environment due to the increasing of human life along the coast is of great importance. There are numerous valuable structures and heritages which are located in defence sites and waterfront area. Marine structures such as groins, sea walls and detached breakwaters are constructed in coast to improve the coast stability against bed erosion due to changing wave and climate pattern. Marine mechanisms and interaction with the shore protection structures need to be intensively studied. Groins are one of the most prominent structures that are used in shore protection to create a safe environment for coastal area by maintaining the land against progressive coastal erosion. The main structural function of a groin is to control the long shore current and littoral sediment transport. This structure can be submerged and provide the necessary beach protection without negative environmental impact. However, for submerged structures adopted for beach protection, the shoreline response to these structures is not well understood at present. Nowadays, modelling and computer simulation are used to assess beach morphology in the vicinity of marine structures to reduce their environmental impact. The objective of this study is to predict the beach morphology in the vicinity of submerged groins and comparison with non-submerged groins with focus on a part of the coast located in Dahane sar Sefidrood, Guilan province, Iran where serious coast erosion has occurred recently. The simulations were obtained using a one-line model which can be used as a first approximation of shoreline prediction in the vicinity of groins. The results of the proposed model are compared with field measurements to determine the shape of the coast. Finally, the results of the present study show that using submerged groins can have a good efficiency to control the beach erosion without causing severe environmental impact to the coast. The important outcome from this study can be employed in optimum designing of defence sites in the coastal cities to improve their efficiency in terms of re-using the heritage lands.Keywords: submerged structures, groin, shore protective structures, coastal cities
Procedia PDF Downloads 316347 Structural Analysis and Modelling in an Evolving Iron Ore Operation
Authors: Sameh Shahin, Nannang Arrys
Abstract:
Optimizing pit slope stability and reducing strip ratio of a mining operation are two key tasks in geotechnical engineering. With a growing demand for minerals and an increasing cost associated with extraction, companies are constantly re-evaluating the viability of mineral deposits and challenging their geological understanding. Within Rio Tinto Iron Ore, the Structural Geology (SG) team investigate and collect critical data, such as point based orientations, mapping and geological inferences from adjacent pits to re-model deposits where previous interpretations have failed to account for structurally controlled slope failures. Utilizing innovative data collection methods and data-driven investigation, SG aims to address the root causes of slope instability. Committing to a resource grid drill campaign as the primary source of data collection will often bias data collection to a specific orientation and significantly reduce the capability to identify and qualify complexity. Consequently, these limitations make it difficult to construct a realistic and coherent structural model that identifies adverse structural domains. Without the consideration of complexity and the capability of capturing these structural domains, mining operations run the risk of inadequately designed slopes that may fail and potentially harm people. Regional structural trends have been considered in conjunction with surface and in-pit mapping data to model multi-batter fold structures that were absent from previous iterations of the structural model. The risk is evident in newly identified dip-slope and rock-mass controlled sectors of the geotechnical design rather than a ubiquitous dip-slope sector across the pit. The reward is two-fold: 1) providing sectors of rock-mass controlled design in previously interpreted structurally controlled domains and 2) the opportunity to optimize the slope angle for mineral recovery and reduced strip ratio. Furthermore, a resulting high confidence model with structures and geometries that can account for historic slope instabilities in structurally controlled domains where design assumptions failed.Keywords: structural geology, geotechnical design, optimization, slope stability, risk mitigation
Procedia PDF Downloads 46346 A 1T1R Nonvolatile Memory with Al/TiO₂/Au and Sol-Gel Processed Barium Zirconate Nickelate Gate in Pentacene Thin Film Transistor
Authors: Ke-Jing Lee, Cheng-Jung Lee, Yu-Chi Chang, Li-Wen Wang, Yeong-Her Wang
Abstract:
To avoid the cross-talk issue of only resistive random access memory (RRAM) cell, one transistor and one resistor (1T1R) architecture with a TiO₂-based RRAM cell connected with solution barium zirconate nickelate (BZN) organic thin film transistor (OTFT) device is successfully demonstrated. The OTFT were fabricated on a glass substrate. Aluminum (Al) as the gate electrode was deposited via a radio-frequency (RF) magnetron sputtering system. The barium acetate, zirconium n-propoxide, and nickel II acetylacetone were synthesized by using the sol-gel method. After the BZN solution was completely prepared using the sol-gel process, it was spin-coated onto the Al/glass substrate as the gate dielectric. The BZN layer was baked at 100 °C for 10 minutes under ambient air conditions. The pentacene thin film was thermally evaporated on the BZN layer at a deposition rate of 0.08 to 0.15 nm/s. Finally, gold (Au) electrode was deposited using an RF magnetron sputtering system and defined through shadow masks as both the source and drain. The channel length and width of the transistors were 150 and 1500 μm, respectively. As for the manufacture of 1T1R configuration, the RRAM device was fabricated directly on drain electrodes of TFT device. A simple metal/insulator/metal structure, which consisting of Al/TiO₂/Au structures, was fabricated. First, Au was deposited to be a bottom electrode of RRAM device by RF magnetron sputtering system. Then, the TiO₂ layer was deposited on Au electrode by sputtering. Finally, Al was deposited as the top electrode. The electrical performance of the BZN OTFT was studied, showing superior transfer characteristics with the low threshold voltage of −1.1 V, good saturation mobility of 5 cm²/V s, and low subthreshold swing of 400 mV/decade. The integration of the BZN OTFT and TiO₂ RRAM devices was finally completed to form 1T1R configuration with low power consumption of 1.3 μW, the low operation current of 0.5 μA, and reliable data retention. Based on the I-V characteristics, the different polarities of bipolar switching are found to be determined by the compliance current with the different distribution of the internal oxygen vacancies used in the RRAM and 1T1R devices. Also, this phenomenon can be well explained by the proposed mechanism model. It is promising to make the 1T1R possible for practical applications of low-power active matrix flat-panel displays.Keywords: one transistor and one resistor (1T1R), organic thin-film transistor (OTFT), resistive random access memory (RRAM), sol-gel
Procedia PDF Downloads 354345 Islamic Finance and Trade Promotion in the African Continental Free Trade Area: An Exploratory Study
Authors: Shehu Usman Rano Aliyu
Abstract:
Despite the significance of finance as a major trade lubricant, evidence in the literature alludes to its scarcity and increasing cost, especially in developing countries where small and medium-scale enterprises are worst affected. The creation of the African Continental Free Trade Area (AFCFTA) in 2018, an organ of the African Union (AU), was meant to serve as a beacon for deepening economic integration through the removal of trade barriers inhibiting intra-African trade and movement of persons, among others. Hence, this research explores the role Islamic trade finance (ITF) could play in spurring intra- and inter-African trade. The study involves six countries; Egypt, Kenya, Malaysia, Morocco, Nigeria, and Saudi Arabia, and employs survey research, a total of 430 sample data, and SmartPLS Structural Equation Modelling (SEM) techniques in its analyses. We find strong evidence that Shari’ah, legal and regulatory compliance issues of the ITF institutions rhythm with the internal, national, and international compliance requirements equally as the unique instruments applied in ITF. In addition, ITF was found to be largely driven by global economic and political stability, socially responsible finance, ethical and moral considerations, risk-sharing, and resilience of the global Islamic finance industry. Further, SMEs, Governments, and Importers are the major beneficiary sectors. By and large, AfCFTA’s protocols align with the principles of ITF and are therefore suited for the proliferation of Islamic finance in the continent. And, while AML/KYC and BASEL requirements, compliance to AAOIFI and IFSB standards, paucity of Shari'ah experts, threats to global security, and increasing global economic uncertainty pose as major impediments, the future of ITF would be shaped by a greater need for institutional and policy support, global economic cum political stability, robust regulatory framework, and digital technology/fintech. The study calls for the licensing of more ITF institutions in the continent, participation of multilateral institutions in ITF, and harmonization of Shariah standards.Keywords: AfCFTA, islamic trade finance, murabaha, letter of credit, forwarding
Procedia PDF Downloads 56344 An Overview of the Wind and Wave Climate in the Romanian Nearshore
Authors: Liliana Rusu
Abstract:
The goal of the proposed work is to provide a more comprehensive picture of the wind and wave climate in the Romanian nearshore, using the results provided by numerical models. The Romanian coastal environment is located in the western side of the Black Sea, the more energetic part of the sea, an area with heavy maritime traffic and various offshore operations. Information about the wind and wave climate in the Romanian waters is mainly based on observations at Gloria drilling platform (70 km from the coast). As regards the waves, the measurements of the wave characteristics are not so accurate due to the method used, being also available for a limited period. For this reason, the wave simulations that cover large temporal and spatial scales represent an option to describe better the wave climate. To assess the wind climate in the target area spanning 1992–2016, data provided by the NCEP-CFSR (U.S. National Centers for Environmental Prediction - Climate Forecast System Reanalysis) and consisting in wind fields at 10m above the sea level are used. The high spatial and temporal resolution of the wind fields is good enough to represent the wind variability over the area. For the same 25-year period, as considered for the wind climate, this study characterizes the wave climate from a wave hindcast data set that uses NCEP-CFSR winds as input for a model system SWAN (Simulating WAves Nearshore) based. The wave simulation results with a two-level modelling scale have been validated against both in situ measurements and remotely sensed data. The second level of the system, with a higher resolution in the geographical space (0.02°×0.02°), is focused on the Romanian coastal environment. The main wave parameters simulated at this level are used to analyse the wave climate. The spatial distributions of the wind speed, wind direction and the mean significant wave height have been computed as the average of the total data. As resulted from the amount of data, the target area presents a generally moderate wave climate that is affected by the storm events developed in the Black Sea basin. Both wind and wave climate presents high seasonal variability. All the results are computed as maps that help to find the more dangerous areas. A local analysis has been also employed in some key locations corresponding to highly sensitive areas, as for example the main Romanian harbors.Keywords: numerical simulations, Romanian nearshore, waves, wind
Procedia PDF Downloads 344343 Media Impression and Its Impact on Foreign Policy Making: A Study of India-China Relations
Authors: Rosni Lakandri
Abstract:
With the development of science and technology, there has been a complete transformation in the domain of information technology. Particularly after the Second World War and Cold War period, the role of media and communication technology in shaping the political, economic, socio-cultural proceedings across the world has been tremendous. It performs as a channel between the governing bodies of the state and the general masses. As we have seen the international community constantly talking about the onset of Asian Century, India and China happens to be the major player in this. Both have the civilization history, both are neighboring countries, both are witnessing a huge economic growth and, important of all, both are considered the rising powers of Asia. Not negating the fact that both countries have gone to war with each other in 1962 and the common people and even the policy makers of both the sides view each other till now from this prism. A huge contribution to this perception of people goes to the media coverage of both sides, even if there are spaces of cooperation which they share, the negative impacts of media has tended to influence the people’s opinion and government’s perception about each other. Therefore, analysis of media’s impression in both the countries becomes important in order to know their effect on the larger implications of foreign policy towards each other. It is usually said that media not only acts as the information provider but also acts as ombudsman to the government. They provide a kind of check and balance to the governments in taking proper decisions for the people of the country but in attempting to answer this hypothesis we have to analyze does the media really helps in shaping the political landscape of any country? Therefore, this study rests on the following questions; 1.How do China and India depict each other through their respective News media? 2.How much and what influences they make on the policy making process of each country? How do they shape the public opinion in both the countries? In order to address these enquiries, the study employs both primary and secondary sources available, and in generating data and other statistical information, primary sources like reports, government documents, and cartography, agreements between the governments have been used. Secondary sources like books, articles and other writings collected from various sources and opinion from visual media sources like news clippings, videos in this topic are also included as a source of on ground information as this study is not based on field study. As the findings suggest in case of China and India, media has certainly affected people’s knowledge about the political and diplomatic issues at the same time has affected the foreign policy making of both the countries. They have considerable impact on the foreign policy formulation and we can say there is some mediatization happening in foreign policy issues in both the countries.Keywords: China, foreign policy, India, media, public opinion
Procedia PDF Downloads 151342 Predictability of Thermal Response in Housing: A Case Study in Australia, Adelaide
Authors: Mina Rouhollahi, J. Boland
Abstract:
Changes in cities’ heat balance due to rapid urbanization and the urban heat island (UHI) have increased energy demands for space cooling and have resulted in uncomfortable living conditions for urban residents. Climate resilience and comfortable living spaces can be addressed through well-designed urban development. The sustainable housing can be more effective in controlling high levels of urban heat. In Australia, to mitigate the effects of UHIs and summer heat waves, one solution to sustainable housing has been the trend to compact housing design and the construction of energy efficient dwellings. This paper analyses whether current housing configurations and orientations are effective in avoiding increased demands for air conditioning and having an energy efficient residential neighborhood. A significant amount of energy is consumed to ensure thermal comfort in houses. This paper reports on the modelling of heat transfer within the homes using the measurements of radiation, convection and conduction between exterior/interior wall surfaces and outdoor/indoor environment respectively. The simulation was tested on selected 7.5-star energy efficient houses constructed of typical material elements and insulation in Adelaide, Australia. The chosen design dwellings were analyzed in extremely hot weather through one year. The data were obtained via a thermal circuit to accurately model the fundamental heat transfer mechanisms on both boundaries of the house and through the multi-layered wall configurations. The formulation of the Lumped capacitance model was considered in discrete time steps by adopting a non-linear model method. The simulation results focused on the effects of orientation of the solar radiation on the dynamic thermal characteristics of the houses orientations. A high star rating did not necessarily coincide with a decrease in peak demands for cooling. A more effective approach to avoid increasing the demands for air conditioning and energy may be to integrate solar–climatic data to evaluate the performance of energy efficient houses.Keywords: energy-efficient residential building, heat transfer, neighborhood orientation, solar–climatic data
Procedia PDF Downloads 133341 Investigation of Dry-Blanching and Freezing Methods of Fruits
Authors: Epameinondas Xanthakis, Erik Kaunisto, Alain Le-Bail, Lilia Ahrné
Abstract:
Fruits and vegetables are characterized as perishable food matrices due to their short shelf life as several deterioration mechanisms are being involved. Prior to the common preservation methods like freezing or canning, fruits and vegetables are being blanched in order to inactivate deteriorative enzymes. Both conventional blanching pretreatments and conventional freezing methods hide drawbacks behind their beneficial impacts on the preservation of those matrices. Conventional blanching methods may require longer processing times, leaching of minerals and nutrients due to the contact with the warm water which in turn leads to effluent production with large BOD. An important issue of freezing technologies is the size of the formed ice crystals which is also critical for the final quality of the frozen food as it can cause irreversible damage to the cellular structure and subsequently to degrade the texture and the colour of the product. Herein, the developed microwave blanching methodology and the results regarding quality aspects and enzyme inactivation will be presented. Moreover, heat transfer phenomena, mass balance, temperature distribution, and enzyme inactivation (such as Pectin Methyl Esterase and Ascorbic Acid Oxidase) of our microwave blanching approach will be evaluated based on measurements and computer modelling. The present work is part of the COLDμWAVE project which aims to the development of an innovative environmentally sustainable process for blanching and freezing of fruits and vegetables with improved textural and nutritional quality. In this context, COLDµWAVE will develop tailored equipment for MW blanching of vegetables that has very high energy efficiency and no water consumption. Furthermore, the next steps of this project regarding the development of innovative pathways in MW assisted freezing to improve the quality of frozen vegetables, by exploring in depth previous results acquired by the authors, will be presented. The application of MW assisted freezing process on fruits and vegetables it is expected to lead to improved quality characteristics compared to the conventional freezing. Acknowledgments: COLDμWAVE has received funding from the European Union’s Horizon 2020 research and innovation programme under the Marie Sklodowska-Curie grand agreement No 660067.Keywords: blanching, freezing, fruits, microwave blanching, microwave
Procedia PDF Downloads 267340 Microfluidic Plasmonic Bio-Sensing of Exosomes by Using a Gold Nano-Island Platform
Authors: Srinivas Bathini, Duraichelvan Raju, Simona Badilescu, Muthukumaran Packirisamy
Abstract:
A bio-sensing method, based on the plasmonic property of gold nano-islands, has been developed for detection of exosomes in a clinical setting. The position of the gold plasmon band in the UV-Visible spectrum depends on the size and shape of gold nanoparticles as well as on the surrounding environment. By adsorbing various chemical entities, or binding them, the gold plasmon band will shift toward longer wavelengths and the shift is proportional to the concentration. Exosomes transport cargoes of molecules and genetic materials to proximal and distal cells. Presently, the standard method for their isolation and quantification from body fluids is by ultracentrifugation, not a practical method to be implemented in a clinical setting. Thus, a versatile and cutting-edge platform is required to selectively detect and isolate exosomes for further analysis at clinical level. The new sensing protocol, instead of antibodies, makes use of a specially synthesized polypeptide (Vn96), to capture and quantify the exosomes from different media, by binding the heat shock proteins from exosomes. The protocol has been established and optimized by using a glass substrate, in order to facilitate the next stage, namely the transfer of the protocol to a microfluidic environment. After each step of the protocol, the UV-Vis spectrum was recorded and the position of gold Localized Surface Plasmon Resonance (LSPR) band was measured. The sensing process was modelled, taking into account the characteristics of the nano-island structure, prepared by thermal convection and annealing. The optimal molar ratios of the most important chemical entities, involved in the detection of exosomes were calculated as well. Indeed, it was found that the results of the sensing process depend on the two major steps: the molar ratios of streptavidin to biotin-PEG-Vn96 and, the final step, the capture of exosomes by the biotin-PEG-Vn96 complex. The microfluidic device designed for sensing of exosomes consists of a glass substrate, sealed by a PDMS layer that contains the channel and a collecting chamber. In the device, the solutions of linker, cross-linker, etc., are pumped over the gold nano-islands and an Ocean Optics spectrometer is used to measure the position of the Au plasmon band at each step of the sensing. The experiments have shown that the shift of the Au LSPR band is proportional to the concentration of exosomes and, thereby, exosomes can be accurately quantified. An important advantage of the method is the ability to discriminate between exosomes having different origins.Keywords: exosomes, gold nano-islands, microfluidics, plasmonic biosensing
Procedia PDF Downloads 172339 Automatic Aggregation and Embedding of Microservices for Optimized Deployments
Authors: Pablo Chico De Guzman, Cesar Sanchez
Abstract:
Microservices are a software development methodology in which applications are built by composing a set of independently deploy-able, small, modular services. Each service runs a unique process and it gets instantiated and deployed in one or more machines (we assume that different microservices are deployed into different machines). Microservices are becoming the de facto standard for developing distributed cloud applications due to their reduced release cycles. In principle, the responsibility of a microservice can be as simple as implementing a single function, which can lead to the following issues: - Resource fragmentation due to the virtual machine boundary. - Poor communication performance between microservices. Two composition techniques can be used to optimize resource fragmentation and communication performance: aggregation and embedding of microservices. Aggregation allows the deployment of a set of microservices on the same machine using a proxy server. Aggregation helps to reduce resource fragmentation, and is particularly useful when the aggregated services have a similar scalability behavior. Embedding deals with communication performance by deploying on the same virtual machine those microservices that require a communication channel (localhost bandwidth is reported to be about 40 times faster than cloud vendor local networks and it offers better reliability). Embedding can also reduce dependencies on load balancer services since the communication takes place on a single virtual machine. For example, assume that microservice A has two instances, a1 and a2, and it communicates with microservice B, which also has two instances, b1 and b2. One embedding can deploy a1 and b1 on machine m1, and a2 and b2 are deployed on a different machine m2. This deployment configuration allows each pair (a1-b1), (a2-b2) to communicate using the localhost interface without the need of a load balancer between microservices A and B. Aggregation and embedding techniques are complex since different microservices might have incompatible runtime dependencies which forbid them from being installed on the same machine. There is also a security concern since the attack surface between microservices can be larger. Luckily, container technology allows to run several processes on the same machine in an isolated manner, solving the incompatibility of running dependencies and the previous security concern, thus greatly simplifying aggregation/embedding implementations by just deploying a microservice container on the same machine as the aggregated/embedded microservice container. Therefore, a wide variety of deployment configurations can be described by combining aggregation and embedding to create an efficient and robust microservice architecture. This paper presents a formal method that receives a declarative definition of a microservice architecture and proposes different optimized deployment configurations by aggregating/embedding microservices. The first prototype is based on i2kit, a deployment tool also submitted to ICWS 2018. The proposed prototype optimizes the following parameters: network/system performance, resource usage, resource costs and failure tolerance.Keywords: aggregation, deployment, embedding, resource allocation
Procedia PDF Downloads 203338 Modeling Acceptability of a Personalized and Contextualized Radio Embedded in Vehicles
Authors: Ludivine Gueho, Sylvain Fleury, Eric Jamet
Abstract:
Driver distraction is known to be a major contributing factor of car accidents. Since many years, constructors have been designing embedded technologies to face this problem and reduce distraction. Being able to predict user acceptance would further be helpful in the development process to build appropriate systems. The present research aims at modelling the acceptability of a specific system, an innovative personalized and contextualized embedded radio, through an online survey of 202 people in France that assessed the psychological variables determining intentions to use the system. The questionnaire instantiated the dimensions of the extended version of the UTAUT acceptability model. Because of the specific features of the system assessed, we added 4 dimensions: perceived security, anxiety, trust and privacy concerns. Results showed that hedonic motivation, i.e., the fun or pleasure derived from using a technology, and performance expectancy, i.e., the degree to which individuals believe that the characteristics of the system meet their needs, are the most important dimensions in determining behavioral intentions about the innovative radio. To a lesser extent, social influence, i.e., the degree to which individuals think they can use the system while respecting their social group’s norms and while giving a positive image of themselves, had an effect on behavioral intentions. Moreover, trust, that is, the positive belief about the perceived reliability of, dependability of, and confidence in a person, object or process, had a significant effect, mediated by performance expectancy. In an applicative way, the present research reveals that, to be accepted, in-car embedded new technology has to address individual needs, for instance by facilitating the driving activity or by providing useful information. If it shows hedonic qualities by being entertaining, pretty or comfortable, this may improve the intentions to use it. Therefore, it is clearly important to include reflection about user experience in the design process. Finally, the users have to be reassured on the system’s reliability. For example, improving the transparency of the system by providing information about the system functioning, could improve trust. These results bring some highlights on determinant of acceptance of an in-vehicle technology and are useful for constructors to design acceptable systems.Keywords: acceptability, innovative embedded radio, structural equation, user-centric evaluation, UTAUT
Procedia PDF Downloads 269337 Deep Mill Level Zone (DMLZ) of Ertsberg East Skarn System, Papua; Correlation between Structure and Mineralization to Determined Characteristic Orebody of DMLZ Mine
Authors: Bambang Antoro, Lasito Soebari, Geoffrey de Jong, Fernandy Meiriyanto, Michael Siahaan, Eko Wibowo, Pormando Silalahi, Ruswanto, Adi Budirumantyo
Abstract:
The Ertsberg East Skarn System (EESS) is located in the Ertsberg Mining District, Papua, Indonesia. EESS is a sub-vertical zone of copper-gold mineralization hosted in both diorite (vein-style mineralization) and skarn (disseminated and vein style mineralization). Deep Mill Level Zone (DMLZ) is a mining zone in the lower part of East Ertsberg Skarn System (EESS) that product copper and gold. The Deep Mill Level Zone deposit is located below the Deep Ore Zone deposit between the 3125m to 2590m elevation, measures roughly 1,200m in length and is between 350 and 500m in width. DMLZ planned start mined on Q2-2015, being mined at an ore extraction rate about 60,000 tpd by the block cave mine method (the block cave contain 516 Mt). Mineralization and associated hydrothermal alteration in the DMLZ is hosted and enclosed by a large stock (The Main Ertsberg Intrusion) that is barren on all sides and above the DMLZ. Late porphyry dikes that cut through the Main Ertsberg Intrusion are spatially associated with the center of the DMLZ hydrothermal system. DMLZ orebody hosted in diorite and skarn, both dominantly by vein style mineralization. Percentage Material Mined at DMLZ compare with current Reserves are diorite 46% (with 0.46% Cu; 0.56 ppm Au; and 0.83% EqCu); Skarn is 39% (with 1.4% Cu; 0.95 ppm Au; and 2.05% EqCu); Hornfels is 8% (with 0.84% Cu; 0.82 ppm Au; and 1.39% EqCu); and Marble 7 % possible mined waste. Correlation between Ertsberg intrusion, major structure, and vein style mineralization is important to determine characteristic orebody in DMLZ Mine. Generally Deep Mill Level Zone has 2 type of vein filling mineralization from both hosted (diorite and skarn), in diorite hosted the vein system filled by chalcopyrite-bornite-quartz and pyrite, in skarn hosted the vein filled by chalcopyrite-bornite-pyrite and magnetite without quartz. Based on orientation the stockwork vein at diorite hosted and shallow vein in skarn hosted was generally NW-SE trending and NE-SW trending with shallow-moderate dipping. Deep Mill Level Zone control by two main major faults, geologist founded and verified local structure between major structure with NW-SE trending and NE-SW trending with characteristics slickenside, shearing, gauge, water-gas channel, and some has been re-healed.Keywords: copper-gold, DMLZ, skarn, structure
Procedia PDF Downloads 501336 Spanish Language Violence Corpus: An Analysis of Offensive Language in Twitter
Authors: Beatriz Botella-Gil, Patricio Martínez-Barco, Lea Canales
Abstract:
The Internet and ICT are an integral element of and omnipresent in our daily lives. Technologies have changed the way we see the world and relate to it. The number of companies in the ICT sector is increasing every year, and there has also been an increase in the work that occurs online, from sending e-mails to the way companies promote themselves. In social life, ICT’s have gained momentum. Social networks are useful for keeping in contact with family or friends that live far away. This change in how we manage our relationships using electronic devices and social media has been experienced differently depending on the age of the person. According to currently available data, people are increasingly connected to social media and other forms of online communication. Therefore, it is no surprise that violent content has also made its way to digital media. One of the important reasons for this is the anonymity provided by social media, which causes a sense of impunity in the victim. Moreover, it is not uncommon to find derogatory comments, attacking a person’s physical appearance, hobbies, or beliefs. This is why it is necessary to develop artificial intelligence tools that allow us to keep track of violent comments that relate to violent events so that this type of violent online behavior can be deterred. The objective of our research is to create a guide for detecting and recording violent messages. Our annotation guide begins with a study on the problem of violent messages. First, we consider the characteristics that a message should contain for it to be categorized as violent. Second, the possibility of establishing different levels of aggressiveness. To download the corpus, we chose the social network Twitter for its ease of obtaining free messages. We chose two recent, highly visible violent cases that occurred in Spain. Both of them experienced a high degree of social media coverage and user comments. Our corpus has a total of 633 messages, manually tagged, according to the characteristics we considered important, such as, for example, the verbs used, the presence of exclamations or insults, and the presence of negations. We consider it necessary to create wordlists that are present in violent messages as indicators of violence, such as lists of negative verbs, insults, negative phrases. As a final step, we will use automatic learning systems to check the data obtained and the effectiveness of our guide.Keywords: human language technologies, language modelling, offensive language detection, violent online content
Procedia PDF Downloads 131335 Process Performance and Nitrogen Removal Kinetics in Anammox Hybrid Reactor
Authors: Swati Tomar, Sunil Kumar Gupta
Abstract:
Anammox is a promising and cost effective alternative to conventional treatment systems that facilitates direct oxidation of ammonium nitrogen under anaerobic conditions with nitrite as an electron acceptor without addition of any external carbon sources. The present study investigates the process kinetics of laboratory scale anammox hybrid reactor (AHR) which combines the dual advantages of attached and suspended growth. The performance & behaviour of AHR was studied under varying hydraulic retention time (HRTs) and nitrogen loading rate (NLRs). The experimental unit consisted of 4 numbers of 5L capacity anammox hybrid reactor inoculated with mixed seed culture containing anoxic and activated sludge. Pseudo steady state (PSS) ammonium and nitrite removal efficiencies of 90.6% and 95.6%, respectively, were achieved during acclimation phase. After establishment of PSS, the performance of AHR was monitored at seven different HRTs of 3.0, 2.5, 2.0, 1.5, 1.0, 0.5 and 0.25 d with increasing NLR from 0.4 to 4.8 kg N/m3d. The results showed that with increase in NLR and decrease in HRT (3.0 to 0.25 d), AHR registered appreciable decline in nitrogen removal efficiency from 92.9% to 67.4 %, respectively. The HRT of 2.0 d was considered optimal to achieve substantial nitrogen removal of 89%, because on further decrease in HRT below 1.5 days, remarkable decline in the values of nitrogen removal efficiency were observed. Analysis of data indicated that attached growth system contributes an additional 15.4 % ammonium removal and reduced the sludge washout rate (additional 29% reduction). This enhanced performance may be attributed to 25% increase in sludge retention time due to the attached growth media. Three kinetic models, namely, first order, Monod and Modified Stover-Kincannon model were applied to assess the substrate removal kinetics of nitrogen removal in AHR. Validation of the models were carried out by comparing experimental set of data with the predicted values obtained from the respective models. For substrate removal kinetics, model validation revealed that Modified Stover-Kincannon is most precise (R2=0.943) and can be suitably applied to predict the kinetics of nitrogen removal in AHR. Lawrence and McCarty model described the kinetics of bacterial growth. The predicted value of yield coefficient and decay constant were in line with the experimentally observed values.Keywords: anammox, kinetics, modelling, nitrogen removal, sludge wash out rate, AHR
Procedia PDF Downloads 315334 A Conceptual Model of the Factors Affecting Saudi Citizens' Use of Social Media to Communicate with the Government
Authors: Reemiah Alotaibi, Muthu Ramachandran, Ah-Lian Kor, Amin Hosseinian-Far
Abstract:
In the past decade, developers of Web 2.0 technologies have shown increasing interest in the topic of e-government. There has been a rapid growth in social media technology because of its significant role in backing up some essential social needs. Its importance and power is derived from its capacity to support two-way communication. Governments are curious to get engaged in these websites, hoping to benefit from the new forms of communication and interaction offered by such technology. Greater participation by the public can be viewed as a chief indicator of effective government communication. Yet, the level of public participation in government 2.0 is not quite satisfactory. In general, it is still at the early stage in most developing countries, including Saudi Arabia. Although it is a fact that Saudi people are among the most active in using social media, the number of people who use social media to communicate with the public institutions is not high. Furthermore, most of the governmental organisations are not using social media tools to communicate with the public. They use these platforms to disseminate information. Our study focuses on the factors affecting citizens’ adoption of social media in Saudi Arabia. Our research question is: what are the factors affecting Saudi citizens’ use of social media to communicate with the government? To answer this research question, the research aims to validate the UTAUT model for examining social media tools from the citizen perspective. An amendment will be proposed to fit the adoption of social media platforms as a communication channel in government by using a developed conceptual model which integrates constructs from the UTAUT model and others external variables based on the literature review. The set of potential factors that affect these citizens' decisions to adopt social media to communicate with their government has been identified as perceived encouragement, trust and cultural influence. The connection between the above-mentioned constructs from the basis for the research hypothesis will be examined in the light of a quantitative methodology. Data collection will be performed through a survey targeting a number of Saudi citizens who are social media users. The data collected from the primary survey will later be analysed by using statistical methods. The outcomes of this research project are argued to have potential contributions to the fields of social media and e-Government adoption, both on the theoretical and practical levels. It is believed that this research project is the first of its type that attempts to identify the factors that affect citizens’ adoption of social media to communicate with the government. The importance of identifying these factors stems from the potential use of them to enhance the government’s implementation of social media and help in making more accurate decisions and strategies based on comprehending the most important factors that affect citizens’ decisions.Keywords: social media, adoption, citizen, UTAUT model
Procedia PDF Downloads 418333 Measuring the Visibility of the European Open Access Journals with Bibliometric Indicators
Authors: Maja Jokić, Andrea Mervar, Stjepan Mateljan
Abstract:
Peer review journals, as the main communication channel among researchers, fully achieve their objective if they are available to the global research community, which is accomplished through open access. In the EU countries, the idea of open access has spread over the years through various projects, initiatives, and strategic documents. Consequently, in this paper we want to analyze, using various bibliometric indicators, visibility, and significance of open access peer review journals compared to the conventional (non-open access) ones. We examine the sample of open access (OA) journals in 28 EU countries in addition to open access journals in three EU candidate countries (Bosnia and Herzegovina, FYR Macedonia and Serbia), all indexed by Scopus (N=1,522). These journals comprise 42% of the total number of OA journals indexed by Scopus. The distribution of OA journals in our sample according to the subject fields indicates that the largest share has OA journals in Health Sciences, 29% followed by Social Sciences and Physical Sciences with 25%, and 21% in Life Sciences. At the same time, the distribution according to countries (N=31) shows the dominance of EU15 countries with the share of 68.3% (N=1041) while post-socialist European countries (EU11 plus three candidate EU countries) have the share of 31.6% (N=481). Bibliometric indicators are derived from the SCImago Journal Ranking database. The analysis of OA journals according to their quartile scores (that reflect the relation between number of articles and their citations) shows that the largest number of OA journals from our sample was in the third quartile in 2015. For comparison, the majority of all academic journals indexed in Scopus from the countries in our sample were in the same year in the first quartile. The median of SJR indicator (SCImago Journal Rankings) for 2015 that measures the journal's prestige, amounted 0.297 for OA journals from the sample, while it was modestly lower for all OA journals, 0.284. The value of the same indicator for all journals indexed by Scopus (N=11,086) from our group of countries was 0.358, which is significantly different from the one for OA journals. Apart from the number of OA journals we also confirm significant differences between EU15 and post-socialist countries in bibliometric status of OA journals. The median SJR indicator for 2015 for EU15 countries was 0.394, while for post-socialist countries it amounted to 0.226. The changes in bibliometric indicators: quartile score, SJR (SCImago Journal Rankings), SNIP (Sources Normalised Impact by Paper) and IPP (Impact per Publication) of OA journals during 2012-2015 period, as well as H-index for the main four subject fields (Life Sciences, Physical Sciences, Social Sciences and Health Sciences) in the whole sample as well as in two main groups of European countries, show increasing trend of acceptance and visibility of OA journals within the academic community. More comprehensive insights into the visibility of OA journals could be reached by using additional qualitative research methods such as for example, interviews with researchers.Keywords: bibliometric analysis, European countries, journal evaluation, open access journals
Procedia PDF Downloads 222332 Reliable and Error-Free Transmission through Multimode Polymer Optical Fibers in House Networks
Authors: Tariq Ahamad, Mohammed S. Al-Kahtani, Taisir Eldos
Abstract:
Optical communications technology has made enormous and steady progress for several decades, providing the key resource in our increasingly information-driven society and economy. Much of this progress has been in finding innovative ways to increase the data carrying capacity of a single optical fiber. In this research article we have explored basic issues in terms of security and reliability for secure and reliable information transfer through the fiber infrastructure. Conspicuously, one potentially enormous source of improvement has however been left untapped in these systems: fibers can easily support hundreds of spatial modes, but today’s commercial systems (single-mode or multi-mode) make no attempt to use these as parallel channels for independent signals. Bandwidth, performance, reliability, cost efficiency, resiliency, redundancy, and security are some of the demands placed on telecommunications today. Since its initial development, fiber optic systems have had the advantage of most of these requirements over copper-based and wireless telecommunications solutions. The largest obstacle preventing most businesses from implementing fiber optic systems was cost. With the recent advancements in fiber optic technology and the ever-growing demand for more bandwidth, the cost of installing and maintaining fiber optic systems has been reduced dramatically. With so many advantages, including cost efficiency, there will continue to be an increase of fiber optic systems replacing copper-based communications. This will also lead to an increase in the expertise and the technology needed to tap into fiber optic networks by intruders. As ever before, all technologies have been subject to hacking and criminal manipulation, fiber optics is no exception. Researching fiber optic security vulnerabilities suggests that not everyone who is responsible for their networks security is aware of the different methods that intruders use to hack virtually undetected into fiber optic cables. With millions of miles of fiber optic cables stretching across the globe and carrying information including but certainly not limited to government, military, and personal information, such as, medical records, banking information, driving records, and credit card information; being aware of fiber optic security vulnerabilities is essential and critical. Many articles and research still suggest that fiber optics is expensive, impractical and hard to tap. Others argue that it is not only easily done, but also inexpensive. This paper will briefly discuss the history of fiber optics, explain the basics of fiber optic technologies and then discuss the vulnerabilities in fiber optic systems and how they can be better protected. Knowing the security risks and knowing the options available may save a company a lot embarrassment, time, and most importantly money.Keywords: in-house networks, fiber optics, security risk, money
Procedia PDF Downloads 420331 Algae Biofertilizers Promote Sustainable Food Production and Nutrient Efficiency: An Integrated Empirical-Modeling Study
Authors: Zeenat Rupawalla, Nicole Robinson, Susanne Schmidt, Sijie Li, Selina Carruthers, Elodie Buisset, John Roles, Ben Hankamer, Juliane Wolf
Abstract:
Agriculture has radically changed the global biogeochemical cycle of nitrogen (N). Fossil fuel-enabled synthetic N-fertiliser is a foundation of modern agriculture but applied to soil crops only use about half of it. To address N-pollution from cropping and the large carbon and energy footprint of N-fertiliser synthesis, new technologies delivering enhanced energy efficiency, decarbonisation, and a circular nutrient economy are needed. We characterised algae fertiliser (AF) as an alternative to synthetic N-fertiliser (SF) using empirical and modelling approaches. We cultivated microalgae in nutrient solution and modelled up-scaled production in nutrient-rich wastewater. Over four weeks, AF released 63.5% of N as ammonium and nitrate, and 25% of phosphorous (P) as phosphate to the growth substrate, while SF released 100% N and 20% P. To maximise crop N-use and minimise N-leaching, we explored AF and SF dose-response-curves with spinach in glasshouse conditions. AF-grown spinach produced 36% less biomass than SF-grown plants due to AF’s slower and linear N-release, while SF resulted in 5-times higher N-leaching loss than AF. Optimised blends of AF and SF boosted crop yield and minimised N-loss due to greater synchrony of N-release and crop uptake. Additional benefits of AF included greener leaves, lower leaf nitrate concentration, and higher microbial diversity and water holding capacity in the growth substrate. Life-cycle-analysis showed that replacing the most effective SF dosage with AF lowered the carbon footprint of fertiliser production from 2.02 g CO₂ (C-producing) to -4.62 g CO₂ (C-sequestering), with a further 12% reduction when AF is produced on wastewater. Embodied energy was lowest for AF-SF blends and could be reduced by 32% when cultivating algae on wastewater. We conclude that (i) microalgae offer a sustainable alternative to synthetic N-fertiliser in spinach production and potentially other crop systems, and (ii) microalgae biofertilisers support the circular nutrient economy and several sustainable development goals.Keywords: bioeconomy, decarbonisation, energy footprint, microalgae
Procedia PDF Downloads 137330 Field Synergy Analysis of Combustion Characteristics in the Afterburner of Solid Oxide Fuel Cell System
Authors: Shing-Cheng Chang, Cheng-Hao Yang, Wen-Sheng Chang, Chih-Chia Lin, Chun-Han Li
Abstract:
The solid oxide fuel cell (SOFC) is a promising green technology which can achieve a high electrical efficiency. Due to the high operating temperature of SOFC stack, the off-gases at high temperature from anode and cathode outlets are introduced into an afterburner to convert the chemical energy into thermal energy by combustion. The heat is recovered to preheat the fresh air and fuel gases before they pass through the stack during the SOFC power generation system operation. For an afterburner of the SOFC system, the temperature control with a good thermal uniformity is important. A burner with a well-designed geometry usually can achieve a satisfactory performance. To design an afterburner for an SOFC system, the computational fluid dynamics (CFD) simulation is adoptable. In this paper, the hydrogen combustion characteristics in an afterburner with simple geometry are studied by using CFD. The burner is constructed by a cylinder chamber with the configuration of a fuel gas inlet, an air inlet, and an exhaust outlet. The flow field and temperature distributions inside the afterburner under different fuel and air flow rates are analyzed. To improve the temperature uniformity of the afterburner during the SOFC system operation, the flow paths of anode/cathode off-gases are varied by changing the positions of fuels and air inlet channel to improve the heat and flow field synergy in the burner furnace. Because the air flow rate is much larger than the fuel gas, the flow structure and heat transfer in the afterburner is dominated by the air flow path. The present work studied the effects of fluid flow structures on the combustion characteristics of an SOFC afterburner by three simulation models with a cylindrical combustion chamber and a tapered outlet. All walls in the afterburner are assumed to be no-slip and adiabatic. In each case, two set of parameters are simulated to study the transport phenomena of hydrogen combustion. The equivalence ratios are in the range of 0.08 to 0.1. Finally, the pattern factor for the simulation cases is calculated to investigate the effect of gas inlet locations on the temperature uniformity of the SOFC afterburner. The results show that the temperature uniformity of the exhaust gas can be improved by simply adjusting the position of the gas inlet. The field synergy analysis indicates the design of the fluid flow paths should be in the way that can significantly contribute to the heat transfer, i.e. the field synergy angle should be as small as possible. In the study cases, the averaged synergy angle of the burner is about 85̊, 84̊, and 81̊ respectively.Keywords: afterburner, combustion, field synergy, solid oxide fuel cell
Procedia PDF Downloads 137329 Endoscopic Stenting of the Main Pancreatic Duct in Patients With Pancreatic Fluid Collections After Pancreas Transplantation
Authors: Y. Teterin, S. Suleymanova, I. Dmitriev, P. Yartcev
Abstract:
Introduction: One of the most common complications after pancreas transplantation are pancreatic fluid collections (PFCs), which are often complicated not only by infection and subsequent disfunction of the pancreatoduodenal graft (PDG), but also with a rather high mortality rate of recipients. Drainage is not always effective and often requires repeated open surgical interventions, which worsens the outcome of the surgery. Percutaneous drainage of PFCs combined with endoscopic stenting of the main pancreatic duct of the pancreatoduodenal graft (MPDPDG) showed high efficiency in the treatment of PFCs. Aims & Methods: From 01.01.2012 to 31.12.2021 at the Sklifosovsky Research Institute for Emergency Medicine were performed 64 transplantations of PDG. In 11 cases (17.2%), the early postoperative period was complicated by the formation of PFCs. Of these, 7 patients underwent percutaneous drainage of pancreonecrosis with high efficiency and did not required additional methods of treatment. In the remaining 4 patients, drainage was ineffective and was an indication for endoscopic stenting of the MPDPDG. They were the ones who made up the study group. Among them were 3 men and 1 woman. The mean age of the patients was 36,4 years.PFCs in these patients formed on days 1, 12, 18, and 47 after PDG transplantation. We used a gastroscope to stent the MPDPDG, due to anatomical features of the location of the duodenoduodenal anastomosis after PDG transplantation. Through the endoscope channel was performed selective catheterization of the MPDPDG, using a catheter and a guidewire, followed by its contrasting with a water-soluble contrast agent. Due to the extravasation of the contrast, was determined the localization of the defect in the PDG duct system. After that, a plastic pancreatic stent with a diameter of 7 Fr. and a length of 7 cm. was installed along guidewire. The stent was installed in such a way that its proximal edge completely covered the defect zone, and the distal one was determined in the intestinal lumen. Results: In all patients PDG pancreaticography revealed extravasation of a contrast in the area of the isthmus and body of the pancreas, which required stenting of the MPDPDG. In 1 (25%) case, the patient had a dislocation of the stent into the intestinal lumen (III degree according to Clavien-Dindo (2009)). This patient underwent repeated endoscopic stenting of the MPDPDG. On average 23 days after endoscopic stenting of the MPDPDG, the drainage tubes were removed and after approximately 40 days all patients were discharged in a satisfactory condition with follow-up endocrinologist and surgeon consultation. Pancreatic stents were removed after 6 months ± 7 days. Conclusion: Endoscopic stenting of the main pancreatic duct of the donor pancreas is by far the most highly effective and minimally invasive method in the treatment of PFCs after transplantation of the pancreatoduodenal complex.Keywords: pancreas transplantation, endoscopy surgery, diabetes, stenting, main pancreatic duct
Procedia PDF Downloads 86328 Assessing the Blood-Brain Barrier (BBB) Permeability in PEA-15 Mutant Cat Brain using Magnetization Transfer (MT) Effect at 7T
Authors: Sultan Z. Mahmud, Emily C. Graff, Adil Bashir
Abstract:
Phosphoprotein enriched in astrocytes 15 kDa (PEA-15) is a multifunctional adapter protein which is associated with the regulation of apoptotic cell death. Recently it has been discovered that PEA-15 is crucial in normal neurodevelopment of domestic cats, a gyrencephalic animal model, although the exact function of PEA-15 in neurodevelopment is unknown. This study investigates how PEA-15 affects the blood-brain barrier (BBB) permeability in cat brain, which can cause abnormalities in tissue metabolite and energy supplies. Severe polymicrogyria and microcephaly have been observed in cats with a loss of function PEA-15 mutation, affecting the normal neurodevelopment of the cat. This suggests that the vital role of PEA-15 in neurodevelopment is associated with gyrification. Neurodevelopment is a highly energy demanding process. The mammalian brain depends on glucose as its main energy source. PEA-15 plays a very important role in glucose uptake and utilization by interacting with phospholipase D1 (PLD1). Mitochondria also plays a critical role in bioenergetics and essential to supply adequate energy needed for neurodevelopment. Cerebral blood flow regulates adequate metabolite supply and recent findings also showed that blood plasma contains mitochondria as well. So the BBB can play a very important role in regulating metabolite and energy supply in the brain. In this study the blood-brain permeability in cat brain was measured using MRI magnetization transfer (MT) effect on the perfusion signal. Perfusion is the tissue mass normalized supply of blood to the capillary bed. Perfusion also accommodates the supply of oxygen and other metabolites to the tissue. A fraction of the arterial blood can diffuse to the tissue, which depends on the BBB permeability. This fraction is known as water extraction fraction (EF). MT is a process of saturating the macromolecules, which has an effect on the blood that has been diffused into the tissue while having minimal effect on intravascular blood water that has not been exchanged with the tissue. Measurement of perfusion signal with and without MT enables to estimate the microvascular blood flow, EF and permeability surface area product (PS) in the brain. All the experiments were performed with Siemens 7T Magnetom with 32 channel head coil. Three control cats and three PEA-15 mutant cats were used for the study. Average EF in white and gray matter was 0.9±0.1 and 0.86±0.15 respectively, perfusion in white and gray matter was 85±15 mL/100g/min and 97±20 mL/100g/min respectively, PS in white and gray matter was 201±25 mL/100g/min and 225±35 mL/100g/min respectively for control cats. For PEA-15 mutant cats, average EF in white and gray matter was 0.81±0.15 and 0.77±0.2 respectively, perfusion in white and gray matter was 140±25 mL/100g/min and 165±18 mL/100g/min respectively, PS in white and gray matter was 240±30 mL/100g/min and 259±21 mL/100g/min respectively. This results show that BBB is compromised in PEA-15 mutant cat brain, where EF is decreased and perfusion as well as PS are increased in the mutant cats compared to the control cats. This findings might further explain the function of PEA-15 in neurodevelopment.Keywords: BBB, cat brain, magnetization transfer, PEA-15
Procedia PDF Downloads 143327 Leveraging xAPI in a Corporate e-Learning Environment to Facilitate the Tracking, Modelling, and Predictive Analysis of Learner Behaviour
Authors: Libor Zachoval, Daire O Broin, Oisin Cawley
Abstract:
E-learning platforms, such as Blackboard have two major shortcomings: limited data capture as a result of the limitations of SCORM (Shareable Content Object Reference Model), and lack of incorporation of Artificial Intelligence (AI) and machine learning algorithms which could lead to better course adaptations. With the recent development of Experience Application Programming Interface (xAPI), a large amount of additional types of data can be captured and that opens a window of possibilities from which online education can benefit. In a corporate setting, where companies invest billions on the learning and development of their employees, some learner behaviours can be troublesome for they can hinder the knowledge development of a learner. Behaviours that hinder the knowledge development also raise ambiguity about learner’s knowledge mastery, specifically those related to gaming the system. Furthermore, a company receives little benefit from their investment if employees are passing courses without possessing the required knowledge and potential compliance risks may arise. Using xAPI and rules derived from a state-of-the-art review, we identified three learner behaviours, primarily related to guessing, in a corporate compliance course. The identified behaviours are: trying each option for a question, specifically for multiple-choice questions; selecting a single option for all the questions on the test; and continuously repeating tests upon failing as opposed to going over the learning material. These behaviours were detected on learners who repeated the test at least 4 times before passing the course. These findings suggest that gauging the mastery of a learner from multiple-choice questions test scores alone is a naive approach. Thus, next steps will consider the incorporation of additional data points, knowledge estimation models to model knowledge mastery of a learner more accurately, and analysis of the data for correlations between knowledge development and identified learner behaviours. Additional work could explore how learner behaviours could be utilised to make changes to a course. For example, course content may require modifications (certain sections of learning material may be shown to not be helpful to many learners to master the learning outcomes aimed at) or course design (such as the type and duration of feedback).Keywords: artificial intelligence, corporate e-learning environment, knowledge maintenance, xAPI
Procedia PDF Downloads 121326 Using Rainfall Simulators to Design and Assess the Post-Mining Erosional Stability
Authors: Ashraf M. Khalifa, Hwat Bing So, Greg Maddocks
Abstract:
Changes to the mining environmental approvals process in Queensland have been rolled out under the MERFP Act (2018). This includes requirements for a Progressive Rehabilitation and Closure Plan (PRC Plan). Key considerations of the landform design report within the PRC Plan must include: (i) identification of materials available for landform rehabilitation, including their ability to achieve the required landform design outcomes, (ii) erosion assessments to determine landform heights, gradients, profiles, and material placement, (iii) slope profile design considering the interactions between soil erodibility, rainfall erosivity, landform height, gradient, and vegetation cover to identify acceptable erosion rates over a long-term average, (iv) an analysis of future stability based on the factors described above e.g., erosion and /or landform evolution modelling. ACARP funded an extensive and thorough erosion assessment program using rainfall simulators from 1998 to 2010. The ACARP program included laboratory assessment of 35 soil and spoil samples from 16 coal mines and samples from a gold mine in Queensland using 3 x 0.8 m laboratory rainfall simulator. The reliability of the laboratory rainfall simulator was verified through field measurements using larger flumes 20 x 5 meters and catchment scale measurements at three sites (3 different catchments, average area of 2.5 ha each). Soil cover systems are a primary component of a constructed mine landform. The primary functions of a soil cover system are to sustain vegetation and limit the infiltration of water and oxygen into underlying reactive mine waste. If the external surface of the landform erodes, the functions of the cover system cannot be maintained, and the cover system will most likely fail. Assessing a constructed landform’s potential ‘long-term’ erosion stability requires defensible erosion rate thresholds below which rehabilitation landform designs are considered acceptably erosion-resistant or ‘stable’. The process used to quantify erosion rates using rainfall simulators (flumes) to measure rill and inter-rill erosion on bulk samples under laboratory conditions or on in-situ material under field conditions will be explained.Keywords: open-cut, mining, erosion, rainfall simulator
Procedia PDF Downloads 101