Search results for: large storage
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 8684

Search results for: large storage

7814 Profile of Cross-Reactivity Allergens Highlighted by Multiplex Technology “Alex Microchip Technique” in the Diagnosis of Type I Hypersensitivity

Authors: Gadiri Sabiha

Abstract:

Introduction: Current allergy diagnostic tools using Multiplex technology have made it possible to increase the efficiency of the search for specific IgE. This opportunity is provided by the newly developed “Alex Biochip”, consisting of a panel of 282 allergens in native and molecular form, a CCD inhibitor, and the potential for detecting cross-reactive allergens. We evaluated the performance of this technology in detecting cross-reactivity in previously explored patients. Material/Method: The sera of 39 patients presenting sensitization and polysensitization profiles were explored. The search for specific IgE is carried out by the Alex ® IgE Biochip, and the results are analyzed by nature and by molecular family of allergens using specific software. Results/Discussion: The analysis gave a particular profile of cross-reactivity allergens: 33% for the Ole e1 family, 31% for NPC2, 26% for storage proteins, 20% for Tropomyosin, 10% for LTPs, 10% for Arginine Kinase and 10% for Uteroglobin CCDs were absent in all patients. The “Ole e1” allergen is responsible for a pollen-pollen cross allergy. The storage proteins found and LTP are not species-specific, causing cross-pollen-food allergy. The nDer p2 of the NPC2 family is responsible for cross-reactivity between mite species. Conclusion: The cross-reactivities responsible for mixed syndromes at diagnosis in our patients were dominated by pollen-pollen and pollen-food syndromes. They allow the identification of severity factors linked to the prognosis and the best-adapted immunotherapy.

Keywords: specific IgE, allergy, cross reactivity, molecular allergens

Procedia PDF Downloads 62
7813 The Effect of Organic Matter Maturation and Porosity Evolution on Methane Storage Potential in Shale-Gas Reservoirs

Authors: T. Topór, A. Derkowski, P. Ziemiański

Abstract:

Formation of organic matter (OM)-hosted nanopores upon thermal maturation are one of the key factor controlling methane storage potential in unconventional shale-gas reservoirs. In this study, the subcritical CO₂ and N₂ gas adsorption measurements combined with scanning electron microscopy and supercritical methane adsorption have been used to characterize pore system and methane storage potential in black shales from the Baltic Basin (Poland). The samples were collected from a virtually equivalent Llandovery strata across the basin and represent a complete digenetic sequence, from thermally immature to overmature. The results demonstrate that the thermal maturation is a dominant mechanism controlling the formation of OM micro- and mesopores in the Baltic Basin shales. The formation of micro- and mesopores occurs in the oil window (vitrinite reflectance; leavedVR; ~0.5-0.9%) as a result of oil expulsion from kerogenleft OM highly porous. The generated hydrocarbons then turn into solid bitumen causing pore blocking and substantial decrease in micro- and mesopore volume in late-mature shales (VR ~0.9-1.2%). Both micro- and mesopores are regenerated in a middle of the catagenesis range (VR 1.4-1.9%) due to secondary cracking of OM and gas formation. The micropore volume in investigated shales is almost exclusively controlled by the OM content. The contribution of clay minerals to micropore volume is insignificant and masked by a strong contribution from OM. Methane adsorption capacity in the Baltic Basin shales is predominantly controlled by microporous OM with pores < 1.5 nm. The mesopore volume (2-50 nm) and mesopore surface area have no effect on methane sorption behavior. The adsorbed methane density equivalent, calculated as absolute methane adsorption divided by micropore volume, reviled a decrease of the methane loading potential in micropores with increasing maturity. The highest methane loading potential in micropores is observed for OM before metagenesis (VR < 2%), where the adsorbed methane density equivalent is greater than the density of liquid methane. This implies that, in addition to physical adsorption, absorption of methane in OM may occur before metagenesis. After OM content reduction using NaOCl solution methane adoption capacity substantially decreases, suggesting significantly greater adsorption potential for OM microstructure than for the clay minerals matrix.

Keywords: maturation, methane sorption, organic matter, porosity, shales

Procedia PDF Downloads 233
7812 Rheological and Microstructural Characterization of Concentrated Emulsions Prepared by Fish Gelatin

Authors: Helen S. Joyner (Melito), Mohammad Anvari

Abstract:

Concentrated emulsions stabilized by proteins are systems of great importance in food, pharmaceutical and cosmetic products. Controlling emulsion rheology is critical for ensuring desired properties during formation, storage, and consumption of emulsion-based products. Studies on concentrated emulsions have focused on rheology of monodispersed systems. However, emulsions used for industrial applications are polydispersed in nature, and this polydispersity is regarded as an important parameter that also governs the rheology of the concentrated emulsions. Therefore, the objective of this study was to characterize rheological (small and large deformation behaviors) and microstructural properties of concentrated emulsions which were not truly monodispersed as usually encountered in food products such as margarines, mayonnaise, creams, spreads, and etc. The concentrated emulsions were prepared at different concentrations of fish gelatin (0.2, 0.4, 0.8% w/v in the whole emulsion system), oil-water ratio 80-20 (w/w), homogenization speed 10000 rpm, and 25oC. Confocal laser scanning microscopy (CLSM) was used to determine the microstructure of the emulsions. To prepare samples for CLSM analysis, FG solutions were stained by Fluorescein isothiocyanate dye. Emulsion viscosity profiles were determined using shear rate sweeps (0.01 to 100 1/s). The linear viscoelastic regions (LVRs) of the emulsions were determined using strain sweeps (0.01 to 100% strain) for each sample. Frequency sweeps were performed in the LVR (0.1% strain) from 0.6 to 100 rad/s. Large amplitude oscillatory shear (LAOS) testing was conducted by collecting raw waveform data at 0.05, 1, 10, and 100% strain at 4 different frequencies (0.5, 1, 10, and 100 rad/s). All measurements were performed in triplicate at 25oC. The CLSM results revealed that increased fish gelatin concentration resulted in more stable oil-in-water emulsions with homogeneous, finely dispersed oil droplets. Furthermore, the protein concentration had a significant effect on emulsion rheological properties. Apparent viscosity and dynamic moduli at small deformations increased with increasing fish gelatin concentration. These results were related to increased inter-droplet network connections caused by increased fish gelatin adsorption at the surface of oil droplets. Nevertheless, all samples showed shear-thinning and weak gel behaviors over shear rate and frequency sweeps, respectively. Lissajous plots, or plots of stress versus strain, and phase lag values were used to determine nonlinear behavior of the emulsions in LAOS testing. Greater distortion in the elliptical shape of the plots followed by higher phase lag values was observed at large strains and frequencies in all samples, indicating increased nonlinear behavior. Shifts from elastic-dominated to viscous dominated behavior were also observed. These shifts were attributed to damage to the sample microstructure (e.g. gel network disruption), which would lead to viscous-type behaviors such as permanent deformation and flow. Unlike the small deformation results, the LAOS behavior of the concentrated emulsions was not dependent on fish gelatin concentration. Systems with different microstructures showed similar nonlinear viscoelastic behaviors. The results of this study provided valuable information that can be used to incorporate concentrated emulsions in emulsion-based food formulations.

Keywords: concentrated emulsion, fish gelatin, microstructure, rheology

Procedia PDF Downloads 272
7811 Color Image Compression/Encryption/Contour Extraction using 3L-DWT and SSPCE Method

Authors: Ali A. Ukasha, Majdi F. Elbireki, Mohammad F. Abdullah

Abstract:

Data security needed in data transmission, storage, and communication to ensure the security. This paper is divided into two parts. This work interests with the color image which is decomposed into red, green and blue channels. The blue and green channels are compressed using 3-levels discrete wavelet transform. The Arnold transform uses to changes the locations of red image channel pixels as image scrambling process. Then all these channels are encrypted separately using the key image that has same original size and are generating using private keys and modulo operations. Performing the X-OR and modulo operations between the encrypted channels images for image pixel values change purpose. The extracted contours from color images recovery can be obtained with accepted level of distortion using single step parallel contour extraction (SSPCE) method. Experiments have demonstrated that proposed algorithm can fully encrypt 2D Color images and completely reconstructed without any distortion. Also shown that the analyzed algorithm has extremely large security against some attacks like salt and pepper and Jpeg compression. Its proof that the color images can be protected with a higher security level. The presented method has easy hardware implementation and suitable for multimedia protection in real time applications such as wireless networks and mobile phone services.

Keywords: SSPCE method, image compression and salt and peppers attacks, bitplanes decomposition, Arnold transform, color image, wavelet transform, lossless image encryption

Procedia PDF Downloads 515
7810 The Search of Anomalous Higgs Boson Couplings at the Large Hadron Electron Collider and Future Circular Electron Hadron Collider

Authors: Ilkay Turk Cakir, Murat Altinli, Zekeriya Uysal, Abdulkadir Senol, Olcay Bolukbasi Yalcinkaya, Ali Yilmaz

Abstract:

The Higgs boson was discovered by the ATLAS and CMS experimental groups in 2012 at the Large Hadron Collider (LHC). Production and decay properties of the Higgs boson, Standard Model (SM) couplings, and limits on effective scale of the Higgs boson’s couplings with other bosons are investigated at particle colliders. Deviations from SM estimates are parametrized by effective Lagrangian terms to investigate Higgs couplings. This is a model-independent method for describing the new physics. In this study, sensitivity to neutral gauge boson anomalous couplings with the Higgs boson is investigated using the parameters of the Large Hadron electron Collider (LHeC) and the Future Circular electron-hadron Collider (FCC-eh) with a model-independent approach. By using MadGraph5_aMC@NLO multi-purpose event generator with the parameters of LHeC and FCC-eh, the bounds on the anomalous Hγγ, HγZ and HZZ couplings in e− p → e− q H process are obtained. Detector simulations are also taken into account in the calculations.

Keywords: anomalos couplings, FCC-eh, Higgs, Z boson

Procedia PDF Downloads 205
7809 Thermodynamic and Spectroscopic Investigation of Binary 2,2-Dimethyl-1-Propanol+ CO₂ Gas Hydrates

Authors: Seokyoon Moon, Yun-Ho Ahn, Heejoong Kim, Sujin Hong, Yunseok Lee, Youngjune Park

Abstract:

Gas hydrate is a non-stoichiometric crystalline compound consisting of host water-framework and low molecular weight guest molecules. Small gaseous molecules such as CH₄, CO₂, and N₂ can be captured in the host water framework lattices of the gas hydrate with specific temperature and pressure conditions. The three well-known crystal structures of structure I (sI), structure II (sII), and structure H (sH) are determined by the size and shape of guest molecules. In this study, we measured the phase equilibria of binary (2,2-dimethyl-1-propanol + CO₂, CH₄, N₂) hydrates to explore their fundamental thermodynamic characteristics. We identified the structure of the binary gas hydrate by employing synchrotron high-resolution powder diffraction (HRPD), and the guest distributions in the lattice of gas hydrate were investigated via dispersive Raman and ¹³C solid-state nuclear magnetic resonance (NMR) spectroscopies. The end-to-end distance of 2,2-dimethyl-1-propanol was calculated to be 7.76 Å, which seems difficult to be enclathrated in large cages of sI or sII. However, due to the flexibility of the host water framework, binary hydrates of sI or sII types can be formed with the help of small gas molecule. Also, the synchrotron HRPD patterns revealed that the binary hydrate structure highly depends on the type of help gases; a cubic Fd3m sII hydrate was formed with CH₄ or N₂, and a cubic Pm3n sI hydrate was formed with CO₂. Interestingly, dispersive Raman and ¹³C NMR spectra showed that the unique tuning phenomenon occurred in binary (2,2-dimethyl-1-propanol + CO₂) hydrate. By optimizing the composition of NPA, we can achieve both thermodynamic stability and high CO₂ storage capacity for the practical application to CO₂ capture.

Keywords: clathrate, gas hydrate, neopentyl alcohol, CO₂, tuning phenomenon

Procedia PDF Downloads 237
7808 Shape Management Method of Large Structure Based on Octree Space Partitioning

Authors: Gichun Cha, Changgil Lee, Seunghee Park

Abstract:

The objective of the study is to construct the shape management method contributing to the safety of the large structure. In Korea, the research of the shape management is lack because of the new attempted technology. Terrestrial Laser Scanning (TLS) is used for measurements of large structures. TLS provides an efficient way to actively acquire accurate the point clouds of object surfaces or environments. The point clouds provide a basis for rapid modeling in the industrial automation, architecture, construction or maintenance of the civil infrastructures. TLS produce a huge amount of point clouds. Registration, Extraction and Visualization of data require the processing of a massive amount of scan data. The octree can be applied to the shape management of the large structure because the scan data is reduced in the size but, the data attributes are maintained. The octree space partitioning generates the voxel of 3D space, and the voxel is recursively subdivided into eight sub-voxels. The point cloud of scan data was converted to voxel and sampled. The experimental site is located at Sungkyunkwan University. The scanned structure is the steel-frame bridge. The used TLS is Leica ScanStation C10/C5. The scan data was condensed 92%, and the octree model was constructed with 2 millimeter in resolution. This study presents octree space partitioning for handling the point clouds. The basis is created by shape management of the large structures such as double-deck tunnel, building and bridge. The research will be expected to improve the efficiency of structural health monitoring and maintenance. "This work is financially supported by 'U-City Master and Doctor Course Grant Program' and the National Research Foundation of Korea(NRF) grant funded by the Korea government (MSIP) (NRF- 2015R1D1A1A01059291)."

Keywords: 3D scan data, octree space partitioning, shape management, structural health monitoring, terrestrial laser scanning

Procedia PDF Downloads 295
7807 Assessing the Effectiveness of Warehousing Facility Management: The Case of Mantrac Ghana Limited

Authors: Kuhorfah Emmanuel Mawuli

Abstract:

Generally, for firms to enhance their operational efficiency of logistics, it is imperative to assess the logistics function. The cost of logistics conventionally represents a key consideration in the pricing decisions of firms, which suggests that cost efficiency in logistics can go a long way to improve margins. Warehousing, which is a key part of logistics operations, has the prospect of influencing operational efficiency in logistics management as well as customer value, but this potential has often not been recognized. It has been found that there is a paucity of research that evaluates the efficiency of warehouses. Indeed, limited research has been conducted to examine potential barriers to effective warehousing management. Due to this paucity of research, there is limited knowledge on how to address the obstacles associated with warehousing management. In order for warehousing management to become profitable, there is the need to integrate, balance, and manage the economic inputs and outputs of the entire warehouse operations, something that many firms tend to ignore. Management of warehousing is not solely related to storage functions. Instead, effective warehousing management requires such practices as maximum possible mechanization and automation of operations, optimal use of space and capacity of storage facilities, organization through "continuous flow" of goods, a planned system of storage operations, and safety of goods. For example, there is an important need for space utilization of the warehouse surface as it is a good way to evaluate the storing operation and pick items per hour. In the setting of Mantrac Ghana, not much knowledge regarding the management of the warehouses exists. The researcher has personally observed many gaps in the management of the warehouse facilities in the case organization Mantrac Ghana. It is important, therefore, to assess the warehouse facility management of the case company with the objective of identifying weaknesses for improvement. The study employs an in-depth qualitative research approach using interviews as a mode of data collection. Respondents in the study mainly comprised warehouse facility managers in the studied company. A total of 10 participants were selected for the study using a purposive sampling strategy. Results emanating from the study demonstrate limited warehousing effectiveness in the case company. Findings further reveal that the major barriers to effective warehousing facility management comprise poor layout, poor picking optimization, labour costs, and inaccurate orders; policy implications of the study findings are finally outlined.

Keywords: assessing, warehousing, facility, management

Procedia PDF Downloads 60
7806 Cataract Surgery and Sustainability: Comparative Study of Single-Use Versus Reusable Cassettes in Phacoemulsification

Authors: Oscar Kallay

Abstract:

Objective: This study compares the sustainability, financial implications, and surgical efficiency of two phacoemulsification cassette systems for cataract surgery: a machine with single-use cassettes and another with daily, reusable ones. Methods: The observational study involves retrospective cataract surgery data collection at the Centre Médical de l'Alliance, Braine-L’alleud, Belgium, a tertiary eye care center. Information on cassette weight, quantities, and transport volume was obtained from routine procedures and purchasing records. The costs for each machine were calculated by reviewing the invoices received from the accounting department. Results: We found significant differences across comparisons. The reusable cassette machine, when compared to the single-use machine, used 306.7 kg less plastic (75.3% reduction), required 2,494 cubic meters less storage per 1000 surgeries (67.7% decrease), and cost €54.16 less per 10 procedures (16.9% reduction). The machine with daily reusable cassettes also exhibited a 7-minute priming time advantage for 10 procedures, reducing downtime between cases. Conclusions: Our findings underscore the benefits of adopting reusable cassette systems: reduced plastic consumption, storage volume, and priming time, as well as enhanced efficiency and cost savings. Healthcare professionals and institutions are encouraged to embrace environmentally conscious initiatives. The use of reusable cassette systems for cataract surgeries offers a pathway to sustainable practices.

Keywords: cataract, epidemiolog, surgery treatment, lens and zonules, public health

Procedia PDF Downloads 13
7805 Molecular Dynamics Simulation of the Effect of the Solid Gas Interface Nanolayer on Enhanced Thermal Conductivity of Copper-CO2 Nanofluid

Authors: Zeeshan Ahmed, Ajinkya Sarode, Pratik Basarkar, Atul Bhargav, Debjyoti Banerjee

Abstract:

The use of CO2 in oil recovery and in CO2 capture and storage is gaining traction in recent years. These applications involve heat transfer between CO2 and the base fluid, and hence, there arises a need to improve the thermal conductivity of CO2 to increase the process efficiency and reduce cost. One way to improve the thermal conductivity is through nanoparticle addition in the base fluid. The nanofluid model in this study consisted of copper (Cu) nanoparticles in varying concentrations with CO2 as a base fluid. No experimental data are available on thermal conductivity of CO2 based nanofluid. Molecular dynamics (MD) simulations are an increasingly adopted tool to perform preliminary assessments of nanoparticle (NP) fluid interactions. In this study, the effect of the formation of a nanolayer (or molecular layering) at the gas-solid interface on thermal conductivity is investigated using equilibrium MD simulations by varying NP diameter and keeping the volume fraction (1.413%) of nanofluid constant to check the diameter effect of NP on the nanolayer and thermal conductivity. A dense semi-solid fluid layer was seen to be formed at the NP-gas interface, and the thickness increases with increase in particle diameter, which also moves with the NP Brownian motion. Density distribution has been done to see the effect of nanolayer, and its thickness around the NP. These findings are extremely beneficial, especially to industries employed in oil recovery as increased thermal conductivity of CO2 will lead to enhanced oil recovery and thermal energy storage.

Keywords: copper-CO2 nanofluid, molecular dynamics simulation, molecular interfacial layer, thermal conductivity

Procedia PDF Downloads 330
7804 The Utilization of Salicylic Acid of the Extract from Avocado Skin as an Inhibitor of Ethylene Production to Keep the Quality of Banana in Storage

Authors: Adira Nofeadri Ryofi, Alvin Andrianus, Anna Khairunnisa, Anugrah Cahyo Widodo, Arbhyando Tri Putrananda, Arsy Imanda N. Raswati, Gita Rahmaningsih, Ina Agustina

Abstract:

The consumption level of fresh bananas from 2005 until 2010, increased from 8.2 to 10 kg/capita/year. The commercial scale of banana generally harvested when it still green to make the banana avoid physical damage, chemical, and disease after harvest and ripe fruit. That first metabolism activity can be used as a synthesis reaction. Ripening fruit was influenced by ethylene hormone that synthesized in fruit which is experiencing ripe and including hormone in the ripening fruit process in klimaterik phase. This ethylene hormone is affected by the respiration level that would speed up the restructuring of carbohydrates inside the fruit, so the weighting of fruit will be decreased. Compared to other klimaterik fruit, banana is a fruit that has a medium ethylene production rate and the rate of respiration is low. The salicylic acid can regulate the result number of the growth process or the development of fruits and plants. Salicylic acid serves to hinder biosynthesis ethylene and delay senses. The research aims to understand the influence of salicylic acid concentration that derived from the waste of avocado skin in inhibition process to ethylene production that the maturation can be controlled, so it can keep the quality of banana for storage. It is also to increase the potential value of the waste of avocado skin that were still used in industrial cosmetics.

Keywords: ethylene hormone, extract avocado skin, inhibitor, salicylic acid

Procedia PDF Downloads 234
7803 Influence of the Substitution of C for Mg and Ni on the Microstructure and Hydrogen Storage Characteristics of Mg2Ni Alloys

Authors: Sajad Haghanifar, Seyed-Farshid Kashani Bozorg

Abstract:

Nano-crystalline Mg2Ni-based powder was produced by mechanical alloying technique using binary and ternary powder mixtures with stoichiometric compositions of Mg2Ni, Mg1.9C0.1Ni and Mg2C0.1Ni0.9. The structures and morphologies of the milled products were studied by XRD, SEM and HRTEM. Their electrochemical hydrogen storage characteristics were investigated in 6 M KOH solution. X-Ray diffraction, scanning and transmission electron microscopy of the milled products showed the formation of Mg2Ni-based nano-crystallites after 5, 15 and 30 h of milling using the initial powder mixtures of Mg1.9C0.1Ni, Mg2Ni and Mg2C0.1Ni0.9, respectively. It was found that partial substitution of C for Mg has beneficial effect on the formation kinetic of nano-crystalline Mg2Ni. Contrary to this, partial substitution of C for Ni was resulted in retardation of formation kinetic of nano-crystalline Mg2Ni. In addition, the negative electrode made from Mg1.9C0.1Ni ternary milled product after 30 hour of milling exhibited the highest initial discharge capacity and longest discharge life. Thus, partial substitution of C for Mg is beneficial to electrode properties of the Mg2Ni-based crystallites. The relation between the discharge capacity and cycling number of mechanically alloyed products was proposed on the basis of the fact that the degradation of discharge capacity was mainly caused by the oxidation of magnesium and nickel. The experimental data fitted the deduced equation well.

Keywords: Mg2Ni, hydrogen absorbing materials, electrochemical properties, nano-crystalline, amorphous, mechanical alloying, carbon

Procedia PDF Downloads 429
7802 Dispersions of Carbon Black in Microemulsions

Authors: Mohamed Youssry, Dominique Guyomard, Bernard Lestriez

Abstract:

In order to enhance the energy and power densities of electrodes for energy storage systems, the formulation and processing of electrode slurries proved to be a critical issue in determining the electrode performance. In this study, we introduce novel approach to formulate carbon black slurries based on microemulsion and lyotropic liquid crystalline phases (namely, lamellar phase) composed of non-ionic surfactant (Triton X100), decanol and water. Simultaneous measurements of electrical properties of slurries under shear flow (rheology) have been conducted to elucidate the microstructure evolution with the surfactant concentration and decanol/water ratio at rest, as well as, the structural transition under steady-shear which has been confirmed by rheo-microscopy. Interestingly, the carbon black slurries at low decanol/water ratio are weak-gel (flowable) with higher electrical conductivity than those at higher ratio which behave strong-gel viscoelastic response. In addition, the slurries show recoverable electrical behaviour under shear flow in tandem with the viscosity trend. It is likely that oil-in-water microemulsion enhances slurries’ stability without affecting on the percolating network of carbon black. On the other hand, the oil-in-water analogous and bilayer structure of lamellar phase cause the slurries less conductive as a consequence of losing the network percolation. These findings are encouraging to formulate microemulsion-based electrodes for energy storage system (lithium-ion batteries).

Keywords: electrode slurries, microemulsion, microstructure transition, rheo-electrical properties

Procedia PDF Downloads 261
7801 Load Balancing Technique for Energy - Efficiency in Cloud Computing

Authors: Rani Danavath, V. B. Narsimha

Abstract:

Cloud computing is emerging as a new paradigm of large scale distributed computing. Cloud computing is a model for enabling ubiquitous, convenient, on-demand network access to a shared pool of configurable computing resources (e.g., three service models, and four deployment networks, servers, storage, applications, and services) that can be rapidly provisioned and released with minimal management effort or service provider interaction. This cloud model is composed of five essential characteristics models. Load balancing is one of the main challenges in cloud computing, which is required to distribute the dynamic workload across multiple nodes, to ensure that no single node is overloaded. It helps in optimal utilization of resources, enhancing the performance of the system. The goal of the load balancing is to minimize the resource consumption and carbon emission rate, that is the direct need of cloud computing. This determined the need of new metrics energy consumption and carbon emission for energy-efficiency load balancing techniques in cloud computing. Existing load balancing techniques mainly focuses on reducing overhead, services, response time and improving performance etc. In this paper we introduced a Technique for energy-efficiency, but none of the techniques have considered the energy consumption and carbon emission. Therefore, our proposed work will go towards energy – efficiency. So this energy-efficiency load balancing technique can be used to improve the performance of cloud computing by balancing the workload across all the nodes in the cloud with the minimum resource utilization, in turn, reducing energy consumption, and carbon emission to an extent, which will help to achieve green computing.

Keywords: cloud computing, distributed computing, energy efficiency, green computing, load balancing, energy consumption, carbon emission

Procedia PDF Downloads 445
7800 Imputation of Incomplete Large-Scale Monitoring Count Data via Penalized Estimation

Authors: Mohamed Dakki, Genevieve Robin, Marie Suet, Abdeljebbar Qninba, Mohamed A. El Agbani, Asmâa Ouassou, Rhimou El Hamoumi, Hichem Azafzaf, Sami Rebah, Claudia Feltrup-Azafzaf, Nafouel Hamouda, Wed a.L. Ibrahim, Hosni H. Asran, Amr A. Elhady, Haitham Ibrahim, Khaled Etayeb, Essam Bouras, Almokhtar Saied, Ashrof Glidan, Bakar M. Habib, Mohamed S. Sayoud, Nadjiba Bendjedda, Laura Dami, Clemence Deschamps, Elie Gaget, Jean-Yves Mondain-Monval, Pierre Defos Du Rau

Abstract:

In biodiversity monitoring, large datasets are becoming more and more widely available and are increasingly used globally to estimate species trends and con- servation status. These large-scale datasets challenge existing statistical analysis methods, many of which are not adapted to their size, incompleteness and heterogeneity. The development of scalable methods to impute missing data in incomplete large-scale monitoring datasets is crucial to balance sampling in time or space and thus better inform conservation policies. We developed a new method based on penalized Poisson models to impute and analyse incomplete monitoring data in a large-scale framework. The method al- lows parameterization of (a) space and time factors, (b) the main effects of predic- tor covariates, as well as (c) space–time interactions. It also benefits from robust statistical and computational capability in large-scale settings. The method was tested extensively on both simulated and real-life waterbird data, with the findings revealing that it outperforms six existing methods in terms of missing data imputation errors. Applying the method to 16 waterbird species, we estimated their long-term trends for the first time at the entire North African scale, a region where monitoring data suffer from many gaps in space and time series. This new approach opens promising perspectives to increase the accuracy of species-abundance trend estimations. We made it freely available in the r package ‘lori’ (https://CRAN.R-project.org/package=lori) and recommend its use for large- scale count data, particularly in citizen science monitoring programmes.

Keywords: biodiversity monitoring, high-dimensional statistics, incomplete count data, missing data imputation, waterbird trends in North-Africa

Procedia PDF Downloads 150
7799 Using the Semantic Web Technologies to Bring Adaptability in E-Learning Systems

Authors: Fatima Faiza Ahmed, Syed Farrukh Hussain

Abstract:

The last few decades have seen a large proportion of our population bending towards e-learning technologies, starting from learning tools used in primary and elementary schools to competency based e-learning systems specifically designed for applications like finance and marketing. The huge diversity in this crowd brings about a large number of challenges for the designers of these e-learning systems, one of which is the adaptability of such systems. This paper focuses on adaptability in the learning material in an e-learning course and how artificial intelligence and the semantic web can be used as an effective tool for this purpose. The study proved that the semantic web, still a hot topic in the area of computer science can prove to be a powerful tool in designing and implementing adaptable e-learning systems.

Keywords: adaptable e-learning, HTMLParser, information extraction, semantic web

Procedia PDF Downloads 323
7798 Technical and Economic Potential of Partial Electrification of Railway Lines

Authors: Rafael Martins Manzano Silva, Jean-Francois Tremong

Abstract:

Electrification of railway lines allows to increase speed, power, capacity and energetic efficiency of rolling stocks. However, this process of electrification is complex and costly. An electrification project is not just about design of catenary. It also includes installation of structures around electrification, as substation installation, electrical isolation, signalling, telecommunication and civil engineering structures. France has more than 30,000 km of railways, whose only 53% are electrified. The others 47% of railways use diesel locomotive and represent only 10% of the circulation (tons.km). For this reason, a new type of electrification, less expensive than the usual, is requested to enable the modernization of these railways. One solution could be the use of hybrids trains. This technology opens up new opportunities for less expensive infrastructure development such as the partial electrification of railway lines. In a partially electrified railway, the power supply of theses hybrid trains could be made either by the catenary or by the on-board energy storage system (ESS). Thus, the on-board ESS would feed the energetic needs of the train along the non-electrified zones while in electrified zones, the catenary would feed the train and recharge the on-board ESS. This paper’s objective deals with the technical and economic potential identification of partial electrification of railway lines. This study provides different scenarios of electrification by replacing the most expensive places to electrify using on-board ESS. The target is to reduce the cost of new electrification projects, i.e. reduce the cost of electrification infrastructures while not increasing the cost of rolling stocks. In this study, scenarios are constructed in function of the electrification’s cost of each structure. The electrification’s cost varies considerably because of the installation of catenary support in tunnels, bridges and viaducts is much more expensive than in others zones of the railway. These scenarios will be used to describe the power supply system and to choose between the catenary and the on-board energy storage depending on the position of the train on the railway. To identify the influence of each partial electrification scenario in the sizing of the on-board ESS, a model of the railway line and of the rolling stock is developed for a real case. This real case concerns a railway line located in the south of France. The energy consumption and the power demanded at each point of the line for each power supply (catenary or on-board ESS) are provided at the end of the simulation. Finally, the cost of a partial electrification is obtained by adding the civil engineering costs of the zones to be electrified plus the cost of the on-board ESS. The study of the technical and economic potential ends with the identification of the most economically interesting scenario of electrification.

Keywords: electrification, hybrid, railway, storage

Procedia PDF Downloads 425
7797 Exploring the Sources of Innovation in Food Processing SMEs of Kerala

Authors: Bhumika Gupta, Jeayaram Subramanian, Hardik Vachhrajani, Avinash Shivdas

Abstract:

Indian food processing industry is one of the largest in the world in terms of production, consumption, exports and growth opportunities. SMEs play a crucial role within this. Large manufacturing firms largely dominate innovation studies in India. Innovation sources used by SMEs are often different from that of large firms. This paper focuses on exploring various sources of innovation adopted by food processing SMEs in Kerala, South India. Outcome suggests that SMEs use various sources like suppliers, competitors, employees, government/research institutions and customers to get new ideas.

Keywords: food processing, innovation, SMEs, sources of innovation

Procedia PDF Downloads 410
7796 Modeling of Virtual Power Plant

Authors: Muhammad Fanseem E. M., Rama Satya Satish Kumar, Indrajeet Bhausaheb Bhavar, Deepak M.

Abstract:

Keeping the right balance of electricity between the supply and demand sides of the grid is one of the most important objectives of electrical grid operation. Power generation and demand forecasting are the core of power management and generation scheduling. Large, centralized producing units were used in the construction of conventional power systems in the past. A certain level of balance was possible since the generation kept up with the power demand. However, integrating renewable energy sources into power networks has proven to be a difficult challenge due to its intermittent nature. The power imbalance caused by rising demands and peak loads is negatively affecting power quality and dependability. Demand side management and demand response were one of the solutions, keeping generation the same but altering or rescheduling or shedding completely the load or demand. However, shedding the load or rescheduling is not an efficient way. There comes the significance of virtual power plants. The virtual power plant integrates distributed generation, dispatchable load, and distributed energy storage organically by using complementing control approaches and communication technologies. This would eventually increase the utilization rate and financial advantages of distributed energy resources. Most of the writing on virtual power plant models ignored technical limitations, and modeling was done in favor of a financial or commercial viewpoint. Therefore, this paper aims to address the modeling intricacies of VPPs and their technical limitations, shedding light on a holistic understanding of this innovative power management approach.

Keywords: cost optimization, distributed energy resources, dynamic modeling, model quality tests, power system modeling

Procedia PDF Downloads 57
7795 Methods Used to Achieve Airtightness of 0.07 Ach@50Pa for an Industrial Building

Authors: G. Wimmers

Abstract:

The University of Northern British Columbia needed a new laboratory building for the Master of Engineering in Integrated Wood Design Program and its new Civil Engineering Program. Since the University is committed to reducing its environmental footprint and because the Master of Engineering Program is actively involved in research of energy efficient buildings, the decision was made to request the energy efficiency of the Passive House Standard in the Request for Proposals. The building is located in Prince George in Northern British Columbia, a city located at the northern edge of climate zone 6 with an average low between -8 and -10.5 in the winter months. The footprint of the building is 30m x 30m with a height of about 10m. The building consists of a large open space for the shop and laboratory with a small portion of the floorplan being two floors, allowing for a mezzanine level with a few offices as well as mechanical and storage rooms. The total net floor area is 1042m² and the building’s gross volume 9686m³. One key requirement of the Passive House Standard is the airtight envelope with an airtightness of < 0.6 ach@50Pa. In the past, we have seen that this requirement can be challenging to reach for industrial buildings. When testing for air tightness, it is important to test in both directions, pressurization, and depressurization, since the airflow through all leakages of the building will, in reality, happen simultaneously in both directions. A specific detail or situation such as overlapping but not sealed membranes might be airtight in one direction, due to the valve effect, but are opening up when tested in the opposite direction. In this specific project, the advantage was the overall very compact envelope and the good volume to envelope area ratio. The building had to be very airtight and the details for the windows and doors installation as well as all transitions from walls to roof and floor, the connections of the prefabricated wall panels and all penetrations had to be carefully developed to allow for maximum airtightness. The biggest challenges were the specific components of this industrial building, the large bay door for semi-trucks and the dust extraction system for the wood processing machinery. The testing was carried out in accordance with EN 132829 (method A) as specified in the International Passive House Standard and the volume calculation was also following the Passive House guideline resulting in a net volume of 7383m3, excluding all walls, floors and suspended ceiling volumes. This paper will explore the details and strategies used to achieve an airtightness of 0.07 ach@50Pa, to the best of our knowledge the lowest value achieved in North America so far following the test protocol of the International Passive House Standard and discuss the crucial steps throughout the project phases and the most challenging details.

Keywords: air changes, airtightness, envelope design, industrial building, passive house

Procedia PDF Downloads 146
7794 An Efficient Algorithm for Solving the Transmission Network Expansion Planning Problem Integrating Machine Learning with Mathematical Decomposition

Authors: Pablo Oteiza, Ricardo Alvarez, Mehrdad Pirnia, Fuat Can

Abstract:

To effectively combat climate change, many countries around the world have committed to a decarbonisation of their electricity, along with promoting a large-scale integration of renewable energy sources (RES). While this trend represents a unique opportunity to effectively combat climate change, achieving a sound and cost-efficient energy transition towards low-carbon power systems poses significant challenges for the multi-year Transmission Network Expansion Planning (TNEP) problem. The objective of the multi-year TNEP is to determine the necessary network infrastructure to supply the projected demand in a cost-efficient way, considering the evolution of the new generation mix, including the integration of RES. The rapid integration of large-scale RES increases the variability and uncertainty in the power system operation, which in turn increases short-term flexibility requirements. To meet these requirements, flexible generating technologies such as energy storage systems must be considered within the TNEP as well, along with proper models for capturing the operational challenges of future power systems. As a consequence, TNEP formulations are becoming more complex and difficult to solve, especially for its application in realistic-sized power system models. To meet these challenges, there is an increasing need for developing efficient algorithms capable of solving the TNEP problem with reasonable computational time and resources. In this regard, a promising research area is the use of artificial intelligence (AI) techniques for solving large-scale mixed-integer optimization problems, such as the TNEP. In particular, the use of AI along with mathematical optimization strategies based on decomposition has shown great potential. In this context, this paper presents an efficient algorithm for solving the multi-year TNEP problem. The algorithm combines AI techniques with Column Generation, a traditional decomposition-based mathematical optimization method. One of the challenges of using Column Generation for solving the TNEP problem is that the subproblems are of mixed-integer nature, and therefore solving them requires significant amounts of time and resources. Hence, in this proposal we solve a linearly relaxed version of the subproblems, and trained a binary classifier that determines the value of the binary variables, based on the results obtained from the linearized version. A key feature of the proposal is that we integrate the binary classifier into the optimization algorithm in such a way that the optimality of the solution can be guaranteed. The results of a study case based on the HRP 38-bus test system shows that the binary classifier has an accuracy above 97% for estimating the value of the binary variables. Since the linearly relaxed version of the subproblems can be solved with significantly less time than the integer programming counterpart, the integration of the binary classifier into the Column Generation algorithm allowed us to reduce the computational time required for solving the problem by 50%. The final version of this paper will contain a detailed description of the proposed algorithm, the AI-based binary classifier technique and its integration into the CG algorithm. To demonstrate the capabilities of the proposal, we evaluate the algorithm in case studies with different scenarios, as well as in other power system models.

Keywords: integer optimization, machine learning, mathematical decomposition, transmission planning

Procedia PDF Downloads 82
7793 Hyperspectral Image Classification Using Tree Search Algorithm

Authors: Shreya Pare, Parvin Akhter

Abstract:

Remotely sensing image classification becomes a very challenging task owing to the high dimensionality of hyperspectral images. The pixel-wise classification methods fail to take the spatial structure information of an image. Therefore, to improve the performance of classification, spatial information can be integrated into the classification process. In this paper, the multilevel thresholding algorithm based on a modified fuzzy entropy function is used to perform the segmentation of hyperspectral images. The fuzzy parameters of the MFE function have been optimized by using a new meta-heuristic algorithm based on the Tree-Search algorithm. The segmented image is classified by a large distribution machine (LDM) classifier. Experimental results are shown on a hyperspectral image dataset. The experimental outputs indicate that the proposed technique (MFE-TSA-LDM) achieves much higher classification accuracy for hyperspectral images when compared to state-of-art classification techniques. The proposed algorithm provides accurate segmentation and classification maps, thus becoming more suitable for image classification with large spatial structures.

Keywords: classification, hyperspectral images, large distribution margin, modified fuzzy entropy function, multilevel thresholding, tree search algorithm, hyperspectral image classification using tree search algorithm

Procedia PDF Downloads 173
7792 Synthesis of Nickel-Platinum Composite Nanoparticles and Silica-Coating of Them

Authors: Masato Yanase, Noriko Yamauchi, Shohei Tada, Yoshio Kobayashi

Abstract:

Nickel (Ni) and Ni-containing nanoparticles are promising materials due to their unique magnetic properties. In a preliminary experiment, aggregates of such nanoparticles formed after they were synthesized. Since the nanoparticle aggregation may deteriorate their unique properties, a method to suppress the nanoparticle aggregation has been required. In the present work, compositing with nickel (Ni) and platinum (Pt) and silica-coating of them were examined for suppression. Ni-Pt nanoparticles were synthesized in water exposed to air, in which nickel (II) acetate tetrahydrate, hexachloroplatinate (IV) hexahydrate, and sodium borohydride were used as a Ni source, a Pt source, and a reducing reagent, respectively. Polyvinylpyrrolidone, poly (sodium 4-styrene sulfonate), and citric acid were used as the stabilizers. Silica-coating of Ni-Pt nanoparticles was performed by adding tetraethylorthosilicate(TEOS)/ethanol solution to the Ni-Pt nanoparticle colloidal solution (Ni-Pt/SiO₂). The morphology of Ni-Pt nanoparticles was dependent on the reaction time and the species of stabilizer. The Ni-Pt/SiO₂ nanoparticles were composed of Ni-Pt nanoparticles as core and SiO₂ as shell, and their morphology depended on the TEOS concentrations. Furthermore, the Ni-Pt/SiO₂ nanoparticles were more dispersed compared to uncoated Ni-Pt nanoparticles. This suggested that the silica-coating had an effect on controlling the aggregation. The liquid-phase synthesis process involving the sol-gel method used in this study is advantageous for achieving monodispersion of particles and high-purity of materials. If the challenges of optimizing reaction conditions to achieve them during scale-up can be addressed, the proposed method holds great potential for large-scale production of particles in applications such as magnetic storage devices and biomedical imaging.

Keywords: metal, composite, nickel, platinum, nanoparticle, silica-coating

Procedia PDF Downloads 22
7791 Quality and Shelf life of UHT Milk Produced in Tripoli, Libya

Authors: Faozia A. S. Abuhtana, Yahia S. Abujnah, Said O. Gnann

Abstract:

Ultra High Temperature (UHT) processed milk is widely distributed and preferred in numerous countries all over the world due its relatively high quality and long shelf life. Because of the notable high consumption rate of UHT in Libya in addition to negligible studies related to such product on the local level, this study was designed to assess the shelf life of locally produced as well as imported reconstituted sterilized whole milk samples marketed in Tripoli, Libya . Four locally produced vs. three imported brands were used in this study. All samples were stored at room temperature (25± 2C ) for 8 month long period, and subjected to physical, chemical, microbiological and sensory tests. These tests included : measurement of pH, specific gravity, percent acidity, and determination of fat, protein and melamine content. Microbiological tests included total aerobic count, total psychotropic bacteria, total spore forming bacteria and total coliform counts. Results indicated no detection of microbial growth of any type during the study period, in addition to no detection of melamine in all samples. On the other hand, a gradual decline in pH accompanied with gradual increase in % acidity of both locally produced and imported samples was observed. Such changes in both pH and % acidity reached their lowest and highest values respectively during the 24th week of storage. For instance pH values were (6.40, 6.55, 6.55, 6.15) and (6.30, 6.50, 6.20) for local and imported brands respectively. On the other hand, % acidity reached (0.185, 0181, 0170, 0183) and (0180, 0.180, 0.171) at the 24th week for local and imported brands respectively. Similar pattern of decline was also observed in specific gravity, fat and protein content in some local and imported samples especially at later stages of the study. In both cases, some of the recorded pH values, % acidity, sp. gravity and fat content were in violation of the accepted limits set by Libyan standard no. 356 for sterilized milk. Such changes in pH, % acidity and other UHT sterilized milk constituents during storage were coincided with a gradual decrease in the degree of acceptance of the stored milk samples of both types as shown by sensory scores recorded by the panelists. In either case degree of acceptance was significantly low at late stages of storage and most milk samples became relatively unacceptable after the 18th and 20th week for both untrained and trained panelists respectively.

Keywords: UHT milk, shelf life, quality, gravity, bacteria

Procedia PDF Downloads 333
7790 Steepest Descent Method with New Step Sizes

Authors: Bib Paruhum Silalahi, Djihad Wungguli, Sugi Guritman

Abstract:

Steepest descent method is a simple gradient method for optimization. This method has a slow convergence in heading to the optimal solution, which occurs because of the zigzag form of the steps. Barzilai and Borwein modified this algorithm so that it performs well for problems with large dimensions. Barzilai and Borwein method results have sparked a lot of research on the method of steepest descent, including alternate minimization gradient method and Yuan method. Inspired by previous works, we modified the step size of the steepest descent method. We then compare the modification results against the Barzilai and Borwein method, alternate minimization gradient method and Yuan method for quadratic function cases in terms of the iterations number and the running time. The average results indicate that the steepest descent method with the new step sizes provide good results for small dimensions and able to compete with the results of Barzilai and Borwein method and the alternate minimization gradient method for large dimensions. The new step sizes have faster convergence compared to the other methods, especially for cases with large dimensions.

Keywords: steepest descent, line search, iteration, running time, unconstrained optimization, convergence

Procedia PDF Downloads 538
7789 Freshwater Lens Observation: Case Study of Laura Island, Majuro Atoll, Republic of the Marshall Islands

Authors: Kazuhisa Koda, Tsutomu Kobayashi, Rebecca Lorennji, Alington Robert, Halston DeBrum, Julious Lucky, Paul Paul

Abstract:

Atolls are low-lying small islands with highly permeable ground that does not allow rivers and lakes to develop. As the water resources on these atolls basically rely on precipitation, groundwater becomes a very important water resource during droughts. Freshwater lenses develop as groundwater on relatively large atoll islands and play a key role in the stable water supply. Atoll islands in the Pacific Ocean sometimes suffer from drought due to El Nino. The global warming effects are noticeable, particularly on atoll islands. The Republic of the Marshall Islands in Oceania is burdened with the problems common to atoll islands. About half of its population lives in the capital, Majuro, and securing water resources for these people is a crucial issue. There is a freshwater lens on the largest, Laura Island, which serves as a water source for the downtown area. A serious drought that occurred in 1998 resulted in excessive water intake from the freshwater lens on Laura Island causing up-coning. Up-coning mixes saltwater into groundwater pumped from water-intake wells. Because up-coning makes the freshwater lens unusable, there was a need to investigate the freshwater lens on Laura Island. In this study, we observed the electrical conductivities of the groundwater at different depths in existing monitoring wells to determine the total storage volume of the freshwater lens on Laura Island from 2010 to 2013. Our results indicated that most of the groundwater that seeped into the freshwater lens had flowed out into the sea.

Keywords: Atoll islands, drought, El-Nino, freshwater lens, groundwater observation

Procedia PDF Downloads 315
7788 Efficient Implementation of Finite Volume Multi-Resolution Weno Scheme on Adaptive Cartesian Grids

Authors: Yuchen Yang, Zhenming Wang, Jun Zhu, Ning Zhao

Abstract:

An easy-to-implement and robust finite volume multi-resolution Weighted Essentially Non-Oscillatory (WENO) scheme is proposed on adaptive cartesian grids in this paper. Such a multi-resolution WENO scheme is combined with the ghost cell immersed boundary method (IBM) and wall-function technique to solve Navier-Stokes equations. Unlike the k-exact finite volume WENO schemes which involve large amounts of extra storage, repeatedly solving the matrix generated in a least-square method or the process of calculating optimal linear weights on adaptive cartesian grids, the present methodology only adds very small overhead and can be easily implemented in existing edge-based computational fluid dynamics (CFD) codes with minor modifications. Also, the linear weights of this adaptive finite volume multi-resolution WENO scheme can be any positive numbers on condition that their sum is one. It is a way of bypassing the calculation of the optimal linear weights and such a multi-resolution WENO scheme avoids dealing with the negative linear weights on adaptive cartesian grids. Some benchmark viscous problems are numerical solved to show the efficiency and good performance of this adaptive multi-resolution WENO scheme. Compared with a second-order edge-based method, the presented method can be implemented into an adaptive cartesian grid with slight modification for big Reynolds number problems.

Keywords: adaptive mesh refinement method, finite volume multi-resolution WENO scheme, immersed boundary method, wall-function technique.

Procedia PDF Downloads 144
7787 Graph-Oriented Summary for Optimized Resource Description Framework Graphs Streams Processing

Authors: Amadou Fall Dia, Maurras Ulbricht Togbe, Aliou Boly, Zakia Kazi Aoul, Elisabeth Metais

Abstract:

Existing RDF (Resource Description Framework) Stream Processing (RSP) systems allow continuous processing of RDF data issued from different application domains such as weather station measuring phenomena, geolocation, IoT applications, drinking water distribution management, and so on. However, processing window phase often expires before finishing the entire session and RSP systems immediately delete data streams after each processed window. Such mechanism does not allow optimized exploitation of the RDF data streams as the most relevant and pertinent information of the data is often not used in a due time and almost impossible to be exploited for further analyzes. It should be better to keep the most informative part of data within streams while minimizing the memory storage space. In this work, we propose an RDF graph summarization system based on an explicit and implicit expressed needs through three main approaches: (1) an approach for user queries (SPARQL) in order to extract their needs and group them into a more global query, (2) an extension of the closeness centrality measure issued from Social Network Analysis (SNA) to determine the most informative parts of the graph and (3) an RDF graph summarization technique combining extracted user query needs and the extended centrality measure. Experiments and evaluations show efficient results in terms of memory space storage and the most expected approximate query results on summarized graphs compared to the source ones.

Keywords: centrality measures, RDF graphs summary, RDF graphs stream, SPARQL query

Procedia PDF Downloads 196
7786 Restored CO₂ from Flue Gas and Utilization by Converting to Methanol by 3 Step Processes: Steam Reforming, Reverse Water Gas Shift and Hydrogenation

Authors: Rujira Jitrwung, Kuntima Krekkeitsakul, Weerawat Patthaveekongka, Chiraphat Kumpidet, Jarukit Tepkeaw, Krissana Jaikengdee, Anantachai Wannajampa

Abstract:

Flue gas discharging from coal fired or gas combustion power plant contains around 12% Carbon dioxide (CO₂), 6% Oxygen (O₂), and 82% Nitrogen (N₂).CO₂ is a greenhouse gas which has been concerned to the global warming. Carbon Capture, Utilization, and Storage (CCUS) is a topic which is a tool to deal with this CO₂ realization. Flue gas is drawn down from the chimney and filtered, then it is compressed to build up the pressure until 8 bar. This compressed flue gas is sent to three stages Pressure Swing Adsorption (PSA), which is filled with activated carbon. Experiments were showed the optimum adsorption pressure at 7bar, which CO₂ can be adsorbed step by step in 1st, 2nd, and 3rd stage, obtaining CO₂ concentration 29.8, 66.4, and 96.7 %, respectively. The mixed gas concentration from the last step is composed of 96.7% CO₂,2.7% N₂, and 0.6%O₂. This mixed CO₂product gas obtained from 3 stages PSA contained high concentration CO₂, which is ready to use for methanol synthesis. The mixed CO₂ was experimented in 5 Liter/Day of methanol synthesis reactor skid by 3 step processes as followed steam reforming, reverse water gas shift, and then hydrogenation. The result showed that proportional of mixed CO₂ and CH₄ 70/30, 50/50, 30/70 % (v/v), and 10/90 yielded methanol 2.4, 4.3, 5.6, and 6.0 Liter/day and save CO₂ 40, 30, 20, and 5 % respectively. The optimum condition resulted both methanol yield and CO₂ consumption using CO₂/CH₄ ratio 43/57 % (v/v), which yielded 4.8 Liter/day methanol and save CO₂ 27% comparing with traditional methanol production from methane steam reforming (5 Liter/day)and absent CO₂ consumption.

Keywords: carbon capture utilization and storage, pressure swing adsorption, reforming, reverse water gas shift, methanol

Procedia PDF Downloads 181
7785 Hybrid Heat Pump for Micro Heat Network

Authors: J. M. Counsell, Y. Khalid, M. J. Stewart

Abstract:

Achieving nearly zero carbon heating continues to be identified by UK government analysis as an important feature of any lowest cost pathway to reducing greenhouse gas emissions. Heat currently accounts for 48% of UK energy consumption and approximately one third of UK’s greenhouse gas emissions. Heat Networks are being promoted by UK investment policies as one means of supporting hybrid heat pump based solutions. To this effect the RISE (Renewable Integrated and Sustainable Electric) heating system project is investigating how an all-electric heating sourceshybrid configuration could play a key role in long-term decarbonisation of heat.  For the purposes of this study, hybrid systems are defined as systems combining the technologies of an electric driven air source heat pump, electric powered thermal storage, a thermal vessel and micro-heat network as an integrated system.  This hybrid strategy allows for the system to store up energy during periods of low electricity demand from the national grid, turning it into a dynamic supply of low cost heat which is utilized only when required. Currently a prototype of such a system is being tested in a modern house integrated with advanced controls and sensors. This paper presents the virtual performance analysis of the system and its design for a micro heat network with multiple dwelling units. The results show that the RISE system is controllable and can reduce carbon emissions whilst being competitive in running costs with a conventional gas boiler heating system.

Keywords: gas boilers, heat pumps, hybrid heating and thermal storage, renewable integrated and sustainable electric

Procedia PDF Downloads 414