Search results for: discrete optimization
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 3850

Search results for: discrete optimization

550 Solving LWE by Pregressive Pumps and Its Optimization

Authors: Leizhang Wang, Baocang Wang

Abstract:

General Sieve Kernel (G6K) is considered as currently the fastest algorithm for the shortest vector problem (SVP) and record holder of open SVP challenge. We study the lattice basis quality improvement effects of the Workout proposed in G6K, which is composed of a series of pumps to solve SVP. Firstly, we use a low-dimensional pump output basis to propose a predictor to predict the quality of high-dimensional Pumps output basis. Both theoretical analysis and experimental tests are performed to illustrate that it is more computationally expensive to solve the LWE problems by using a G6K default SVP solving strategy (Workout) than these lattice reduction algorithms (e.g. BKZ 2.0, Progressive BKZ, Pump, and Jump BKZ) with sieving as their SVP oracle. Secondly, the default Workout in G6K is optimized to achieve a stronger reduction and lower computational cost. Thirdly, we combine the optimized Workout and the Pump output basis quality predictor to further reduce the computational cost by optimizing LWE instances selection strategy. In fact, we can solve the TU LWE challenge (n = 65, q = 4225, = 0:005) 13.6 times faster than the G6K default Workout. Fourthly, we consider a combined two-stage (Preprocessing by BKZ- and a big Pump) LWE solving strategy. Both stages use dimension for free technology to give new theoretical security estimations of several LWE-based cryptographic schemes. The security estimations show that the securities of these schemes with the conservative Newhope’s core-SVP model are somewhat overestimated. In addition, in the case of LAC scheme, LWE instances selection strategy can be optimized to further improve the LWE-solving efficiency even by 15% and 57%. Finally, some experiments are implemented to examine the effects of our strategies on the Normal Form LWE problems, and the results demonstrate that the combined strategy is four times faster than that of Newhope.

Keywords: LWE, G6K, pump estimator, LWE instances selection strategy, dimension for free

Procedia PDF Downloads 60
549 Geological and Geotechnical Approach for Stabilization of Cut-Slopes in Power House Area of Luhri HEP Stage-I (210 MW), India

Authors: S. P. Bansal, Mukesh Kumar Sharma, Ankit Prabhakar

Abstract:

Luhri Hydroelectric Project Stage-I (210 MW) is a run of the river type development with a dam toe surface powerhouse (122m long, 50.50m wide, and 65.50m high) on the right bank of river Satluj in Himachal Pradesh, India. The project is located in the inner lesser Himalaya between Dhauladhar Range in the south and higher Himalaya in the north in the seismically active region. At the project, the location river is confined within narrow V-shaped valleys with little or no flat areas close to the river bed. Nearly 120m high cut slopes behind the powerhouse are proposed from the powerhouse foundation level of 795m to ± 915m to accommodate the surface powerhouse. The stability of 120m high cut slopes is a prime concern for the reason of risk involved. The slopes behind the powerhouse will be excavated in mainly in augen gneiss, fresh to weathered in nature, and biotite rich at places. The foliation joints are favorable and dipping inside the hill. Two valleys dipping steeper joints will be encountered on the slopes, which can cause instability during excavation. Geological exploration plays a vital role in designing and optimization of cut slopes. SWEDGE software has been used to analyze the geometry and stability of surface wedges in cut slopes. The slopes behind powerhouse have been analyzed in three zones for stability analysis by providing a break in the continuity of cut slopes, which shall provide quite substantial relief for slope stabilization measure. Pseudo static analysis has been carried out for the stabilization of wedges. The results indicate that many large wedges are forming, which have a factor of safety less than 1. The stability measures (support system, bench width, slopes) have been planned so that no wedge failure may occur in the future.

Keywords: cut slopes, geotechnical investigations, Himalayan geology, surface powerhouse, wedge failure

Procedia PDF Downloads 117
548 A 0-1 Goal Programming Approach to Optimize the Layout of Hospital Units: A Case Study in an Emergency Department in Seoul

Authors: Farhood Rismanchian, Seong Hyeon Park, Young Hoon Lee

Abstract:

This paper proposes a method to optimize the layout of an emergency department (ED) based on real executions of care processes by considering several planning objectives simultaneously. Recently, demand for healthcare services has been dramatically increased. As the demand for healthcare services increases, so do the need for new healthcare buildings as well as the need for redesign and renovating existing ones. The importance of implementation of a standard set of engineering facilities planning and design techniques has been already proved in both manufacturing and service industry with many significant functional efficiencies. However, high complexity of care processes remains a major challenge to apply these methods in healthcare environments. Process mining techniques applied in this study to tackle the problem of complexity and to enhance care process analysis. Process related information such as clinical pathways extracted from the information system of an ED. A 0-1 goal programming approach is then proposed to find a single layout that simultaneously satisfies several goals. The proposed model solved by optimization software CPLEX 12. The solution reached using the proposed method has 42.2% improvement in terms of walking distance of normal patients and 47.6% improvement in walking distance of critical patients at minimum cost of relocation. It has been observed that lots of patients must unnecessarily walk long distances during their visit to the emergency department because of an inefficient design. A carefully designed layout can significantly decrease patient walking distance and related complications.

Keywords: healthcare operation management, goal programming, facility layout problem, process mining, clinical processes

Procedia PDF Downloads 295
547 Myomectomy and Blood Loss: A Quality Improvement Project

Authors: Ena Arora, Rong Fan, Aleksandr Fuks, Kolawole Felix Akinnawonu

Abstract:

Introduction: Leiomyomas are benign tumors that are derived from the overgrowth of uterine smooth muscle cells. Women with symptomatic leiomyomas who desire future fertility, myomectomy should be the standard surgical treatment. Perioperative hemorrhage is a common complication in myomectomy. We performed the study to investigate blood transfusion rate in abdominal myomectomies, risk factors influencing blood loss and modalities to improve perioperative blood loss. Methods: Retrospective chart review was done for patients who underwent myomectomy from 2016 to 2022 at Queens hospital center, New York. We looked at preoperative patient demographics, clinical characteristics, intraoperative variables, and postoperative outcomes. Mann-Whitney U test were used for parametric and non-parametric continuous variable comparisons, respectively. Results: A total of 159 myomectomies were performed between 2016 and 2022, including 1 laparoscopic, 65 vaginal and 93 abdominal. 44 patients received blood transfusion during or within 72 hours of abdominal myomectomy. The blood transfusion rate was 47.3%. Blood transfusion rate was found to be twice higher than the average documented rate in literature which is 20%. Risk factors identified were black race, preoperative hematocrit<30%, preoperative blood transfusion within 72 hours, large fibroid burden, prolonged surgical time, and abdominal approach. Conclusion: Preoperative optimization with iron supplements or GnRH agonists is important for patients undergoing myomectomy. Interventions to decrease intra operative blood loss should include cell saver, tourniquet, vasopressin, misoprostol, tranexamic acid and gelatin-thrombin matrix hemostatic sealant.

Keywords: myomectomy, perioperative blood loss, cell saver, tranexamic acid

Procedia PDF Downloads 85
546 Loss Function Optimization for CNN-Based Fingerprint Anti-Spoofing

Authors: Yehjune Heo

Abstract:

As biometric systems become widely deployed, the security of identification systems can be easily attacked by various spoof materials. This paper contributes to finding a reliable and practical anti-spoofing method using Convolutional Neural Networks (CNNs) based on the types of loss functions and optimizers. The types of CNNs used in this paper include AlexNet, VGGNet, and ResNet. By using various loss functions including Cross-Entropy, Center Loss, Cosine Proximity, and Hinge Loss, and various loss optimizers which include Adam, SGD, RMSProp, Adadelta, Adagrad, and Nadam, we obtained significant performance changes. We realize that choosing the correct loss function for each model is crucial since different loss functions lead to different errors on the same evaluation. By using a subset of the Livdet 2017 database, we validate our approach to compare the generalization power. It is important to note that we use a subset of LiveDet and the database is the same across all training and testing for each model. This way, we can compare the performance, in terms of generalization, for the unseen data across all different models. The best CNN (AlexNet) with the appropriate loss function and optimizers result in more than 3% of performance gain over the other CNN models with the default loss function and optimizer. In addition to the highest generalization performance, this paper also contains the models with high accuracy associated with parameters and mean average error rates to find the model that consumes the least memory and computation time for training and testing. Although AlexNet has less complexity over other CNN models, it is proven to be very efficient. For practical anti-spoofing systems, the deployed version should use a small amount of memory and should run very fast with high anti-spoofing performance. For our deployed version on smartphones, additional processing steps, such as quantization and pruning algorithms, have been applied in our final model.

Keywords: anti-spoofing, CNN, fingerprint recognition, loss function, optimizer

Procedia PDF Downloads 136
545 The Use of Unmanned Aerial System (UAS) in Improving the Measurement System on the Example of Textile Heaps

Authors: Arkadiusz Zurek

Abstract:

The potential of using drones is visible in many areas of logistics, especially in terms of their use for monitoring and control of many processes. The technologies implemented in the last decade concern new possibilities for companies that until now have not even considered them, such as warehouse inventories. Unmanned aerial vehicles are no longer seen as a revolutionary tool for Industry 4.0, but rather as tools in the daily work of factories and logistics operators. The research problem is to develop a method for measuring the weight of goods in a selected link of the clothing supply chain by drones. However, the purpose of this article is to analyze the causes of errors in traditional measurements, and then to identify adverse events related to the use of drones for the inventory of a heap of textiles intended for production purposes. On this basis, it will be possible to develop guidelines to eliminate the causes of these events in the measurement process using drones. In a real environment, work was carried out to determine the volume and weight of textiles, including, among others, weighing a textile sample to determine the average density of the assortment, establishing a local geodetic network, terrestrial laser scanning and photogrammetric raid using an unmanned aerial vehicle. As a result of the analysis of measurement data obtained in the facility, the volume and weight of the assortment and the accuracy of their determination were determined. In this article, this work presents how such heaps are currently being tested, what adverse events occur, indicate and describes the current use of photogrammetric techniques of this type of measurements so far performed by external drones for the inventory of wind farms or construction of the station and compare them with the measurement system of the aforementioned textile heap inside a large-format facility.

Keywords: drones, unmanned aerial system, UAS, indoor system, security, process automation, cost optimization, photogrammetry, risk elimination, industry 4.0

Procedia PDF Downloads 86
544 Model-Based Approach as Support for Product Industrialization: Application to an Optical Sensor

Authors: Frederic Schenker, Jonathan J. Hendriks, Gianluca Nicchiotti

Abstract:

In a product industrialization perspective, the end-product shall always be at the peak of technological advancement and developed in the shortest time possible. Thus, the constant growth of complexity and a shorter time-to-market calls for important changes on both the technical and business level. Undeniably, the common understanding of the system is beclouded by its complexity which leads to the communication gap between the engineers and the sale department. This communication link is therefore important to maintain and increase the information exchange between departments to ensure a punctual and flawless delivery to the end customer. This evolution brings engineers to reason with more hindsight and plan ahead. In this sense, they use new viewpoints to represent the data and to express the model deliverables in an understandable way that the different stakeholder may identify their needs and ideas. This article focuses on the usage of Model-Based System Engineering (MBSE) in a perspective of system industrialization and reconnect the engineering with the sales team. The modeling method used and presented in this paper concentrates on displaying as closely as possible the needs of the customer. Firstly, by providing a technical solution to the sales team to help them elaborate commercial offers without omitting technicalities. Secondly, the model simulates between a vast number of possibilities across a wide range of components. It becomes a dynamic tool for powerful analysis and optimizations. Thus, the model is no longer a technical tool for the engineers, but a way to maintain and solidify the communication between departments using different views of the model. The MBSE contribution to cost optimization during New Product Introduction (NPI) activities is made explicit through the illustration of a case study describing the support provided by system models to architectural choices during the industrialization of a novel optical sensor.

Keywords: analytical model, architecture comparison, MBSE, product industrialization, SysML, system thinking

Procedia PDF Downloads 161
543 Technological Development of a Biostimulant Bioproduct for Fruit Seedlings: An Engineering Overview

Authors: Andres Diaz Garcia

Abstract:

The successful technological development of any bioproduct, including those of the biostimulant type, requires to adequately completion of a series of stages allied to different disciplines that are related to microbiological, engineering, pharmaceutical chemistry, legal and market components, among others. Engineering as a discipline has a key contribution in different aspects of fermentation processes such as the design and optimization of culture media, the standardization of operating conditions within the bioreactor and the scaling of the production process of the active ingredient that it will be used in unit operations downstream. However, all aspects mentioned must take into account many biological factors of the microorganism such as the growth rate, the level of assimilation to various organic and inorganic sources and the mechanisms of action associated with its biological activity. This paper focuses on the practical experience within the Colombian Corporation for Agricultural Research (Agrosavia), which led to the development of a biostimulant bioproduct based on native rhizobacteria Bacillus amyloliquefaciens, oriented mainly to plant growth promotion in cape gooseberry nurseries and fruit crops in Colombia, and the challenges that were overcome from the expertise in the area of engineering. Through the application of strategies and engineering tools, a culture medium was optimized to obtain concentrations higher than 1E09 CFU (colony form units)/ml in liquid fermentation, the process of biomass production was standardized and a scale-up strategy was generated based on geometric (H/D of bioreactor relationships), and operational criteria based on a minimum dissolved oxygen concentration and that took into account the differences in the capacity of control of the process in the laboratory and pilot scales. Currently, the bioproduct obtained through this technological process is in stages of registration in Colombia for cape gooseberry fruits for export.

Keywords: biochemical engineering, liquid fermentation, plant growth promoting, scale-up process

Procedia PDF Downloads 112
542 Biodiesel Production from Yellow Oleander Seed Oil

Authors: S. Rashmi, Devashish Das, N. Spoorthi, H. V. Manasa

Abstract:

Energy is essential and plays an important role for overall development of a nation. The global economy literally runs on energy. The use of fossil fuels as energy is now widely accepted as unsustainable due to depleting resources and also due to the accumulation of greenhouse gases in the environment, renewable and carbon neutral biodiesel are necessary for environment and economic sustainability. Unfortunately biodiesel produced from oil crop, waste cooking oil and animal fats are not able to replace fossil fuel. Fossil fuels remain the dominant source of primary energy, accounting for 84% of the overall increase in demand. Today biodiesel has come to mean a very specific chemical modification of natural oils. Objectives: To produce biodiesel from yellow oleander seed oil, to test the yield of biodiesel using different types of catalyst (KOH & NaOH). Methodology: Oil is extracted from dried yellow oleander seeds using Soxhlet extractor and oil expeller (bulk). The FFA content of the oil is checked and depending on the FFA value either two steps or single step process is followed to produce biodiesel. Two step processes includes esterfication and transesterification, single step includes only transesterification. The properties of biodiesel are checked. Engine test is done for biodiesel produced. Result: It is concluded that biodiesel quality parameters such as yield(85% & 90%), flash point(1710C & 1760C),fire point(1950C & 1980C), viscosity(4.9991 and 5.21 mm2/s) for the biodiesel from seed oil of Thevetiaperuviana produced by using KOH & NaOH respectively. Thus the seed oil of Thevetiaperuviana is a viable feedstock for good quality fuel.The outcomes of our project are a substitute for conventional fuel, to reduce petro diesel requirement,improved performance in terms of emissions. Future prospects: Optimization of biodiesel production using response surface method.

Keywords: yellow oleander seeds, biodiesel, quality parameters, renewable sources

Procedia PDF Downloads 446
541 FWGE Production From Wheat Germ Using Co-culture of Saccharomyces cerevisiae and Lactobacillus plantarum

Authors: Valiollah Babaeipour, Mahdi Rahaie

Abstract:

food supplements are rich in specific nutrients and bioactive compounds that eliminate free radicals and improve cellular metabolism. The major bioactive compounds are found in bran and cereal sprouts. Secondary metabolites of these microorganisms have antioxidant properties that can be used alone or in combination with chemotherapy and radiation therapy to treat cancer. Biologically active compounds such as benzoquinone derivatives extracted from fermented wheat germ extract (FWGE) have several positive effects on the overall state of human health and strengthen the immune system. The present work describes the discontinuous fermentation of raw wheat germ for FWGE production through the simultaneous culture process using the probiotic strains of Saccharomyces cerevisiae, Lactobacillus plantarum, and the possibility of using solid waste. To increase production efficiency, first to select important factors in the optimization of each fermentation process, using a factorial statistical scheme of stirring fraction (120 to 200 rpm), dilution of solids to solvent (1 to 8-12), fermentation time (16 to 24 hours) and strain to wheat germ ratio (20% to 50%) were studied and then simultaneous culture was performed to increase the yields of 2 and 6 dimethoxybenzoquinone (2,6-DMBQ). Since 2 and 6 dimethoxy benzoquinone were fermented as the main biologically active compound in wheat germ extract, UV-Vis analysis was performed to confirm the presence of 2 and 6 dimethoxy benzoquinone in the final product. In addition, 2,6-DMBQ of some products was isolated in a non-polar C-18 column and quantified using high performance liquid chromatography (HPLC). Based on our findings, it can be concluded that the increase of 2 and 6 dimethoxybenzoquinone in the simultaneous culture of Saccharomyces cerevisiae - Lactobacillus plantarum compared to pure culture of Saccharomyces cerevisiae (from 1.89 mg / g) to 28.9% (2.66 mg / g) Increased.

Keywords: wheat germ, FWGE, saccharomyces cerevisiae, lactobacillus plantarum, co-culture, 2, 6-DMBQ

Procedia PDF Downloads 130
540 Enhanced Photocatalytic H₂ Production from H₂S on Metal Modified Cds-Zns Semiconductors

Authors: Maali-Amel Mersel, Lajos Fodor, Otto Horvath

Abstract:

Photocatalytic H₂ production by H₂S decomposition is regarded to be an environmentally friendly process to produce carbon-free energy through direct solar energy conversion. For this purpose, sulphide-based materials, as photocatalysts, were widely used due to their excellent solar spectrum responses and high photocatalytic activity. The loading of proper co-catalysts that are based on cheap and earth-abundant materials on those semiconductors was shown to play an important role in the improvement of their efficiency. In this research, CdS-ZnS composite was studied because of its controllable band gap and excellent performance for H₂ evolution under visible light irradiation. The effects of the modification of this photocatalyst with different types of materials and the influence of the preparation parameters on its H₂ production activity were investigated. The CdS-ZnS composite with an enhanced photocatalytic activity for H₂ production was synthesized from ammine complexes. Two types of modification were used: compounds of Ni-group metals (NiS, PdS, and Pt) were applied as co-catalyst on the surface of CdS-ZnS semiconductor, while NiS, MnS, CoS, Ag₂S, and CuS were used as a dopant in the bulk of the catalyst. It was found that 0.1% of noble metals didn’t remarkably influence the photocatalytic activity, while the modification with 0.5% of NiS was shown to be more efficient in the bulk than on the surface. The modification with other types of metals results in a decrease of the rate of H₂ production, while the co-doping seems to be more promising. The preparation parameters (such as the amount of ammonia to form the ammine complexes, the order of the preparation steps together with the hydrothermal treatment) were also found to highly influence the rate of H₂ production. SEM, EDS and DRS analyses were made to reveal the structure of the most efficient photocatalysts. Moreover, the detection of the conduction band electron on the surface of the catalyst was also investigated. The excellent photoactivity of the CdS-ZnS catalysts with and without modification encourages further investigations to enhance the hydrogen generation by optimization of the reaction conditions.

Keywords: H₂S, photoactivity, photocatalytic H₂ production, CdS-ZnS

Procedia PDF Downloads 131
539 Generative Design Method for Cooled Additively Manufactured Gas Turbine Parts

Authors: Thomas Wimmer, Bernhard Weigand

Abstract:

The improvement of gas turbine efficiency is one of the main drivers of research and development in the gas turbine market. This has led to elevated gas turbine inlet temperatures beyond the melting point of the utilized materials. The turbine parts need to be actively cooled in order to withstand these harsh environments. However, the usage of compressor air as coolant decreases the overall gas turbine efficiency. Thus, coolant consumption needs to be minimized in order to gain the maximum advantage from higher turbine inlet temperatures. Therefore, sophisticated cooling designs for gas turbine parts aim to minimize coolant mass flow. New design space is accessible as additive manufacturing is maturing to industrial usage for the creation of hot gas flow path parts. By making use of this technology more efficient cooling schemes can be manufacture. In order to find such cooling schemes a generative design method is being developed. It generates cooling schemes randomly which adhere to a set of rules. These assure the sanity of the design. A huge amount of different cooling schemes are generated and implemented in a simulation environment where it is validated. Criteria for the fitness of the cooling schemes are coolant mass flow, maximum temperature and temperature gradients. This way the whole design space is sampled and a Pareto optimum front can be identified. This approach is applied to a flat plate, which resembles a simplified section of a hot gas flow path part. Realistic boundary conditions are applied and thermal barrier coating is accounted for in the simulation environment. The resulting cooling schemes are presented and compared to representative conventional cooling schemes. Further development of this method can give access to cooling schemes with an even better performance having higher complexity, which makes use of the available design space.

Keywords: additive manufacturing, cooling, gas turbine, heat transfer, heat transfer design, optimization

Procedia PDF Downloads 352
538 Multiple-Material Flow Control in Construction Supply Chain with External Storage Site

Authors: Fatmah Almathkour

Abstract:

Managing and controlling the construction supply chain (CSC) are very important components of effective construction project execution. The goals of managing the CSC are to reduce uncertainty and optimize the performance of a construction project by improving efficiency and reducing project costs. The heart of much SC activity is addressing risk, and the CSC is no different. The delivery and consumption of construction materials is highly variable due to the complexity of construction operations, rapidly changing demand for certain components, lead time variability from suppliers, transportation time variability, and disruptions at the job site. Current notions of managing and controlling CSC, involve focusing on one project at a time with a push-based material ordering system based on the initial construction schedule and, then, holding a tremendous amount of inventory. A two-stage methodology was proposed to coordinate the feed-forward control of advanced order placement with a supplier to a feedback local control in the form of adding the ability to transship materials between projects to improve efficiency and reduce costs. It focused on the single supplier integrated production and transshipment problem with multiple products. The methodology is used as a design tool for the CSC because it includes an external storage site not associated with one of the projects. The idea is to add this feature to a highly constrained environment to explore its effectiveness in buffering the impact of variability and maintaining project schedule at low cost. The methodology uses deterministic optimization models with objectives that minimizing the total cost of the CSC. To illustrate how this methodology can be used in practice and the types of information that can be gleaned, it is tested on a number of cases based on the real example of multiple construction projects in Kuwait.

Keywords: construction supply chain, inventory control supply chain, transshipment

Procedia PDF Downloads 122
537 Modelling and Optimization of a Combined Sorption Enhanced Biomass Gasification with Hydrothermal Carbonization, Hot Gas Cleaning and Dielectric Barrier Discharge Plasma Reactor to Produce Pure H₂ and Methanol Synthesis

Authors: Vera Marcantonio, Marcello De Falco, Mauro Capocelli, Álvaro Amado-Fierro, Teresa A. Centeno, Enrico Bocci

Abstract:

Concerns about energy security, energy prices, and climate change led scientific research towards sustainable solutions to fossil fuel as renewable energy sources coupled with hydrogen as an energy vector and carbon capture and conversion technologies. Among the technologies investigated in the last decades, biomass gasification acquired great interest owing to the possibility of obtaining low-cost and CO₂ negative emission hydrogen production from a large variety of everywhere available organic wastes. Upstream and downstream treatment were then studied in order to maximize hydrogen yield, reduce the content of organic and inorganic contaminants under the admissible levels for the technologies which are coupled with, capture, and convert carbon dioxide. However, studies which analyse a whole process made of all those technologies are still missing. In order to fill this lack, the present paper investigated the coexistence of hydrothermal carbonization (HTC), sorption enhance gasification (SEG), hot gas cleaning (HGC), and CO₂ conversion by dielectric barrier discharge (DBD) plasma reactor for H₂ production from biomass waste by means of Aspen Plus software. The proposed model aimed to identify and optimise the performance of the plant by varying operating parameters (such as temperature, CaO/biomass ratio, separation efficiency, etc.). The carbon footprint of the global plant is 2.3 kg CO₂/kg H₂, lower than the latest limit value imposed by the European Commission to consider hydrogen as “clean”, that was set to 3 kg CO₂/kg H₂. The hydrogen yield referred to the whole plant is 250 gH₂/kgBIOMASS.

Keywords: biomass gasification, hydrogen, aspen plus, sorption enhance gasification

Procedia PDF Downloads 78
536 Load Forecasting in Microgrid Systems with R and Cortana Intelligence Suite

Authors: F. Lazzeri, I. Reiter

Abstract:

Energy production optimization has been traditionally very important for utilities in order to improve resource consumption. However, load forecasting is a challenging task, as there are a large number of relevant variables that must be considered, and several strategies have been used to deal with this complex problem. This is especially true also in microgrids where many elements have to adjust their performance depending on the future generation and consumption conditions. The goal of this paper is to present a solution for short-term load forecasting in microgrids, based on three machine learning experiments developed in R and web services built and deployed with different components of Cortana Intelligence Suite: Azure Machine Learning, a fully managed cloud service that enables to easily build, deploy, and share predictive analytics solutions; SQL database, a Microsoft database service for app developers; and PowerBI, a suite of business analytics tools to analyze data and share insights. Our results show that Boosted Decision Tree and Fast Forest Quantile regression methods can be very useful to predict hourly short-term consumption in microgrids; moreover, we found that for these types of forecasting models, weather data (temperature, wind, humidity and dew point) can play a crucial role in improving the accuracy of the forecasting solution. Data cleaning and feature engineering methods performed in R and different types of machine learning algorithms (Boosted Decision Tree, Fast Forest Quantile and ARIMA) will be presented, and results and performance metrics discussed.

Keywords: time-series, features engineering methods for forecasting, energy demand forecasting, Azure Machine Learning

Procedia PDF Downloads 297
535 Research on Localized Operations of Multinational Companies in China

Authors: Zheng Ruoyuan

Abstract:

With the rapid development of economic globalization and increasingly fierce international competition, multinational companies have carried out investment strategy shifts and innovations, and actively promoted localization strategies. Localization strategies have become the main trend in the development of multinational companies. Large-scale entry of multinational companies China has a history of more than 20 years. With the sustained and steady growth of China's economy and the optimization of the investment environment, multinational companies' investment in China has expanded rapidly, which has also had an important impact on the Chinese economy: promoting employment, foreign exchange reserves, and improving the system. etc., has brought a lot of high-tech and advanced management experience; but it has also brought challenges and survival pressure to China's local enterprises. In recent years, multinational companies have gradually regarded China as an important part of their global strategies and began to invest in China. Actively promote localization strategies, including production, marketing, scientific research and development, etc. Many multinational companies have achieved good results in localized operations in China. Not only have their benefits continued to improve, but they have also established a good corporate image and brand in China. image, which has greatly improved their competitiveness in the international market. However, there are also some multinational companies that have difficulties in localized operations in China. This article will closely follow the background of economic globalization and comprehensively use the theory of multinational companies and strategic management theory and business management theory, using data and facts as the entry point, combined with typical cases of representative significance for analysis, to conduct a systematic study of the localized operations of multinational companies in China. At the same time, for each specific link of the operation of multinational companies, we provide multinational enterprises with some inspirations and references.

Keywords: localization, business management, multinational, marketing

Procedia PDF Downloads 49
534 Optimization of Beneficiation Process for Upgrading Low Grade Egyptian Kaolin

Authors: Nagui A. Abdel-Khalek, Khaled A. Selim, Ahmed Hamdy

Abstract:

Kaolin is naturally occurring ore predominantly containing kaolinite mineral in addition to some gangue minerals. Typical impurities present in kaolin ore are quartz, iron oxides, titanoferrous minerals, mica, feldspar, organic matter, etc. The main coloring impurity, particularly in the ultrafine size range, is titanoferrous minerals. Kaolin is used in many industrial applications such as sanitary ware, table ware, ceramic, paint, and paper industries, each of which should be of certain specifications. For most industrial applications, kaolin should be processed to obtain refined clay so as to match with standard specifications. For example, kaolin used in paper and paint industries need to be of high brightness and low yellowness. Egyptian kaolin is not subjected to any beneficiation process and the Egyptian companies apply selective mining followed by, in some localities, crushing and size reduction only. Such low quality kaolin can be used in refractory and pottery production but not in white ware and paper industries. This paper aims to study the amenability of beneficiation of an Egyptian kaolin ore of El-Teih locality, Sinai, to be suitable for different industrial applications. Attrition scrubbing and classification followed by magnetic separation are applied to remove the associated impurities. Attrition scrubbing and classification are used to separate the coarse silica and feldspars. Wet high intensity magnetic separation was applied to remove colored contaminants such as iron oxide and titanium oxide. Different variables affecting of magnetic separation process such as solid percent, magnetic field, matrix loading capacity, and retention time are studied. The results indicated that substantial decrease in iron oxide (from 1.69% to 0.61% ) and TiO2 (from 3.1% to 0.83%) contents as well as improving iso-brightness (from 63.76% to 75.21% and whiteness (from 79.85% to 86.72%) of the product can be achieved.

Keywords: Kaolin, titanoferrous minerals, beneficiation, magnetic separation, attrition scrubbing, classification

Procedia PDF Downloads 360
533 A Segmentation Method for Grayscale Images Based on the Firefly Algorithm and the Gaussian Mixture Model

Authors: Donatella Giuliani

Abstract:

In this research, we propose an unsupervised grayscale image segmentation method based on a combination of the Firefly Algorithm and the Gaussian Mixture Model. Firstly, the Firefly Algorithm has been applied in a histogram-based research of cluster means. The Firefly Algorithm is a stochastic global optimization technique, centered on the flashing characteristics of fireflies. In this context it has been performed to determine the number of clusters and the related cluster means in a histogram-based segmentation approach. Successively these means are used in the initialization step for the parameter estimation of a Gaussian Mixture Model. The parametric probability density function of a Gaussian Mixture Model is represented as a weighted sum of Gaussian component densities, whose parameters are evaluated applying the iterative Expectation-Maximization technique. The coefficients of the linear super-position of Gaussians can be thought as prior probabilities of each component. Applying the Bayes rule, the posterior probabilities of the grayscale intensities have been evaluated, therefore their maxima are used to assign each pixel to the clusters, according to their gray-level values. The proposed approach appears fairly solid and reliable when applied even to complex grayscale images. The validation has been performed by using different standard measures, more precisely: the Root Mean Square Error (RMSE), the Structural Content (SC), the Normalized Correlation Coefficient (NK) and the Davies-Bouldin (DB) index. The achieved results have strongly confirmed the robustness of this gray scale segmentation method based on a metaheuristic algorithm. Another noteworthy advantage of this methodology is due to the use of maxima of responsibilities for the pixel assignment that implies a consistent reduction of the computational costs.

Keywords: clustering images, firefly algorithm, Gaussian mixture model, meta heuristic algorithm, image segmentation

Procedia PDF Downloads 217
532 Chromatographic Preparation and Performance on Zinc Ion Imprinted Monolithic Column and Its Adsorption Property

Authors: X. Han, S. Duan, C. Liu, C. Zhou, W. Zhu, L. Kong

Abstract:

The ionic imprinting technique refers to the three-dimensional rigid structure with the fixed pore sizes, which was formed by the binding interactions of ions and functional monomers and used ions as the template, it has a high level of recognition to the ionic template. The preparation of monolithic column by the in-situ polymerization need to put the compound of template, functional monomers, cross-linking agent and initiating agent into the solution, dissolve it and inject to the column tube, and then the compound will have a polymerization reaction at a certain temperature, after the synthetic reaction, we washed out the unread template and solution. The monolithic columns are easy to prepare, low consumption and cost-effective with fast mass transfer, besides, they have many chemical functions. But the monolithic columns have some problems in the practical application, such as low-efficiency, quantitative analysis cannot be performed accurately because of the peak shape is wide and has tailing phenomena; the choice of polymerization systems is limited and the lack of theoretical foundations. Thus the optimization of components and preparation methods is an important research direction. During the preparation of ionic imprinted monolithic columns, pore-forming agent can make the polymer generate the porous structure, which can influence the physical properties of polymer, what’ s more, it can directly decide the stability and selectivity of polymerization reaction. The compounds generated in the pre-polymerization reaction could directly decide the identification and screening capabilities of imprinted polymer; thus the choice of pore-forming agent is quite critical in the preparation of imprinted monolithic columns. This article mainly focuses on the research that when using different pore-forming agents, the impact of zinc ion imprinted monolithic column on the enrichment performance of zinc ion.

Keywords: high performance liquid chromatography (HPLC), ionic imprinting, monolithic column, pore-forming agent

Procedia PDF Downloads 214
531 Optimum Structural Wall Distribution in Reinforced Concrete Buildings Subjected to Earthquake Excitations

Authors: Nesreddine Djafar Henni, Akram Khelaifia, Salah Guettala, Rachid Chebili

Abstract:

Reinforced concrete shear walls and vertical plate-like elements play a pivotal role in efficiently managing a building's response to seismic forces. This study investigates how the performance of reinforced concrete buildings equipped with shear walls featuring different shear wall-to-frame stiffness ratios aligns with the requirements stipulated in the Algerian seismic code RPA99v2003, particularly in high-seismicity regions. Seven distinct 3D finite element models are developed and evaluated through nonlinear static analysis. Engineering Demand Parameters (EDPs) such as lateral displacement, inter-story drift ratio, shear force, and bending moment along the building height are analyzed. The findings reveal two predominant categories of induced responses: force-based and displacement-based EDPs. Furthermore, as the shear wall-to-frame ratio increases, there is a concurrent increase in force-based EDPs and a decrease in displacement-based ones. Examining the distribution of shear walls from both force and displacement perspectives, model G with the highest stiffness ratio, concentrating stiffness at the building's center, intensifies induced forces. This configuration necessitates additional reinforcements, leading to a conservative design approach. Conversely, model C, with the lowest stiffness ratio, distributes stiffness towards the periphery, resulting in minimized induced shear forces and bending moments, representing an optimal scenario with maximal performance and minimal strength requirements.

Keywords: dual RC buildings, RC shear walls, modeling, static nonlinear pushover analysis, optimization, seismic performance

Procedia PDF Downloads 56
530 Developing Urban Design and Planning Approach to Enhance the Efficiency of Infrastructure and Public Transportation in Order to Reduce GHG Emissions

Authors: A. Rostampouryasouri, A. Maghoul, S. Tahersima

Abstract:

The rapid growth of urbanization and the subsequent increase in population in cities have resulted in the destruction of the environment to cater to the needs of citizens. The industrialization of urban life has led to the production of pollutants, which has significantly contributed to the rise of air pollution. Infrastructure can have both positive and negative effects on air pollution. The effects of infrastructure on air pollution are complex and depend on various factors such as the type of infrastructure, location, and context. This study examines the effects of infrastructure on air pollution, drawing on a range of empirical evidence from Iran and China. Our paper focus for analyzing the data is on the following concepts: 1. Urban design and planning principles and practices 2. Infrastructure efficiency and optimization strategies 3. Public transportation systems and their environmental impact 4. GHG emissions reduction strategies in urban areas 5. Case studies and best practices in sustainable urban development This paper employs a mixed methodology approach with a focus on developmental and applicative purposes. The mixed methods approach combines both quantitative and qualitative research methods to provide a more comprehensive understanding of the research topic. A group of 20 architectural specialists and experts who are proficient in the field of research, design, and implementation of green architecture projects were interviewed in a systematic and purposeful manner. The research method was based on content analysis using MAXQDA2020 software. The findings suggest that policymakers and urban planners should consider the potential impacts of infrastructure on air pollution and take measures to mitigate negative effects while maximizing positive ones. This includes adopting a nature-based approach to urban planning and infrastructure development, investing in information infrastructure, and promoting modern logistic transport infrastructure.

Keywords: GHG emissions, infrastructure efficiency, urban development, urban design

Procedia PDF Downloads 77
529 Developing a Sustainable Transit Planning Index Using Analytical Hierarchy Process Method for ZEB Implementation in Canada

Authors: Mona Ghafouri-Azar, Sara Diamond, Jeremy Bowes, Grace Yuan, Aimee Burnett, Michelle Wyndham-West, Sara Wagner, Anand Pariyarath

Abstract:

Transportation is the fastest growing source of greenhouse gas emissions worldwide. In Canada, it is responsible for 23% of total CO2emissions from fuel combustion, and emissions from the transportation sector are the second largest source of emissions after the oil and gas sector. Currently, most Canadian public transportation systems rely on buses that operateon fossil fuels.Canada is currently investing billions of dollars to replacediesel buses with electric busesas this isperceived to have a significant impact on climate mitigation. This paper focuses on the possible impacts of zero emission buses (ZEB) on sustainable development, considering three dimensions of sustainability; environmental quality, economic growth, and social development.A sustainable transportation system is one that is safe, affordable, accessible, efficient, and resilient and that contributes minimal emissions of carbon and other pollutants.To enable implementation of these goals, relevant indicators were selected and defined that measure progress towards a sustainable transportation system. These were drawn from Canadian and international examples. Studies compare different European cities in terms of development, sustainability, and infrastructures, by using transport performance indicators. A Normalized Transport Sustainability index measures and compares policies in different urban areas and allows fine-tuning of policies. Analysts use a number ofmethods for sustainable analysis, like cost-benefit analysis (CBA) toassess economic benefit, life-cycle assessment (LCA) to assess social, economic, and environment factors and goals, and multi-criteria decision making (MCDM) analysis which can comparediffering stakeholder preferences.A multi criteria decision making approach is an appropriate methodology to plan and evaluate sustainable transit development and to provide insights and meaningful information for decision makers and transit agencies. It is essential to develop a system thataggregates specific discrete indices to assess the sustainability of transportation systems.Theseprioritize indicators appropriate for the differentCanadian transit system agencies and theirpreferences and requirements. This studywill develop an integrating index that alliesexistingdiscrete indexes to supporta reliable comparison between the current transportation system (diesel buses) and the new ZEB system emerging in Canada. As a first step, theindexes for each category are selected, and the index matrix constructed. Second, the selected indicators arenormalized to remove anyinconsistency between them. Next, the normalized matrix isweighted based on the relative importance of each index to the main domains of sustainability using the analytical hierarchy process (AHP) method. This is accomplished through expert judgement around the relative importance of different attributes with respect to the goals through apairwise comparison matrix. The considerationof multiple environmental, economic, and social factors (including equity and health) is integrated intoa sustainable transit planning index (STPI) which supportsrealistic ZEB implementation in Canada and beyond and is useful to different stakeholders, agencies, and ministries.

Keywords: zero emission buses, sustainability, sustainable transit, transportation, analytical hierarchy process, environment, economy, social

Procedia PDF Downloads 128
528 Redesigning the Plant Distribution of an Industrial Laundry in Arequipa

Authors: Ana Belon Hercilla

Abstract:

The study is developed in “Reactivos Jeans” company, in the city of Arequipa, whose main business is the laundry of garments at an industrial level. In 2012 the company initiated actions to provide a dry cleaning service of alpaca fiber garments, recognizing that this item is in a growth phase in Peru. Additionally this company took the initiative to use a new greenwashing technology which has not yet been developed in the country. To accomplish this, a redesign of both the process and the plant layout was required. For redesigning the plant, the methodology used was the Systemic Layout Planning, allowing this study divided into four stages. First stage is the information gathering and evaluation of the initial situation of the company, for which a description of the areas, facilities and initial equipment, distribution of the plant, the production process and flows of major operations was made. Second stage is the development of engineering techniques that allow the logging and analysis procedures, such as: Flow Diagram, Route Diagram, DOP (process flowchart), DAP (analysis diagram). Then the planning of the general distribution is carried out. At this stage, proximity factors of the areas are established, the Diagram Paths (TRA) is developed, and the Relational Diagram Activities (DRA). In order to obtain the General Grouping Diagram (DGC), further information is complemented by a time study and Guerchet method is used to calculate the space requirements for each area. Finally, the plant layout redesigning is presented and the implementation of the improvement is made, making it possible to obtain a model much more efficient than the initial design. The results indicate that the implementation of the new machinery, the adequacy of the plant facilities and equipment relocation resulted in a reduction of the production cycle time by 75.67%, routes were reduced by 68.88%, the number of activities during the process were reduced by 40%, waits and storage were removed 100%.

Keywords: redesign, time optimization, industrial laundry, greenwashing

Procedia PDF Downloads 394
527 Hydrogen Production at the Forecourt from Off-Peak Electricity and Its Role in Balancing the Grid

Authors: Abdulla Rahil, Rupert Gammon, Neil Brown

Abstract:

The rapid growth of renewable energy sources and their integration into the grid have been motivated by the depletion of fossil fuels and environmental issues. Unfortunately, the grid is unable to cope with the predicted growth of renewable energy which would lead to its instability. To solve this problem, energy storage devices could be used. Electrolytic hydrogen production from an electrolyser is considered a promising option since it is a clean energy source (zero emissions). Choosing flexible operation of an electrolyser (producing hydrogen during the off-peak electricity period and stopping at other times) could bring about many benefits like reducing the cost of hydrogen and helping to balance the electric systems. This paper investigates the price of hydrogen during flexible operation compared with continuous operation, while serving the customer (hydrogen filling station) without interruption. The optimization algorithm is applied to investigate the hydrogen station in both cases (flexible and continuous operation). Three different scenarios are tested to see whether the off-peak electricity price could enhance the reduction of the hydrogen cost. These scenarios are: Standard tariff (1 tier system) during the day (assumed 12 p/kWh) while still satisfying the demand for hydrogen; using off-peak electricity at a lower price (assumed 5 p/kWh) and shutting down the electrolyser at other times; using lower price electricity at off-peak times and high price electricity at other times. This study looks at Derna city, which is located on the coast of the Mediterranean Sea (32° 46′ 0 N, 22° 38′ 0 E) with a high potential for wind resource. Hourly wind speed data which were collected over 24½ years from 1990 to 2014 were in addition to data on hourly radiation and hourly electricity demand collected over a one-year period, together with the petrol station data.

Keywords: hydrogen filling station off-peak electricity, renewable energy, off-peak electricity, electrolytic hydrogen

Procedia PDF Downloads 231
526 BIM Modeling of Site and Existing Buildings: Case Study of ESTP Paris Campus

Authors: Rita Sassine, Yassine Hassani, Mohamad Al Omari, Stéphanie Guibert

Abstract:

Building Information Modelling (BIM) is the process of creating, managing, and centralizing information during the building lifecycle. BIM can be used all over a construction project, from the initiation phase to the planning and execution phases to the maintenance and lifecycle management phase. For existing buildings, BIM can be used for specific applications such as lifecycle management. However, most of the existing buildings don’t have a BIM model. Creating a compatible BIM for existing buildings is very challenging. It requires special equipment for data capturing and efforts to convert these data into a BIM model. The main difficulties for such projects are to define the data needed, the level of development (LOD), and the methodology to be adopted. In addition to managing information for an existing building, studying the impact of the built environment is a challenging topic. So, integrating the existing terrain that surrounds buildings into the digital model is essential to be able to make several simulations as flood simulation, energy simulation, etc. Making a replication of the physical model and updating its information in real-time to make its Digital Twin (DT) is very important. The Digital Terrain Model (DTM) represents the ground surface of the terrain by a set of discrete points with unique height values over 2D points based on reference surface (e.g., mean sea level, geoid, and ellipsoid). In addition, information related to the type of pavement materials, types of vegetation and heights and damaged surfaces can be integrated. Our aim in this study is to define the methodology to be used in order to provide a 3D BIM model for the site and the existing building based on the case study of “Ecole Spéciale des Travaux Publiques (ESTP Paris)” school of engineering campus. The property is located on a hilly site of 5 hectares and is composed of more than 20 buildings with a total area of 32 000 square meters and a height between 50 and 68 meters. In this work, the campus precise levelling grid according to the NGF-IGN69 altimetric system and the grid control points are computed according to (Réseau Gédésique Français) RGF93 – Lambert 93 french system with different methods: (i) Land topographic surveying methods using robotic total station, (ii) GNSS (Global Network Satellite sytem) levelling grid with NRTK (Network Real Time Kinematic) mode, (iii) Point clouds generated by laser scanning. These technologies allow the computation of multiple building parameters such as boundary limits, the number of floors, the floors georeferencing, the georeferencing of the 4 base corners of each building, etc. Once the entry data are identified, the digital model of each building is done. The DTM is also modeled. The process of altimetric determination is complex and requires efforts in order to collect and analyze multiple data formats. Since many technologies can be used to produce digital models, different file formats such as DraWinG (DWG), LASer (LAS), Comma-separated values (CSV), Industry Foundation Classes (IFC) and ReViT (RVT) will be generated. Checking the interoperability between BIM models is very important. In this work, all models are linked together and shared on 3DEXPERIENCE collaborative platform.

Keywords: building information modeling, digital terrain model, existing buildings, interoperability

Procedia PDF Downloads 112
525 In Silico Exploration of Quinazoline Derivatives as EGFR Inhibitors for Lung Cancer: A Multi-Modal Approach Integrating QSAR-3D, ADMET, Molecular Docking, and Molecular Dynamics Analyses

Authors: Mohamed Moussaoui

Abstract:

A series of thirty-one potential inhibitors targeting the epidermal growth factor receptor kinase (EGFR), derived from quinazoline, underwent 3D-QSAR analysis using CoMFA and CoMSIA methodologies. The training and test sets of quinazoline derivatives were utilized to construct and validate the QSAR models, respectively, with dataset alignment performed using the lowest energy conformer of the most active compound. The best-performing CoMFA and CoMSIA models demonstrated impressive determination coefficients, with R² values of 0.981 and 0.978, respectively, and Leave One Out cross-validation determination coefficients, Q², of 0.645 and 0.729, respectively. Furthermore, external validation using a test set of five compounds yielded predicted determination coefficients, R² test, of 0.929 and 0.909 for CoMFA and CoMSIA, respectively. Building upon these promising results, eighteen new compounds were designed and assessed for drug likeness and ADMET properties through in silico methods. Additionally, molecular docking studies were conducted to elucidate the binding interactions between the selected compounds and the enzyme. Detailed molecular dynamics simulations were performed to analyze the stability, conformational changes, and binding interactions of the quinazoline derivatives with the EGFR kinase. These simulations provided deeper insights into the dynamic behavior of the compounds within the active site. This comprehensive analysis enhances the understanding of quinazoline derivatives as potential anti-cancer agents and provides valuable insights for lead optimization in the early stages of drug discovery, particularly for developing highly potent anticancer therapeutics

Keywords: 3D-QSAR, CoMFA, CoMSIA, ADMET, molecular docking, quinazoline, molecular dynamic, egfr inhibitors, lung cancer, anticancer

Procedia PDF Downloads 48
524 Review on Implementation of Artificial Intelligence and Machine Learning for Controlling Traffic and Avoiding Accidents

Authors: Neha Singh, Shristi Singh

Abstract:

Accidents involving motor vehicles are more likely to cause serious injuries and fatalities. It also has a host of other perpetual issues, such as the regular loss of life and goods in accidents. To solve these issues, appropriate measures must be implemented, such as establishing an autonomous incident detection system that makes use of machine learning and artificial intelligence. In order to reduce traffic accidents, this article examines the overview of artificial intelligence and machine learning in autonomous event detection systems. The paper explores the major issues, prospective solutions, and use of artificial intelligence and machine learning in road transportation systems for minimising traffic accidents. There is a lot of discussion on additional, fresh, and developing approaches that less frequent accidents in the transportation industry. The study structured the following subtopics specifically: traffic management using machine learning and artificial intelligence and an incident detector with these two technologies. The internet of vehicles and vehicle ad hoc networks, as well as the use of wireless communication technologies like 5G wireless networks and the use of machine learning and artificial intelligence for the planning of road transportation systems, are elaborated. In addition, safety is the primary concern of road transportation. Route optimization, cargo volume forecasting, predictive fleet maintenance, real-time vehicle tracking, and traffic management, according to the review's key conclusions, are essential for ensuring the safety of road transportation networks. In addition to highlighting research trends, unanswered problems, and key research conclusions, the study also discusses the difficulties in applying artificial intelligence to road transport systems. Planning and managing the road transportation system might use the work as a resource.

Keywords: artificial intelligence, machine learning, incident detector, road transport systems, traffic management, automatic incident detection, deep learning

Procedia PDF Downloads 113
523 Optimization and Kinetic Analysis of the Enzymatic Hydrolysis of Oil Palm Empty Fruit Bunch To Xylose Using Crude Xylanase from Trichoderma Viride ITB CC L.67

Authors: Efri Mardawati, Ronny Purwadi, Made Tri Ari Penia Kresnowati, Tjandra Setiadi

Abstract:

EFB are mainly composed of cellulose (≈ 43%), hemicellulose (≈ 23%) and lignin (≈20%). The palm oil empty fruit bunches (EFB) is the lignosellulosic waste from crude palm oil industries mainly compose of (≈ 43%), hemicellulose (≈ 23%) and lignin (≈20%). Xylan, a polymer made of pentose sugar xylose and the most abundant component of hemicellulose in plant cell wall. Further xylose can be used as a raw material for production of a wide variety of chemicals such as xylitol, which is extensively used in food, pharmaceutical and thin coating applications. Currently, xylose is mostly produced from xylan via chemical hydrolysis processes. However, these processes are normally conducted at a high temperature and pressure, which is costly, and the required downstream processes are relatively complex. As an alternative method, enzymatic hydrolysis of xylan to xylose offers an environmentally friendly biotechnological process, which is performed at ambient temperature and pressure with high specificity and at low cost. This process is catalysed by xylanolytic enzymes that can be produced by some fungal species such as Aspergillus niger, Penicillium crysogenum, Tricoderma reseei, etc. Fungal that will be used to produce crude xylanase enzyme in this study is T. Viride ITB CC L.67. It is the purposes of this research to study the influence of pretreatment of EFB for the enzymatic hydrolysis process, optimation of temperature and pH of the hydrolysis process, the influence of substrate and enzyme concentration to the enzymatic hydrolysis process, the dynamics of hydrolysis process and followingly to study the kinetics of this process. Xylose as the product of enzymatic hydrolysis process analyzed by HPLC. The results show that the thermal pretreatment of EFB enhance the enzymatic hydrolysis process. The enzymatic hydrolysis can be well approached by the Michaelis Menten kinetic model, and kinetic parameters are obtained from experimental data.

Keywords: oil palm empty fruit bunches (EFB), xylose, enzymatic hydrolysis, kinetic modelling

Procedia PDF Downloads 389
522 Development of Wave-Dissipating Block Installation Simulation for Inexperienced Worker Training

Authors: Hao Min Chuah, Tatsuya Yamazaki, Ryosui Iwasawa, Tatsumi Suto

Abstract:

In recent years, with the advancement of digital technology, the movement to introduce so-called ICT (Information and Communication Technology), such as computer technology and network technology, to civil engineering construction sites and construction sites is accelerating. As part of this movement, attempts are being made in various situations to reproduce actual sites inside computers and use them for designing and construction planning, as well as for training inexperienced engineers. The installation of wave-dissipating blocks on coasts, etc., is a type of work that has been carried out by skilled workers based on their years of experience and is one of the tasks that is difficult for inexperienced workers to carry out on site. Wave-dissipating blocks are structures that are designed to protect coasts, beaches, and so on from erosion by reducing the energy of ocean waves. Wave-dissipating blocks usually weigh more than 1 t and are installed by being suspended by a crane, so it would be time-consuming and costly for inexperienced workers to train on-site. In this paper, therefore, a block installation simulator is developed based on Unity 3D, a game development engine. The simulator computes porosity. Porosity is defined as the ratio of the total volume of the wave breaker blocks inside the structure to the final shape of the ideal structure. Using the evaluation of porosity, the simulator can determine how well the user is able to install the blocks. The voxelization technique is used to calculate the porosity of the structure, simplifying the calculations. Other techniques, such as raycasting and box overlapping, are employed for accurate simulation. In the near future, the simulator will install an automatic block installation algorithm based on combinatorial optimization solutions and compare the user-demonstrated block installation and the appropriate installation solved by the algorithm.

Keywords: 3D simulator, porosity, user interface, voxelization, wave-dissipating blocks

Procedia PDF Downloads 103
521 Conception of Increasing the Efficiency of Excavation Shoring by Prestressing Diaphragm Walls

Authors: Mateusz Frydrych

Abstract:

The construction of diaphragm walls as excavation shoring as well as part of deep foundations is widely used in geotechnical engineering. Today's design challenges lie in the optimal dimensioning of the cross-section, which is demanded by technological considerations. Also in force is the issue of optimization and sustainable use of construction materials, including reduction of carbon footprint, which is currently a relevant challenge for the construction industry. The author presents the concept of an approach to achieving increased efficiency of diaphragm wall excavation shoring by using structural compression technology. The author proposes to implement prestressed tendons in a non-linear manner in the reinforcement cage. As a result bending moment is reduced, which translates into a reduction in the amount of steel needed in the section, a reduction in displacements, and a reduction in the scratching of the casing, including the achievement of better tightness. This task is rarely seen and has not yet been described in a scientific way in the literature. The author has developed a dynamic numerical model that allows the dimensioning of the cross-section of a prestressed shear wall, as well as the study of casing displacements and cross-sectional forces in any defined computational situation. Numerical software from the Sofistik - open source development environment - was used for the study, and models were validated in Plaxis software . This is an interesting idea that allows for optimizing the execution of construction works and reducing the required resources by using fewer materials and saving time. The author presents the possibilities of a prestressed diaphragm wall, among others, using. The example of a diaphragm wall working as a cantilever at the height of two underground floors without additional strutting or stability protection by using ground anchors. This makes the execution of the work more criminal for the contractor and, as a result, cheaper for the investor.

Keywords: prestressed diaphragm wall, Plaxis, Sofistik, innovation, FEM, optimisation

Procedia PDF Downloads 77