Search results for: structural design optimization
15147 Computational Fluid Dynamics Analysis and Optimization of the Coanda Unmanned Aerial Vehicle Platform
Authors: Nigel Q. Kelly, Zaid Siddiqi, Jin W. Lee
Abstract:
It is known that using Coanda aerosurfaces can drastically augment the lift forces when applied to an Unmanned Aerial Vehicle (UAV) platform. However, Coanda saucer UAVs, which commonly use a dish-like, radially-extending structure, have shown no significant increases in thrust/lift force and therefore have never been commercially successful: the additional thrust/lift generated by the Coanda surface diminishes since the airstreams emerging from the rotor compartment expand radially causing serious loss of momentums and therefore a net loss of total thrust/lift. To overcome this technical weakness, we propose to examine a Coanda surface of straight, cylindrical design and optimize its geometry for highest thrust/lift utilizing computational fluid dynamics software ANSYS Fluent®. The results of this study reveal that a Coanda UAV configured with 4 sides of straight, cylindrical Coanda surface achieve an overall 45% increase in lift compared to conventional Coanda Saucer UAV configurations. This venture integrates with an ongoing research project where a Coanda prototype is being assembled. Additionally, a custom thrust-stand has been constructed for thrust/lift measurement.Keywords: CFD, Coanda, lift, UAV
Procedia PDF Downloads 14315146 The Role of Acoustical Design within Architectural Design in the Early Design Phase
Authors: O. Wright, N. Perkins, M. Donn, M. Halstead
Abstract:
This research responded to anecdotal evidence that suggested inefficiencies within the Architect and Acoustician relationship may lead to ineffective acoustic design decisions. The acoustician spoken to believed that he was approached too late in the design phase. The approached architect valued acoustical qualities, yet, struggled to interpret common measurement parameters. The preliminary investigation of these opinions indicated a gap in the current New Zealand Architectural discourse and currently informs the creation of a 2016 Master of Architecture (Prof) thesis research. Little meaningful information about acoustic intervention in the early design phase could be found from past literature. In the information that was sourced, authors focus on software as an incorporation tool without investigating why the flaws in the relationship originally exist. To further explore this relationship, a survey was designed. It underwent three phases to ensure its consistency, and was delivered to a group of 51 acousticians from one international Acoustics company. The results were then separated between New Zealand and off-shore to identify trends. The survey results suggest that 75% of acousticians meet the architect less than 5 times per project. Instead of regular contact, a mediated method is adopted though a mix of telecommunication and written reports. Acousticians tend to be introduced later into New Zealand building project than the corresponding off-shore building. This delay corresponds to an increase in remedial action for each of the building types in the survey except Auditoria and Office Buildings. 31 participants have had their specifications challenged by an architect. Furthermore, 71% of the acousticians believe that architects do not have the knowledge to understand why the acoustic specifications are in place. The issues raised in this investigation align to the colloquial evidence expressed by the two consultants. It identifies a larger gap in the industry were acoustics is remedially treated rather than identified as a possible design driver. Further research through design is suggested to understand the role of acoustics within architectural design and potential tools for its inclusion during, not after, the design process.Keywords: architectural acoustics, early-design, interdisciplinary communication, remedial response
Procedia PDF Downloads 25515145 Influence of Transverse Steel and Casting Direction on Shear Response and Ductility of Reinforced Ultra High Performance Concrete Beams
Authors: Timothy E. Frank, Peter J. Amaddio, Elizabeth D. Decko, Alexis M. Tri, Darcy A. Farrell, Cole M. Landes
Abstract:
Ultra high performance concrete (UHPC) is a class of cementitious composites with a relatively large percentage of cement generating high compressive strength. Additionally, UHPC contains disbursed fibers, which control crack width, carry the tensile load across narrow cracks, and limit spalling. These characteristics lend themselves to a wide range of structural applications when UHPC members are reinforced with longitudinal steel. Efficient use of fibers and longitudinal steel is required to keep lifecycle cost competitive in reinforced UHPC members; this requires full utilization of both the compressive and tensile qualities of the reinforced cementitious composite. The objective of this study is to investigate the shear response of steel-reinforced UHPC beams to guide design decisions that keep initial costs reasonable, limit serviceability crack widths, and ensure a ductile structural response and failure path. Five small-scale, reinforced UHPC beams were experimentally tested. Longitudinal steel, transverse steel, and casting direction were varied. Results indicate that an increase in transverse steel in short-spanned reinforced UHPC beams provided additional shear capacity and increased the peak load achieved. Beams with very large longitudinal steel reinforcement ratios did not achieve yield and fully utilized the tension properties of the longitudinal steel. Casting the UHPC beams from the end or from the middle affected load-carrying capacity and ductility, but image analysis determined the fiber orientation was not significantly different. It is believed the presence of transverse and longitudinal steel reinforcement minimized the effect of different UHPC casting directions. Results support recent recommendations in the literature suggesting a 1% fiber volume fraction is sufficient within UHPC to prevent spalling and provide compressive fracture toughness under extreme loading conditions.Keywords: fiber orientation, reinforced ultra high performance concrete beams, shear, transverse steel
Procedia PDF Downloads 11315144 The Studies of the Impact of Biomimicry and Sustainability on Urban Design
Authors: Nourhane Mohamed El Haridi, Mostafa El Arabi, Zeyad El Sayad
Abstract:
Biomimicry is defined, by Benyus the natural sciences writer, as imitating or taking inspiration from nature’s forms and processes to solve human problems. Biomimicry is the conscious emulation of life’s genius. As the design community realizes the tremendous impact human constructions have on the world, environmental designers look to new approaches like biomimicry to advance sustainable design. Building leading the declaration made by biomimicry scientists that a full imitation of nature engages form, ecosystem, and process; this paper uses a logic approach to interpret human and environmental wholeness. Designers would benefit from both integrating social theory with environmental thinking and from combining their substantive skills with techniques for getting sustainable biomimic urban design. Integrating biomimicryʹs “Life’s Principles” into a built environment process model will make biomimicry more accessible and thus more widely accepted throughout the industry, and the sustainability of all species will benefit. The Biomimicry Guild hypothesizes the incorporation of these principles, called Lifeʹs Principles, increase the likelihood of sustainability for a respective design, and make it more likely that the design will have a greater impact on sustainability for future generations of all species as mentioned by Benyus in her book. This thesis utilizes Life’s Principles as a foundation for a design process model intended for application on built environment projects at various scales. This paper takes a look at the importance of the integration of biomimicry in urban design to get more sustainable cities and better life, by analyzing the principles of both sustainability and biomimicry, and applying these ideas on futuristic or existing cities to make a biomimic sustainable city more healthier and more conductive to life, and get a better biomimic urban design. A group of experts, architects, biologists, scientists, economists and ecologists should work together to face all the financial and designing difficulties, to have better solutions and good innovative ideas for biomimic sustainable urban design, it is not the only solution, but it is one of the best studies for a better future.Keywords: biomimicry, built environment, sustainability, urban design
Procedia PDF Downloads 52615143 Growth of Non-Polar a-Plane AlGaN Epilayer with High Crystalline Quality and Smooth Surface Morphology
Authors: Abbas Nasir, Xiong Zhang, Sohail Ahmad, Yiping Cui
Abstract:
Non-polar a-plane AlGaN epilayers of high structural quality have been grown on r-sapphire substrate by using metalorganic chemical vapor deposition (MOCVD). A graded non-polar AlGaN buffer layer with variable aluminium concentration was used to improve the structural quality of the non-polar a-plane AlGaN epilayer. The characterisations were carried out by high-resolution X-ray diffraction (HR-XRD), atomic force microscopy (AFM) and Hall effect measurement. The XRD and AFM results demonstrate that the Al-composition-graded non-polar AlGaN buffer layer significantly improved the crystalline quality and the surface morphology of the top layer. A low root mean square roughness 1.52 nm is obtained from AFM, and relatively low background carrier concentration down to 3.9× cm-3 is obtained from Hall effect measurement.Keywords: non-polar AlGaN epilayer, Al composition-graded AlGaN layer, root mean square, background carrier concentration
Procedia PDF Downloads 14415142 The Impact of Technology on Architecture and Graphic Designs
Authors: Feby Zaki Raouf Fawzy
Abstract:
Nowadays, design and architecture are being affected and undergoing change with the rapid advancements in technology, economics, politics, society, and culture. Architecture has been transforming with the latest developments after the inclusion of computers in design. Integration of design into the computational environment has revolutionized architecture and unique perspectives in architecture have been gained. The history of architecture shows the various technological developments and changes in which architecture has transformed with time. Therefore, the analysis of integration between technology and the history of the architectural process makes it possible to build a consensus on the idea of how architecture is to proceed. In this study, each period that occurs with the integration of technology into architecture is addressed within the historical process. At the same time, changes in architecture via technology are identified as important milestones and predictions with regards to the future of architecture have been determined. Developments and changes in technology and the use of technology in architecture within years are analyzed in charts and graphs comparatively. The historical process of architecture and its transformation via technology is supported by a detailed literature review, and they are consolidated with the examination of focal points of 20th-century architecture under the titles parametric design, genetic architecture, simulation, and biomimicry. It is concluded that with the historical research between past and present, the developments in architecture cannot keep up with the advancements in technology, and recent developments in technology overshadow architecture; even technology decides the direction of architecture. As a result, a scenario is presented with regard to the reach of technology in the future of architecture and the role of the architect.Keywords: design and development the information technology architecture, enterprise architecture, enterprise architecture design result, TOGAF architecture development method (ADM)
Procedia PDF Downloads 7015141 A Methodology of Testing Beam to Column Connection under Lateral Impact Load
Authors: A. Al-Rifaie, Z. W. Guan, S. W. Jones
Abstract:
Beam to column connection can be considered as the most important structural part that affects the response of buildings to progressive collapse. However, many studies were conducted to investigate the beam to column connection under accidental loads such as fire, blast and impact load to investigate the connection response. The study is a part of a PhD plan to investigate different types of connections under lateral impact load. The conventional test setups, such as cruciform setup, were designed to apply shear forces and bending moment on the connection, whilst, in the lateral impact case, the connection is subjected to combined tension and moment. Hence, a review is presented to introduce the previous test setup that is used to investigate the connection behaviour. Then, the design and fabrication of the novel test setup is presented. Finally, some trial test results to investigate the efficiency of the proposed setup are discussed. The final results indicate that the setup was efficient in terms of the simplicity and strength.Keywords: connections, impact load, drop hammer, testing methods
Procedia PDF Downloads 30015140 Utilizing Fiber-Based Modeling to Explore the Presence of a Soft Storey in Masonry-Infilled Reinforced Concrete Structures
Authors: Akram Khelaifia, Salah Guettala, Nesreddine Djafar Henni, Rachid Chebili
Abstract:
Recent seismic events have underscored the significant influence of masonry infill walls on the resilience of structures. The irregular positioning of these walls exacerbates their adverse effects, resulting in substantial material and human losses. Research and post-earthquake evaluations emphasize the necessity of considering infill walls in both the design and assessment phases. This study delves into the presence of soft stories in reinforced concrete structures with infill walls. Employing an approximate method relying on pushover analysis results, fiber-section-based macro-modeling is utilized to simulate the behavior of infill walls. The findings shed light on the presence of soft first stories, revealing a notable 240% enhancement in resistance for weak column—strong beam-designed frames due to infill walls. Conversely, the effect is more moderate at 38% for strong column—weak beam-designed frames. Interestingly, the uniform distribution of infill walls throughout the structure's height does not influence soft-story emergence in the same seismic zone, irrespective of column-beam strength. In regions with low seismic intensity, infill walls dissipate energy, resulting in consistent seismic behavior regardless of column configuration. Despite column strength, structures with open-ground stories remain vulnerable to soft first-story emergence, underscoring the crucial role of infill walls in reinforced concrete structural design.Keywords: masonry infill walls, soft Storey, pushover analysis, fiber section, macro-modeling
Procedia PDF Downloads 6815139 Deep Learning Approach to Trademark Design Code Identification
Authors: Girish J. Showkatramani, Arthi M. Krishna, Sashi Nareddi, Naresh Nula, Aaron Pepe, Glen Brown, Greg Gabel, Chris Doninger
Abstract:
Trademark examination and approval is a complex process that involves analysis and review of the design components of the marks such as the visual representation as well as the textual data associated with marks such as marks' description. Currently, the process of identifying marks with similar visual representation is done manually in United States Patent and Trademark Office (USPTO) and takes a considerable amount of time. Moreover, the accuracy of these searches depends heavily on the experts determining the trademark design codes used to catalog the visual design codes in the mark. In this study, we explore several methods to automate trademark design code classification. Based on recent successes of convolutional neural networks in image classification, we have used several different convolutional neural networks such as Google’s Inception v3, Inception-ResNet-v2, and Xception net. The study also looks into other techniques to augment the results from CNNs such as using Open Source Computer Vision Library (OpenCV) to pre-process the images. This paper reports the results of the various models trained on year of annotated trademark images.Keywords: trademark design code, convolutional neural networks, trademark image classification, trademark image search, Inception-ResNet-v2
Procedia PDF Downloads 23415138 Spectroscopic Study of Tb³⁺ Doped Calcium Aluminozincate Phosphor for Display and Solid-State Lighting Applications
Authors: Sumandeep Kaur, Allam Srinivasa Rao, Mula Jayasimhadri
Abstract:
In recent years, rare earth (RE) ions doped inorganic luminescent materials are seeking great attention due to their excellent physical and chemical properties. These materials offer high thermal and chemical stability and exhibit good luminescence properties due to the presence of RE ions. The luminescent properties of these materials are attributed to their intra-configurational f-f transitions in RE ions. A series of Tb³⁺ doped calcium aluminozincate has been synthesized via sol-gel method. The structural and morphological studies have been carried out by recording X-ray diffraction patterns and SEM image. The luminescent spectra have been recorded for a comprehensive study of their luminescence properties. The XRD profile reveals the single-phase orthorhombic crystal structure with an average crystallite size of 65 nm as calculated by using DebyeScherrer equation. The SEM image exhibits completely random, irregular morphology of micron size particles of the prepared samples. The optimization of luminescence has been carried out by varying the dopant Tb³⁺ concentration within the range from 0.5 to 2.0 mol%. The as-synthesized phosphors exhibit intense emission at 544 nm pumped at 478 nm excitation wavelength. The optimized Tb³⁺ concentration has been found to be 1.0 mol% in the present host lattice. The decay curves show bi-exponential fitting for the as-synthesized phosphor. The colorimetric studies show green emission with CIE coordinates (0.334, 0.647) lying in green region for the optimized Tb³⁺ concentration. This report reveals the potential utility of Tb³⁺ doped calcium aluminozincate phosphors for display and solid-state lighting devices.Keywords: concentration quenching, phosphor, photoluminescence, XRD
Procedia PDF Downloads 15515137 Development, Optimization and Characterization of Gastroretentive Multiparticulate Drug Delivery System
Authors: Swapnila V. Vanshiv, Hemant P. Joshi, Atul B. Aware
Abstract:
Current study illustrates the formulation of floating microspheres for purpose of gastroretention of Dipyridamole which shows pH dependent solubility, with the highest solubility in acidic pH. The formulation involved hollow microsphere preparation by using solvent evaporation technique. Concentrations of rate controlling polymer, hydrophilic polymer, internal phase ratio, stirring speed were optimized to get desired responses, namely release of Dipyridamole, buoyancy of microspheres, entrapment efficiency of microspheres. In the formulation, the floating microspheres were prepared by using ethyl cellulose as release retardant and HPMC as a low density hydrophilic swellable polymer. Formulated microspheres were evaluated for their physical properties such as particle size and surface morphology by optical microscopy and SEM. Entrapment efficiency, floating behavior and drug release study as well the formulation was evaluated for in vivo gastroretention in rabbits using gamma scintigraphy. Formulation showed 75% drug release up to 10 hr with entrapment efficiency of 91% and 88% buoyancy till 10 hr. Gamma scintigraphic studies revealed that the optimized system was retained in the gastric region (stomach) for a prolonged period i.e. more than 5 hr.Keywords: Dipyridamole microspheres, gastroretention, HPMC, optimization method
Procedia PDF Downloads 38715136 Density Measurement of Mixed Refrigerants R32+R1234yf and R125+R290 from 0°C to 100°C and at Pressures up to 10 MPa
Authors: Xiaoci Li, Yonghua Huang, Hui Lin
Abstract:
Optimization of the concentration of components in mixed refrigerants leads to potential improvement of either thermodynamic cycle performance or safety performance of heat pumps and refrigerators. R32+R1234yf and R125+R290 are two promising binary mixed refrigerants for the application of heat pumps working in the cold areas. The p-ρ-T data of these mixtures are one of the fundamental and necessary properties for design and evaluation of the performance of the heat pumps. Although the property data of mixtures can be predicted by the mixing models based on the pure substances incorporated in programs such as the NIST database Refprop, direct property measurement will still be helpful to reveal the true state behaviors and verify the models. Densities of the mixtures of R32+R1234yf an d R125+R290 are measured by an Anton Paar U shape oscillating tube digital densimeter DMA-4500 in the range of temperatures from 0°C to 100 °C and pressures up to 10 MPa. The accuracy of the measurement reaches 0.00005 g/cm³. The experimental data are compared with the predictions by Refprop in the corresponding range of pressure and temperature.Keywords: mixed refrigerant, density measurement, densimeter, thermodynamic property
Procedia PDF Downloads 29715135 Design of Wide-Range Variable Fractional-Delay FIR Digital Filters
Authors: Jong-Jy Shyu, Soo-Chang Pei, Yun-Da Huang
Abstract:
In this paper, design of wide-range variable fractional-delay (WR-VFD) finite impulse response (FIR) digital filters is proposed. With respect to the conventional VFD filter which is designed such that its delay is adjustable within one unit, the proposed VFD FIR filter is designed such that its delay can be tunable within a wider range. By the traces of coefficients of the fractional-delay FIR filter, it is found that the conventional method of polynomial substitution for filter coefficients no longer satisfies the design demand, and the circuits perform the sinc function (sinc converter) are added to overcome this problem. In this paper, least-squares method is adopted to design WR-VFD FIR filter. Throughout this paper, several examples will be proposed to demonstrate the effectiveness of the presented methods.Keywords: digital filter, FIR filter, variable fractional-delay (VFD) filter, least-squares approximation
Procedia PDF Downloads 49415134 Integer Programming Model for the Network Design Problem with Facility Dependent Shortest Path Routing
Authors: Taehan Lee
Abstract:
We consider a network design problem which has shortest routing restriction based on the values determined by the installed facilities on each arc. In conventional multicommodity network design problem, a commodity can be routed through any possible path when the capacity is available. But, we consider a problem in which the commodity between two nodes must be routed on a path which has shortest metric value and the link metric value is determined by the installed facilities on the link. By this routing restriction, the problem has a distinct characteristic. We present an integer programming formulation containing the primal-dual optimality conditions to the shortest path routing. We give some computational results for the model.Keywords: integer programming, multicommodity network design, routing, shortest path
Procedia PDF Downloads 42115133 Supervised-Component-Based Generalised Linear Regression with Multiple Explanatory Blocks: THEME-SCGLR
Authors: Bry X., Trottier C., Mortier F., Cornu G., Verron T.
Abstract:
We address component-based regularization of a Multivariate Generalized Linear Model (MGLM). A set of random responses Y is assumed to depend, through a GLM, on a set X of explanatory variables, as well as on a set T of additional covariates. X is partitioned into R conceptually homogeneous blocks X1, ... , XR , viewed as explanatory themes. Variables in each Xr are assumed many and redundant. Thus, Generalised Linear Regression (GLR) demands regularization with respect to each Xr. By contrast, variables in T are assumed selected so as to demand no regularization. Regularization is performed searching each Xr for an appropriate number of orthogonal components that both contribute to model Y and capture relevant structural information in Xr. We propose a very general criterion to measure structural relevance (SR) of a component in a block, and show how to take SR into account within a Fisher-scoring-type algorithm in order to estimate the model. We show how to deal with mixed-type explanatory variables. The method, named THEME-SCGLR, is tested on simulated data.Keywords: Component-Model, Fisher Scoring Algorithm, GLM, PLS Regression, SCGLR, SEER, THEME
Procedia PDF Downloads 39815132 Optimization of Surface Finish in Milling Operation Using Live Tooling via Taguchi Method
Authors: Harish Kumar Ponnappan, Joseph C. Chen
Abstract:
The main objective of this research is to optimize the surface roughness of a milling operation on AISI 1018 steel using live tooling on a HAAS ST-20 lathe. In this study, Taguchi analysis is used to optimize the milling process by investigating the effect of different machining parameters on surface roughness. The L9 orthogonal array is designed with four controllable factors with three different levels each and an uncontrollable factor, resulting in 18 experimental runs. The optimal parameters determined from Taguchi analysis were feed rate – 76.2 mm/min, spindle speed 1150 rpm, depth of cut – 0.762 mm and 2-flute TiN coated high-speed steel as tool material. The process capability Cp and process capability index Cpk values were improved from 0.62 and -0.44 to 1.39 and 1.24 respectively. The average surface roughness values from the confirmation runs were 1.30 µ, decreasing the defect rate from 87.72% to 0.01%. The purpose of this study is to efficiently utilize the Taguchi design to optimize the surface roughness in a milling operation using live tooling.Keywords: live tooling, surface roughness, taguchi analysis, CNC milling operation, CNC turning operation
Procedia PDF Downloads 14315131 Technical Assessment of Utilizing Electrical Variable Transmission Systems in Hybrid Electric Vehicles
Authors: Majid Vafaeipour, Mohamed El Baghdadi, Florian Verbelen, Peter Sergeant, Joeri Van Mierlo, Kurt Stockman, Omar Hegazy
Abstract:
The Electrical Variable Transmission (EVT), an electromechanical device, can be considered as an alternative solution to the conventional transmission system utilized in Hybrid Electric Vehicles (HEVs). This study present comparisons in terms of fuel consumption, power split, and state of charge (SoC) of an HEV containing an EVT to a conventional parallel topology and a series topology. To this end, corresponding simulations of these topologies are all performed in presence of control strategies enabling battery charge-sustaining and efficient power split. The power flow through the components of the vehicle are attained, and fuel consumption results of the considered cases are compared. The investigation of the results indicates utilizing EVT can provide significant added values in HEV configurations. The outcome of the current research paves its path for implementation of design optimization approaches on such systems in further research directions.Keywords: Electrical Variable Transmission (EVT), Hybrid Electric Vehicle (HEV), parallel, series, modeling
Procedia PDF Downloads 23815130 Locomotion Effects of Redundant Degrees of Freedom in Multi-Legged Quadruped Robots
Authors: Hossein Keshavarz, Alejandro Ramirez-Serrano
Abstract:
Energy efficiency and locomotion speed are two key parameters for legged robots; thus, finding ways to improve them are important. This paper proposes a locomotion framework to analyze the energy usage and speed of quadruped robots via a Genetic Algorithm (GA) optimization process. For this, a quadruped robot platform with joint redundancy in its hind legs that we believe will help multi-legged robots improve their speed and energy consumption is used. ContinuO, the quadruped robot of interest, has 14 active degrees of freedom (DoFs), including three DoFs for each front leg, and unlike previously developed quadruped robots, four DoFs for each hind leg. ContinuO aims to realize a cost-effective quadruped robot for real-world scenarios with high speeds and the ability to overcome large obstructions. The proposed framework is used to locomote the robot and analyze its energy consumed at diverse stride lengths and locomotion speeds. The analysis is performed by comparing the obtained results in two modes, with and without the joint redundancy on the robot’s hind legs.Keywords: genetic algorithm optimization, locomotion path planning, quadruped robots, redundant legs
Procedia PDF Downloads 11015129 The Changing Landscape of Fire Safety in Covered Car Parks with the Arrival of Electric Vehicles
Authors: Matt Stallwood, Michael Spearpoint
Abstract:
In 2020, the UK government announced that sales of new petrol and diesel cars would end in 2030, and battery-powered cars made up 1 in 8 new cars sold in 2021 – more than the total from the previous five years. The guidance across the UK for the fire safety design of covered car parks is changing in response to the projected rapid growth in electric vehicle (EV) use. This paper discusses the current knowledge on the fire safety concerns posed by EVs, in particular those powered by lithium-ion batteries, when considering the likelihood of vehicle ignition, fire severity and spread of fire to other vehicles. The paper builds on previous work that has investigated the frequency of fires starting in cars powered by internal combustion engines (ICE), the hazard posed by such fires in covered car parks and the potential for neighboring vehicles to become involved in an incident. Historical data has been used to determine the ignition frequency of ICE car fires, whereas such data is scarce when it comes to EV fires. Should a fire occur, then the fire development has conventionally been assessed to match a ‘medium’ growth rate and to have a 95th percentile peak heat release of 9 MW. The paper examines recent literature in which researchers have measured the burning characteristics of EVs to assess whether these values need to be changed. These findings are used to assess the risk posed by EVs when compared to ICE vehicles. The paper examines what new design guidance is being issued by various organizations across the UK, such as fire and rescue services, insurers, local government bodies and regulators and discusses the impact these are having on the arrangement of parking bays, particularly in residential and mixed-use buildings. For example, the paper illustrates how updated guidance published by the Fire Protection Association (FPA) on the installation of sprinkler systems has increased the hazard classification of parking buildings that can have a considerable impact on the feasibility of a building to meet all its design intents when specifying water supply tanks. Another guidance on the provision of smoke ventilation systems and structural fire resistance is also presented. The paper points to where further research is needed on the fire safety risks posed by EVs in covered car parks. This will ensure that any guidance is commensurate with the need to provide an adequate level of life and property safety in the built environment.Keywords: covered car parks, electric vehicles, fire safety, risk
Procedia PDF Downloads 7415128 The Willingness to Pay of People in Taiwan for Flood Protection Standard of Regions
Authors: Takahiro Katayama, Hsueh-Sheng Chang
Abstract:
Due to the global climate change, it has increased the extreme rainfall that led to serious floods around the world. In recent years, urbanization and population growth also tend to increase the number of impervious surfaces, resulting in significant loss of life and property during floods especially for the urban areas of Taiwan. In the past, the primary governmental response to floods was structural flood control and the only flood protection standards in use were the design standards. However, these design standards of flood control facilities are generally calculated based on current hydrological conditions. In the face of future extreme events, there is a high possibility to surpass existing design standards and cause damages directly and indirectly to the public. To cope with the frequent occurrence of floods in recent years, it has been pointed out that there is a need for a different standard called FPSR (Flood Protection Standard of Regions) in Taiwan. FPSR is mainly used for disaster reduction and used to ensure that hydraulic facilities draining regional flood immediately under specific return period. FPSR could convey a level of flood risk which is useful for land use planning and reflect the disaster situations that a region can bear. However, little has been reported on FPSR and its impacts to the public in Taiwan. Hence, this study proposes a quantity procedure to evaluate the FPSR. This study aimed to examine FPSR of the region and public perceptions of and knowledge about FPSR, as well as the public’s WTP (willingness to pay) for FPSR. The research is conducted via literature review and questionnaire method. Firstly, this study will review the domestic and international research on the FPSR, and provide the theoretical framework of FPSR. Secondly, CVM (Contingent Value Method) has been employed to conduct this survey and using double-bounded dichotomous choice, close-ended format elicits households WTP for raising the protection level to understand the social costs. The samplings of this study are citizens living in Taichung city, Taiwan and 700 samplings were chosen in this study. In the end, this research will continue working on surveys, finding out which factors determining WTP, and provide some recommendations for adaption policies for floods in the future.Keywords: climate change, CVM (Contingent Value Method), FPSR (Flood Protection Standard of Regions), urban flooding
Procedia PDF Downloads 25115127 Hybrid Data-Driven Drilling Rate of Penetration Optimization Scheme Guided by Geological Formation and Historical Data
Authors: Ammar Alali, Mahmoud Abughaban, William Contreras Otalvora
Abstract:
Optimizing the drilling process for cost and efficiency requires the optimization of the rate of penetration (ROP). ROP is the measurement of the speed at which the wellbore is created, in units of feet per hour. It is the primary indicator of measuring drilling efficiency. Maximization of the ROP can indicate fast and cost-efficient drilling operations; however, high ROPs may induce unintended events, which may lead to nonproductive time (NPT) and higher net costs. The proposed ROP optimization solution is a hybrid, data-driven system that aims to improve the drilling process, maximize the ROP, and minimize NPT. The system consists of two phases: (1) utilizing existing geological and drilling data to train the model prior, and (2) real-time adjustments of the controllable dynamic drilling parameters [weight on bit (WOB), rotary speed (RPM), and pump flow rate (GPM)] that direct influence on the ROP. During the first phase of the system, geological and historical drilling data are aggregated. After, the top-rated wells, as a function of high instance ROP, are distinguished. Those wells are filtered based on NPT incidents, and a cross-plot is generated for the controllable dynamic drilling parameters per ROP value. Subsequently, the parameter values (WOB, GPM, RPM) are calculated as a conditioned mean based on physical distance, following Inverse Distance Weighting (IDW) interpolation methodology. The first phase is concluded by producing a model of drilling best practices from the offset wells, prioritizing the optimum ROP value. This phase is performed before the commencing of drilling. Starting with the model produced in phase one, the second phase runs an automated drill-off test, delivering live adjustments in real-time. Those adjustments are made by directing the driller to deviate two of the controllable parameters (WOB and RPM) by a small percentage (0-5%), following the Constrained Random Search (CRS) methodology. These minor incremental variations will reveal new drilling conditions, not explored before through offset wells. The data is then consolidated into a heat-map, as a function of ROP. A more optimum ROP performance is identified through the heat-map and amended in the model. The validation process involved the selection of a planned well in an onshore oil field with hundreds of offset wells. The first phase model was built by utilizing the data points from the top-performing historical wells (20 wells). The model allows drillers to enhance decision-making by leveraging existing data and blending it with live data in real-time. An empirical relationship between controllable dynamic parameters and ROP was derived using Artificial Neural Networks (ANN). The adjustments resulted in improved ROP efficiency by over 20%, translating to at least 10% saving in drilling costs. The novelty of the proposed system lays is its ability to integrate historical data, calibrate based geological formations, and run real-time global optimization through CRS. Those factors position the system to work for any newly drilled well in a developing field event.Keywords: drilling optimization, geological formations, machine learning, rate of penetration
Procedia PDF Downloads 13315126 'Talent Schools' in North Rhine-Westphalia: Aims, Opportunities and Challenges of a 6-Year Study
Authors: Laura Beckmann, Sabrina Rutter, Isabell Van Ackeren, Nina Bremm, Esther Dominique Klein, Kathrin Racherbäumer
Abstract:
Current evidence demonstrates that schools in socially disadvantaged contexts are often characterized by lower school performance and lower educational qualifications among the student body, compared to schools in more privileged socio-spacial contexts. At the same time, national and international findings on schools with structural and social challenges show that certain school and classroom development strategies, as well as human and material resources, can significantly contribute to improved school performance of students. The aim of this contribution is to present a 6-year mixed-methods study (Talent Schools in North Rhine-Westphalia), which is designed as a school experiment addressing the well-acknowledged inequality of educational opportunities in the German school system. Started in the year 2019 and funded by the Ministry for School and Education of the State of North Rhine-Westphalia, the study targets schools in socio-spatially disadvantaged areas, which have increasingly been the focus of both public debate and educational policy. In the German-speaking countries, however, there is little knowledge available on the structure and design of complex strategies for school and classroom development that describe successful approaches to the further development of schools in disadvantaged locations in a process-oriented manner. Given these shortcomings, the present study aims at a longitudinal analysis of school and classroom development processes within 60 ‘talent schools’, whereby concrete micro-progressions within individual schools are documented and aggregated to general processes that may either impede or promote development. The main research question is the following: With the help of which strategies and (teaching) concepts, with which use of resources and with which forms of cooperation can schools contribute to the development of student achievement, including educational qualifications and transition rates in education and employment? Thus, the ‘talent schools’ may serve as examples of how social background can successfully be decoupled from educational success at schools with special structural and procedural challenges. The major chances and challenges of this project will be discussed.Keywords: educational inequality, school development, student achievement, mixed-methods study
Procedia PDF Downloads 12915125 Methaheuristic Bat Algorithm in Training of Feed-Forward Neural Network for Stock Price Prediction
Authors: Marjan Golmaryami, Marzieh Behzadi
Abstract:
Recent developments in stock exchange highlight the need for an efficient and accurate method that helps stockholders make better decision. Since stock markets have lots of fluctuations during the time and different effective parameters, it is difficult to make good decisions. The purpose of this study is to employ artificial neural network (ANN) which can deal with time series data and nonlinear relation among variables to forecast next day stock price. Unlike other evolutionary algorithms which were utilized in stock exchange prediction, we trained our proposed neural network with metaheuristic bat algorithm, with fast and powerful convergence and applied it in stock price prediction for the first time. In order to prove the performance of the proposed method, this research selected a 7 year dataset from Parsian Bank stocks and after imposing data preprocessing, used 3 types of ANN (back propagation-ANN, particle swarm optimization-ANN and bat-ANN) to predict the closed price of stocks. Afterwards, this study engaged MATLAB to simulate 3 types of ANN, with the scoring target of mean absolute percentage error (MAPE). The results may be adapted to other companies stocks too.Keywords: artificial neural network (ANN), bat algorithm, particle swarm optimization algorithm (PSO), stock exchange
Procedia PDF Downloads 55015124 Covariate-Adjusted Response-Adaptive Designs for Semi-Parametric Survival Responses
Authors: Ayon Mukherjee
Abstract:
Covariate-adjusted response-adaptive (CARA) designs use the available responses to skew the treatment allocation in a clinical trial in towards treatment found at an interim stage to be best for a given patient's covariate profile. Extensive research has been done on various aspects of CARA designs with the patient responses assumed to follow a parametric model. However, ranges of application for such designs are limited in real-life clinical trials where the responses infrequently fit a certain parametric form. On the other hand, robust estimates for the covariate-adjusted treatment effects are obtained from the parametric assumption. To balance these two requirements, designs are developed which are free from distributional assumptions about the survival responses, relying only on the assumption of proportional hazards for the two treatment arms. The proposed designs are developed by deriving two types of optimum allocation designs, and also by using a distribution function to link the past allocation, covariate and response histories to the present allocation. The optimal designs are based on biased coin procedures, with a bias towards the better treatment arm. These are the doubly-adaptive biased coin design (DBCD) and the efficient randomized adaptive design (ERADE). The treatment allocation proportions for these designs converge to the expected target values, which are functions of the Cox regression coefficients that are estimated sequentially. These expected target values are derived based on constrained optimization problems and are updated as information accrues with sequential arrival of patients. The design based on the link function is derived using the distribution function of a probit model whose parameters are adjusted based on the covariate profile of the incoming patient. To apply such designs, the treatment allocation probabilities are sequentially modified based on the treatment allocation history, response history, previous patients’ covariates and also the covariates of the incoming patient. Given these information, an expression is obtained for the conditional probability of a patient allocation to a treatment arm. Based on simulation studies, it is found that the ERADE is preferable to the DBCD when the main aim is to minimize the variance of the observed allocation proportion and to maximize the power of the Wald test for a treatment difference. However, the former procedure being discrete tends to be slower in converging towards the expected target allocation proportion. The link function based design achieves the highest skewness of patient allocation to the best treatment arm and thus ethically is the best design. Other comparative merits of the proposed designs have been highlighted and their preferred areas of application are discussed. It is concluded that the proposed CARA designs can be considered as suitable alternatives to the traditional balanced randomization designs in survival trials in terms of the power of the Wald test, provided that response data are available during the recruitment phase of the trial to enable adaptations to the designs. Moreover, the proposed designs enable more patients to get treated with the better treatment during the trial thus making the designs more ethically attractive to the patients. An existing clinical trial has been redesigned using these methods.Keywords: censored response, Cox regression, efficiency, ethics, optimal allocation, power, variability
Procedia PDF Downloads 16515123 Quantifying Automation in the Architectural Design Process via a Framework Based on Task Breakdown Systems and Recursive Analysis: An Exploratory Study
Authors: D. M. Samartsev, A. G. Copping
Abstract:
As with all industries, architects are using increasing amounts of automation within practice, with approaches such as generative design and use of AI becoming more commonplace. However, the discourse on the rate at which the architectural design process is being automated is often personal and lacking in objective figures and measurements. This results in confusion between people and barriers to effective discourse on the subject, in turn limiting the ability of architects, policy makers, and members of the public in making informed decisions in the area of design automation. This paper proposes the use of a framework to quantify the progress of automation within the design process. The use of a reductionist analysis of the design process allows it to be quantified in a manner that enables direct comparison across different times, as well as locations and projects. The methodology is informed by the design of this framework – taking on the aspects of a systematic review but compressed in time to allow for an initial set of data to verify the validity of the framework. The use of such a framework of quantification enables various practical uses such as predicting the future of the architectural industry with regards to which tasks will be automated, as well as making more informed decisions on the subject of automation on multiple levels ranging from individual decisions to policy making from governing bodies such as the RIBA. This is achieved by analyzing the design process as a generic task that needs to be performed, then using principles of work breakdown systems to split the task of designing an entire building into smaller tasks, which can then be recursively split further as required. Each task is then assigned a series of milestones that allow for the objective analysis of its automation progress. By combining these two approaches it is possible to create a data structure that describes how much various parts of the architectural design process are automated. The data gathered in the paper serves the dual purposes of providing the framework with validation, as well as giving insights into the current situation of automation within the architectural design process. The framework can be interrogated in many ways and preliminary analysis shows that almost 40% of the architectural design process has been automated in some practical fashion at the time of writing, with the rate at which progress is made slowly increasing over the years, with the majority of tasks in the design process reaching a new milestone in automation in less than 6 years. Additionally, a further 15% of the design process is currently being automated in some way, with various products in development but not yet released to the industry. Lastly, various limitations of the framework are examined in this paper as well as further areas of study.Keywords: analysis, architecture, automation, design process, technology
Procedia PDF Downloads 10615122 Product Line Design with Customization in the Presence of Demand Uncertainty
Authors: Parisa Bagheri Tookanlou
Abstract:
In this paper, we analyze a product line design problem faced by a manufacturing firm where the product line consists of a customized product in addition to a standard product and is offered in a market in which customers are heterogeneous on aesthetic attributes of the product. The customization level of a product is defined by the fraction of aesthetic attributes of the product that the manufacturer chooses to customize. In contrast to the existing literature on product line design that predominantly assumes deterministic demand, we consider the presence of demand uncertainty and frame the product line design problem in a single period (news vendor) setting. We examine the effect of demand uncertainty on product line decisions. Furthermore, we also examine how product line decisions are influenced by channel structure. While we use the centralized channel as a benchmark, we consider the decentralized dual channel where the customized product is sold through an online channel owned by the manufacturer and the standard product is sold through a retailer. We introduce a supply contract between the manufacturer and the retailer for improving channel efficiency and coordinate the distribution channel.Keywords: product line design, demand uncertainty, customization level, distribution channel
Procedia PDF Downloads 18715121 Analytics Capabilities and Employee Role Stressors: Implications for Organizational Performance
Authors: Divine Agozie, Muesser Nat, Eric Afful-Dadzie
Abstract:
This examination attempts an analysis of the effect of business intelligence and analytics (BI&A) capabilities on organizational role stressors and the implications of such an effect on performance. Two hundred twenty-eight responses gathered from seventy-six firms across Ghana were analyzed using the Partial Least Squares Structural Equation Modelling (PLS-SEM) approach to validate the hypothesized relationships identified in the research model. Findings suggest both endogenous and exogenous dependencies of the sensing capability on the multiple role requirements of personnel. Further, transforming capability increases role conflict, whereas driving capability of BI&A systems impacts role conflict and role ambiguity. This study poses many practical insights to firms seeking to acquire analytics capabilities to drive performance and data-driven decision-making. It is important for firms to consider balancing role changes and task requirements before implementing and post-implementation stages of BI&A innovations.Keywords: business intelligence and analytics, dynamic capabilities view, organizational stressors, structural equation modelling
Procedia PDF Downloads 11415120 Analysis and Re-Design Ergonomic Mineral Water Gallon Trolley
Authors: Dessy Laksyana Utami
Abstract:
Manual material handling activities often make it difficult for humans to work like this. Muscle injury due to incorrect posture.Workers need to facilitate their activities. One tool to assist their activities in the transportation of ordinary materials is a trolley. This tool is very useful because it can be used.It can bring many items without having to spend more energy to operate it. Very Comfortable used a trolley in the community. But the old design still have a complaint by worker, because lack of grip and capacity. After posture analysis with the REBA method, the value of risk need to be increased is obtained tool. Re design use Indonesian anthropometric data with the 50th percentile.Keywords: Material Handling, REBA method, postural assessment, Trolley.
Procedia PDF Downloads 13815119 Internet Optimization by Negotiating Traffic Times
Authors: Carlos Gonzalez
Abstract:
This paper describes a system to optimize the use of the internet by clients requiring downloading of videos at peak hours. The system consists of a web server belonging to a provider of video contents, a provider of internet communications and a software application running on a client’s computer. The client using the application software will communicate to the video provider a list of the client’s future video demands. The video provider calculates which videos are going to be more in demand for download in the immediate future, and proceeds to request the internet provider the most optimal hours to do the downloading. The times of the downloading will be sent to the application software, which will use the information of pre-established hours negotiated between the video provider and the internet provider to download those videos. The videos will be saved in a special protected section of the user’s hard disk, which will only be accessed by the application software in the client’s computer. When the client is ready to see a video, the application will search the list of current existent videos in the area of the hard disk; if it does exist, it will use this video directly without the need for internet access. We found that the best way to optimize the download traffic of videos is by negotiation between the internet communication provider and the video content provider.Keywords: internet optimization, video download, future demands, secure storage
Procedia PDF Downloads 13815118 Virtual Reality Design Platform to Easily Create Virtual Reality Experiences
Authors: J. Casteleiro- Pitrez
Abstract:
The interest in Virtual Reality (VR) keeps increasing among the community of designers. To develop this type of immersive experience, the understanding of new processes and methodologies is as fundamental as its complex implementation which usually implies hiring a specialized team. In this paper, we introduce a case study, a platform that allows designers to easily create complex VR experiences, present its features, and its development process. We conclude that this platform provides a complete solution for the design and development of VR experiences, no-code needed.Keywords: creatives, designers, virtual reality, virtual reality design platform, virtual reality system, no-coding
Procedia PDF Downloads 160