Search results for: asymptotic complexity
1490 The Process of Crisis: Model of Its Development in the Organization
Authors: M. Mikušová
Abstract:
The main aim of this paper is to present a clear and comprehensive picture of the process of a crisis in the organization which will help to better understand its possible developments. For a description of the sequence of individual steps and an indication of their causation and possible variants of the developments, a detailed flow diagram with verbal comment is applied. For simplicity, the process of the crisis is observed in four basic phases called: symptoms of the crisis, diagnosis, action and prevention. The model highlights the complexity of the phenomenon of the crisis and that the various phases of the crisis are interweaving.Keywords: crisis, management, model, organization
Procedia PDF Downloads 2891489 Demand-Oriented Supplier Integration in Agile New Product Development Projects
Authors: Guenther Schuh, Stephan Schroeder, Marcel Faulhaber
Abstract:
Companies are facing an increasing pressure to innovate faster, cheaper and more radical in last years, due to shrinking product lifecycles and higher volatility of markets and customer demands. Especially established companies struggle meeting those demands. Thus, many producing companies are adapting their development processes to address this increasing pressure. One approach taken by many companies is the use of agile, highly iterative development processes to reduce development times and costs as well as to increase the fulfilled customer requirements and the realized level of innovation. At the same time decreasing depths of added value and increasing focus on core competencies as well as a growing product complexity result in a high dependency on suppliers and external development partners during the product development. Thus, a successful introduction of agile development methods into the development of physical products requires also a successful integration of the necessary external partners and suppliers into the new processes and procedures and an adaption of the organizational interfaces to external partners according to the new circumstances and requirements of agile development processes. For an effective and efficient product development, the design of customer-supplier-relationships should be demand-oriented. A significant influence on the required design has the characteristics of the procurement object. Examples therefore are the complexity of technical interfaces between supply object and final product or the importance of the supplied component for the major product functionalities. Thus, this paper presents an approach to derive general requirements on the design of supplier integration according to the characteristics of supply objects. First, therefore the most relevant evaluation criteria and characteristics have been identified based on a thorough literature review. Subsequently the resulting requirements on the design of the supplier integration were derived depending on the different possible values of these criteria.Keywords: iterative development processes, agile new product development, procurement, supplier integration
Procedia PDF Downloads 1711488 Linking Temporal Changes of Climate Factors with Staple Cereal Yields in Southern Burkina Faso
Authors: Pius Borona, Cheikh Mbow, Issa Ouedraogo
Abstract:
In the Sahel, climate variability has been associated with a complex web of direct and indirect impacts. This natural phenomenon has been an impediment to agro-pastoral communities who experience uncertainty while involving in farming activities which is also their key source of livelihood. In this scenario, the role of climate variability in influencing the performance, quantity and quality of staple cereals yields, vital for food and nutrition security has been a topic of importance. This response of crops and subsequent yield variability is also a subject of immense debate due to the complexity of crop development at different stages. This complexity is further compounded by influence of slowly changing non-climatic factors. With these challenges in mind, the present paper initially explores the occurrence of climate variability at an inter annual and inter decadal level in South Burkina Faso. This is evidenced by variation of the total annual rainfall and the number of rainy days among other climatic descriptors. Further, it is shown how district-scale cereal yields in the study area including maize, sorghum and millet casually associate variably to the inter-annual variation of selected climate variables. Statistical models show that the three cereals widely depict sensitivity to the length of the growing period and total dry days in the growing season. Maize yields on the other hand relate strongly to the rainfall amount variation (R2=51.8%) showing high moisture dependence during critical growth stages. Our conclusions emphasize on adoption of efficient water utilization platforms especially those that have evidently increased yields and strengthening of forecasts dissemination.Keywords: climate variability, cereal yields, seasonality, rain fed farming, Burkina Faso, rainfall
Procedia PDF Downloads 2021487 A Framework for Assessing and Implementing Ecological-Based Adaptation Solutions in Urban Areas of Shanghai
Authors: Xin Li
Abstract:
The uncertainty and the complexity of the urban environment combining with the threat of climate change are contributing factors to the vulnerability in multiple-dimensions in Chinese megacities, especially in Shanghai. The urban area occupied high valuable technological infrastructure and density buildings is under the threats of climate change and can provide insufficient ecological service to remain the trade-off on urban sustainable development. Urban ecological-based adaptation (UEbA) combines practices and theoretical work and integrates ecological services into multiple-layers of urban environment planning in order to reduce the impact of the complexity and uncertainty. To understand and to respond to the challenges in the urban level, this paper considers Shanghai as the research objective. It is necessary that its urban adaptation strategies should be reflected and contain the concept and knowledge of EbA. In this paper, we firstly use software to illustrates the visualizing patterns and trends of UEBA research in the current 10 years. Specifically, Citespace software was used for interpreting the significant hubs, landmarks points of peer-reviewed literature on the context of ecological service research in recent 10 years. Secondly, 135 evidence-based EbA literature were reviewed for categorizing the methodologies and framework of evidence-based EbA by the systematic map protocol. Finally, a conceptual framework combined with culture, economic and social components was developed in order to assess the current adaptation strategies in Shanghai. This research founds that the key to reducing urban vulnerability does not only focus on co-benefit arguments but also should pay more attention to the concept of trade-off. This research concludes that the designed framework can provide key knowledge and indicates the essential gap as a valuable tool against climate variability in the process of urban adaptation in Shanghai.Keywords: urban ecological-based adaptation, climate change, sustainable development, climate variability
Procedia PDF Downloads 1541486 Black-Hole Dimension: A Distinct Methodology of Understanding Time, Space and Data in Architecture
Authors: Alp Arda
Abstract:
Inspired by Nolan's ‘Interstellar’, this paper delves into speculative architecture, asking, ‘What if an architect could traverse time to study a city?’ It unveils the ‘Black-Hole Dimension,’ a groundbreaking concept that redefines urban identities beyond traditional boundaries. Moving past linear time narratives, this approach draws from the gravitational dynamics of black holes to enrich our understanding of urban and architectural progress. By envisioning cities and structures as influenced by black hole-like forces, it enables an in-depth examination of their evolution through time and space. The Black-Hole Dimension promotes a temporal exploration of architecture, treating spaces as narratives of their current state interwoven with historical layers. It advocates for viewing architectural development as a continuous, interconnected journey molded by cultural, economic, and technological shifts. This approach not only deepens our understanding of urban evolution but also empowers architects and urban planners to create designs that are both adaptable and resilient. Echoing themes from popular culture and science fiction, this methodology integrates the captivating dynamics of time and space into architectural analysis, challenging established design conventions. The Black-Hole Dimension champions a philosophy that welcomes unpredictability and complexity, thereby fostering innovation in design. In essence, the Black-Hole Dimension revolutionizes architectural thought by emphasizing space-time as a fundamental dimension. It reimagines our built environments as vibrant, evolving entities shaped by the relentless forces of time, space, and data. This groundbreaking approach heralds a future in architecture where the complexity of reality is acknowledged and embraced, leading to the creation of spaces that are both responsive to their temporal context and resilient against the unfolding tapestry of time.Keywords: black-hole, timeline, urbanism, space and time, speculative architecture
Procedia PDF Downloads 691485 Non Linear Stability of Non Newtonian Thin Liquid Film Flowing down an Incline
Authors: Lamia Bourdache, Amar Djema
Abstract:
The effect of non-Newtonian property (power law index n) on traveling waves of thin layer of power law fluid flowing over an inclined plane is investigated. For this, a simplified second-order two-equation model (SM) is used. The complete model is second-order four-equation (CM). It is derived by combining the weighted residual integral method and the lubrication theory. This is due to the fact that at the beginning of the instability waves, a very small number of waves is observed. Using a suitable set of test functions, second order terms are eliminated from the calculus so that the model is still accurate to the second order approximation. Linear, spatial, and temporal stabilities are studied. For travelling waves, a particular type of wave form that is steady in a moving frame, i.e., that travels at a constant celerity without changing its shape is studied. This type of solutions which are characterized by their celerity exists under suitable conditions, when the widening due to dispersion is balanced exactly by the narrowing effect due to the nonlinearity. Changing the parameter of celerity in some range allows exploring the entire spectrum of asymptotic behavior of these traveling waves. The (SM) model is converted into a three dimensional dynamical system. The result is that the model exhibits bifurcation scenarios such as heteroclinic, homoclinic, Hopf, and period-doubling bifurcations for different values of the power law index n. The influence of the non-Newtonian parameter on the nonlinear development of these travelling waves is discussed. It is found at the end that the qualitative characters of bifurcation scenarios are insensitive to the variation of the power law index.Keywords: inclined plane, nonlinear stability, non-Newtonian, thin film
Procedia PDF Downloads 2831484 Influence of Shading on a BIPV System’s Performance in an Urban Context: Case Study of BIPV Systems of the Science Center of Complexity Building of the National and Autonomous University of Mexico in Mexico City
Authors: Viridiana Edith Ardura Perea, José Luis Bermúdez Alcocer
Abstract:
The purpose of this paper is to establish the influence of shading on a Building Integrated Photovoltaic (BIPV) system´s performance in an urban context. The PV systems of the Science Center of Complexity (Centro de Ciencias de la Complejidad) Building based in the Main Campus of the National and Autonomous University of Mexico (UNAM) in Mexico City was taken as case study. The PV systems are placed on the rooftop and on the south façade of the building. The south-façade PV system, operating as sunshades, consists of two strings: one at the ground floor and the other one at the first floor. According to the building’s facility manager, the south-façade PV system generates 42% less electricity per kilowatt peak (kWp) installed than the one on the roof. The methods applied in this study were Solar Radiation Analysis (SRA) simulations performed with the Insight 360 Plug-in from Revit 2018® and an on-site measurement using specialized tools. The results of the SRA simulations showed that the shading casted by the PV system placed on the first floor on top of the PV system of the ground floor decreases its solar incident radiation over 50%. The simulation outcome was compared and validated to the measured data obtained from the on-site measurement. In conclusion, the loss factor achieved from the shading of the PVs is due to the surroundings and the PV system´s own design. The south-façade BIPV system’s deficient design generates critical losses on its performance and decreases its profitability.Keywords: building integrated photovoltaics design, energy analysis software, shading losses, solar radiation analysis
Procedia PDF Downloads 1771483 An Approach to Automate the Modeling of Life Cycle Inventory Data: Case Study on Electrical and Electronic Equipment Products
Authors: Axelle Bertrand, Tom Bauer, Carole Charbuillet, Martin Bonte, Marie Voyer, Nicolas Perry
Abstract:
The complexity of Life Cycle Assessment (LCA) can be identified as the ultimate obstacle to massification. Due to these obstacles, the diffusion of eco-design and LCA methods in the manufacturing sectors could be impossible. This article addresses the research question: How to adapt the LCA method to generalize it massively and improve its performance? This paper aims to develop an approach for automating LCA in order to carry out assessments on a massive scale. To answer this, we proceeded in three steps: First, an analysis of the literature to identify existing automation methods. Given the constraints of large-scale manual processing, it was necessary to define a new approach, drawing inspiration from certain methods and combining them with new ideas and improvements. In a second part, our development of automated construction is presented (reconciliation and implementation of data). Finally, the LCA case study of a conduit is presented to demonstrate the feature-based approach offered by the developed tool. A computerized environment supports effective and efficient decision-making related to materials and processes, facilitating the process of data mapping and hence product modeling. This method is also able to complete the LCA process on its own within minutes. Thus, the calculations and the LCA report are automatically generated. The tool developed has shown that automation by code is a viable solution to meet LCA's massification objectives. It has major advantages over the traditional LCA method and overcomes the complexity of LCA. Indeed, the case study demonstrated the time savings associated with this methodology and, therefore, the opportunity to increase the number of LCA reports generated and, therefore, to meet regulatory requirements. Moreover, this approach also presents the potential of the proposed method for a wide range of applications.Keywords: automation, EEE, life cycle assessment, life cycle inventory, massively
Procedia PDF Downloads 881482 An Empirical Study on the Integration of Listening and Speaking Activities with Writing Instruction for Middles School English Language Learners
Authors: Xueyan Hu, Liwen Chen, Weilin He, Sujie Peng
Abstract:
Writing is an important but challenging skill For English language learners. Due to the small amount of time allocated for writing classes at schools, students have relatively few opportunities to practice writing in the classroom. While the practice of integrating listening and speaking activates with writing instruction has been used for adult English language learners, its application for young English learners has seldom been examined due to the challenge of listening and speaking activities for young English language learners. The study attempted to integrating integrating listening and speaking activities with writing instruction for middle school English language learners so as to improving their writing achievements and writing abilities in terms of the word use, coherence, and complexity in their writings. Guided by Gagne's information processing learning theory and memetics, this study conducted a 8-week writing instruction with an experimental class (n=44) and a control class (n=48) . Students in the experimental class participated in a series of listening and retelling activities about a writing sample the teacher used for writing instruction during each period of writing class. Students in the control class were taught traditionally with teachers’ direction instruction using the writing sample. Using the ANCOVA analysis of the scores of students’ writing, word-use, Chinese-English translation and the text structure, this study showed that the experimental writing instruction can significantly improve students’ writing performance. Compared with the students in the control class, the students in experimental class had significant better performance in word use and complexity in their essays. This study provides useful enlightenment for the teaching of English writing for middle school English language learners. Teachers can skillfully use information technology to integrate listening, speaking, and writing teaching, considering students’ language input and output. Teachers need to select suitable and excellent composition templates for students to ensure their high-quality language input.Keywords: wring instruction, retelling, English language learners, listening and speaking
Procedia PDF Downloads 811481 Complexity in a Leslie-Gower Delayed Prey-Predator Model
Authors: Anuraj Singh
Abstract:
The complex dynamics is explored in a prey predator system with multiple delays. The predator dynamics is governed by Leslie-Gower scheme. The existence of periodic solutions via Hopf bifurcation with respect to delay parameters is established. To substantiate analytical findings, numerical simulations are performed. The system shows rich dynamic behavior including chaos and limit cycles.Keywords: chaos, Hopf bifurcation, stability, time delay
Procedia PDF Downloads 3241480 Quantification and Evaluation of Tumors Heterogeneity Utilizing Multimodality Imaging
Authors: Ramin Ghasemi Shayan, Morteza Janebifam
Abstract:
Tumors are regularly inhomogeneous. Provincial varieties in death, metabolic action, multiplication and body part are watched. There’s expanding proof that strong tumors may contain subpopulations of cells with various genotypes and phenotypes. These unmistakable populaces of malignancy cells can connect during a serious way and may contrast in affectability to medications. Most tumors show organic heterogeneity1–3 remembering heterogeneity for genomic subtypes, varieties inside the statement of development variables and genius, and hostile to angiogenic factors4–9 and varieties inside the tumoural microenvironment. These can present as contrasts between tumors in a few people. for instance, O6-methylguanine-DNA methyltransferase, a DNA fix compound, is hushed by methylation of the quality advertiser in half of glioblastoma (GBM), adding to chemosensitivity, and improved endurance. From the outset, there includes been specific enthusiasm inside the usage of dissemination weighted imaging (DWI) and dynamic complexity upgraded MRI (DCE-MRI). DWI sharpens MRI to water dispersion inside the extravascular extracellular space (EES) and is wiped out with the size and setup of the cell populace. Additionally, DCE-MRI utilizes dynamic obtaining of pictures during and after the infusion of intravenous complexity operator. Signal changes are additionally changed to outright grouping of differentiation permitting examination utilizing pharmacokinetic models. PET scan modality gives one of a kind natural particularity, permitting dynamic or static imaging of organic atoms marked with positron emanating isotopes (for example, 15O, 18F, 11C). The strategy is explained to a colossal radiation portion, which points of confinement rehashed estimations, particularly when utilized together with PC tomography (CT). At long last, it's of incredible enthusiasm to quantify territorial hemoglobin state, which could be joined with DCE-CT vascular physiology estimation to create significant experiences for understanding tumor hypoxia.Keywords: heterogeneity, computerized tomography scan, magnetic resonance imaging, PET
Procedia PDF Downloads 1451479 Evaluating the Effect of Spatial Qualities, Openness and Complexity, on Human Cognitive Performance within Virtual Reality
Authors: Pierre F. Gerard, Frederic F. Leymarie, William Latham
Abstract:
Architects have developed a series of objective evaluations, using spatial analysis tools such as Isovist, that show how certain spatial qualities are beneficial to specific human activities hosted in the built environments. In return, they can build more adapted environments by tuning those spatial qualities in their design. In parallel, virtual reality technologies have been developed by engineers with the dream of creating a system that immerses users in a new form of spatial experiences. They already have demonstrated a useful range of benefits not only in simulating critical events to assist people in acquiring new skills, but also to enhance memory retention, to name just a few. This paper investigates the effects of two spatial qualities, openness, and complexity, on cognitive performance within immersive virtual environments. Isovist measure is used to design a series of room settings with different levels of each spatial qualities. In an empirical study, each room was then used by every participant to solve a navigational puzzle game and give a rating of their spatial experience. They were then asked to fill in a questionnaire before solving the visual-spatial memory quiz, which addressed how well they remembered the different rooms. Findings suggest that those spatial qualities have an effect on some of the measures, including navigation performance and memory retention. In particular, there is an order effect for the navigation puzzle game. Participants tended to spend a longer time in the complex room settings. Moreover, there is an interaction effect while with more open settings, participants tended to perform better when in a simple setting; however, with more closed settings, participants tended to perform better in a more complex setting. For the visual-spatial memory quiz, participants performed significantly better within the more open rooms. We believe this is a first step in using virtual environments to enhance participant cognitive performances through better use of specific spatial qualities.Keywords: architecture, navigation, spatial cognition, virtual reality
Procedia PDF Downloads 1281478 Optimization of Titanium Leaching Process Using Experimental Design
Authors: Arash Rafiei, Carroll Moore
Abstract:
Leaching process as the first stage of hydrometallurgy is a multidisciplinary system including material properties, chemistry, reactor design, mechanics and fluid dynamics. Therefore, doing leaching system optimization by pure scientific methods need lots of times and expenses. In this work, a mixture of two titanium ores and one titanium slag are used for extracting titanium for leaching stage of TiO2 pigment production procedure. Optimum titanium extraction can be obtained from following strategies: i) Maximizing titanium extraction without selective digestion; and ii) Optimizing selective titanium extraction by balancing between maximum titanium extraction and minimum impurity digestion. The main difference between two strategies is due to process optimization framework. For the first strategy, the most important stage of production process is concerned as the main stage and rest of stages would be adopted with respect to the main stage. The second strategy optimizes performance of more than one stage at once. The second strategy has more technical complexity compared to the first one but it brings more economical and technical advantages for the leaching system. Obviously, each strategy has its own optimum operational zone that is not as same as the other one and the best operational zone is chosen due to complexity, economical and practical aspects of the leaching system. Experimental design has been carried out by using Taguchi method. The most important advantages of this methodology are involving different technical aspects of leaching process; minimizing the number of needed experiments as well as time and expense; and concerning the role of parameter interactions due to principles of multifactor-at-time optimization. Leaching tests have been done at batch scale on lab with appropriate control on temperature. The leaching tank geometry has been concerned as an important factor to provide comparable agitation conditions. Data analysis has been done by using reactor design and mass balancing principles. Finally, optimum zone for operational parameters are determined for each leaching strategy and discussed due to their economical and practical aspects.Keywords: titanium leaching, optimization, experimental design, performance analysis
Procedia PDF Downloads 3691477 A Fast Optimizer for Large-scale Fulfillment Planning based on Genetic Algorithm
Authors: Choonoh Lee, Seyeon Park, Dongyun Kang, Jaehyeong Choi, Soojee Kim, Younggeun Kim
Abstract:
Market Kurly is the first South Korean online grocery retailer that guarantees same-day, overnight shipping. More than 1.6 million customers place an average of 4.7 million orders and add 3 to 14 products into a cart per month. The company has sold almost 30,000 kinds of various products in the past 6 months, including food items, cosmetics, kitchenware, toys for kids/pets, and even flowers. The company is operating and expanding multiple dry, cold, and frozen fulfillment centers in order to store and ship these products. Due to the scale and complexity of the fulfillment, pick-pack-ship processes are planned and operated in batches, and thus, the planning that decides the batch of the customers’ orders is a critical factor in overall productivity. This paper introduces a metaheuristic optimization method that reduces the complexity of batch processing in a fulfillment center. The method is an iterative genetic algorithm with heuristic creation and evolution strategies; it aims to group similar orders into pick-pack-ship batches to minimize the total number of distinct products. With a well-designed approach to create initial genes, the method produces streamlined plans, up to 13.5% less complex than the actual plans carried out in the company’s fulfillment centers in the previous months. Furthermore, our digital-twin simulations show that the optimized plans can reduce 3% of operation time for packing, which is the most complex and time-consuming task in the process. The optimization method implements a multithreading design on the Spring framework to support the company’s warehouse management systems in near real-time, finding a solution for 4,000 orders within 5 to 7 seconds on an AWS c5.2xlarge instance.Keywords: fulfillment planning, genetic algorithm, online grocery retail, optimization
Procedia PDF Downloads 821476 A Universal Approach to Categorize Failures in Production
Authors: Konja Knüppel, Gerrit Meyer, Peter Nyhuis
Abstract:
The increasing interconnectedness and complexity of production processes raise the susceptibility of production systems to failure. Therefore, the ability to respond quickly to failures is increasingly becoming a competitive factor. The research project "Sustainable failure management in manufacturing SMEs" is developing a methodology to identify failures in the production and select preventive and reactive measures in order to correct failures and to establish sustainable failure management systems.Keywords: failure categorization, failure management, logistic performance, production optimization
Procedia PDF Downloads 3721475 Communication Anxiety in Nigerian Students Studying English as a Foreign Language: Evidence from Colleges of Education Sector
Authors: Yasàlu Haruna
Abstract:
In every transaction, the use of language is central regardless of form or complexity if any meaning is expected to be harvested therefrom. Students constituting a population group in the learning landscape of Nigeria occupy a central position with a propensity to excel or otherwise in the context of communication, especially in the learning process and social interaction. The nature or quantum of anxiety or confidence in speaking a second language is not only peculiar to societies where the second language is not an official language but to a degree, the linguistic gap created by adoption and adaptation syndrome manifests in created anxiety or lack of confidence especially where mastery of a spoken language becomes a major challenge. This paper explores the manner in which linguistic complexity and cultural barriers combine to widen the adaptation and adoption gap. In much the same way, typical issues of pronouncement, intonation and accent difficulties are vital variables that explain the root cause of anxiety. Using a combination of primary and secondary sources of data expressed in questionnaires, key informant interviews and other available data, the paper concludes that the non-integration of anxiety possibility into the education delivery framework has left a lot to be needed in cultivating second language speakers among students of Nigerian Colleges of Education. In addition, cultural barriers and the absence of integration interfaces in the course of learning within and outside the classroom contribute to further widening the gap. Again, colleagues/mates/conversation partners' mastery of a second language remains a contributory factor largely due to the quality of the preparatory school system in many parts of the country. The paper recommends that national policies and frameworks must be reviewed to consider integration windows where culture and conversation partner deficiencies can be remedied through educational events such as debates, quizzes and symposia; improvements can be attained while commercial advertisements are tailored towards seeking for adoption of second language in commerce and major cultural activities.Keywords: cultural barriers, integration, college of education and adaptation, second language
Procedia PDF Downloads 901474 Design of Robust and Intelligent Controller for Active Removal of Space Debris
Authors: Shabadini Sampath, Jinglang Feng
Abstract:
With huge kinetic energy, space debris poses a major threat to astronauts’ space activities and spacecraft in orbit if a collision happens. The active removal of space debris is required in order to avoid frequent collisions that would occur. In addition, the amount of space debris will increase uncontrollably, posing a threat to the safety of the entire space system. But the safe and reliable removal of large-scale space debris has been a huge challenge to date. While capturing and deorbiting space debris, the space manipulator has to achieve high control precision. However, due to uncertainties and unknown disturbances, there is difficulty in coordinating the control of the space manipulator. To address this challenge, this paper focuses on developing a robust and intelligent control algorithm that controls joint movement and restricts it on the sliding manifold by reducing uncertainties. A neural network adaptive sliding mode controller (NNASMC) is applied with the objective of finding the control law such that the joint motions of the space manipulator follow the given trajectory. A computed torque control (CTC) is an effective motion control strategy that is used in this paper for computing space manipulator arm torque to generate the required motion. Based on the Lyapunov stability theorem, the proposed intelligent controller NNASMC and CTC guarantees the robustness and global asymptotic stability of the closed-loop control system. Finally, the controllers used in the paper are modeled and simulated using MATLAB Simulink. The results are presented to prove the effectiveness of the proposed controller approach.Keywords: GNC, active removal of space debris, AI controllers, MatLabSimulink
Procedia PDF Downloads 1301473 The Tramway in French Cities: Complication of Public Spaces and Complexity of the Design Process
Authors: Elisa Maître
Abstract:
The redeployment of tram networks in French cities has considerably modified public spaces and the way citizens use them. Above and beyond the image that trams have of contributing to the sustainable urban development, the question of safety for users in these spaces has not been studied much. This study is based on an analysis of use of public spaces laid out for trams, from the standpoint of legibility and safety concerns. The study also examines to what extent the complexity of the design process, with many interactions between numerous and varied players in this process has a role in the genesis of these problems. This work is mainly based on the analysis of links between the uses of these re-designed public spaces (through observations, interviews of users and accident studies) and the analysis of the design conditions and processes of the projects studied (mainly based on interviews with the actors of these projects). Practical analyses were based three points of view: that of the planner, that of the user (based on observations and interviews) and that of the road safety expert. The cities of Montpellier, Marseille and Nice are the three fields of study on which the demonstration of this thesis is based. On part, the results of this study allow showing that the insertion of tram poses some problems complication of public areas of French cities. These complications related to the restructuring of public spaces for the tram, create difficulties of use and safety concerns. On the other hand, interviews depth analyses, fully transcribed, have led us to develop particular dysfunction scenarios in the design process. These elements lead to question the way the legibility and safety of these new forms of public spaces are taken into account. Then, an in-depth analysis of the design processes of public spaces with trams systems would also be a way of better understanding the choices made, the compromises accepted, and the conflicts and constraints at work, weighing on the layout of these spaces. The results presented concerning the impact that spaces laid out for trams have on the difficulty of use, suggest different possibilities for improving the way in which safety for all users is taken into account in designing public spaces.Keywords: public spaces, road layout, users, design process of urban projects
Procedia PDF Downloads 2291472 Analysis of Direct Current Motor in LabVIEW
Authors: E. Ramprasath, P. Manojkumar, P. Veena
Abstract:
DC motors have been widely used in the past centuries which are proudly known as the workhorse of industrial systems until the invention of the AC induction motors which makes a huge revolution in industries. Since then, the use of DC machines have been decreased due to enormous factors such as reliability, robustness and complexity but it lost its fame due to the losses. A new methodology is proposed to construct a DC motor through the simulation in LabVIEW to get an idea about its real time performances, if a change in parameter might have bigger improvement in losses and reliability.Keywords: analysis, characteristics, direct current motor, LabVIEW software, simulation
Procedia PDF Downloads 5501471 A Cooperative Signaling Scheme for Global Navigation Satellite Systems
Authors: Keunhong Chae, Seokho Yoon
Abstract:
Recently, the global navigation satellite system (GNSS) such as Galileo and GPS is employing more satellites to provide a higher degree of accuracy for the location service, thus calling for a more efficient signaling scheme among the satellites used in the overall GNSS network. In that the network throughput is improved, the spatial diversity can be one of the efficient signaling schemes; however, it requires multiple antenna that could cause a significant increase in the complexity of the GNSS. Thus, a diversity scheme called the cooperative signaling was proposed, where the virtual multiple-input multiple-output (MIMO) signaling is realized with using only a single antenna in the transmit satellite of interest and with modeling the neighboring satellites as relay nodes. The main drawback of the cooperative signaling is that the relay nodes receive the transmitted signal at different time instants, i.e., they operate in an asynchronous way, and thus, the overall performance of the GNSS network could degrade severely. To tackle the problem, several modified cooperative signaling schemes were proposed; however, all of them are difficult to implement due to a signal decoding at the relay nodes. Although the implementation at the relay nodes could be simpler to some degree by employing the time-reversal and conjugation operations instead of the signal decoding, it would be more efficient if we could implement the operations of the relay nodes at the source node having more resources than the relay nodes. So, in this paper, we propose a novel cooperative signaling scheme, where the data signals are combined in a unique way at the source node, thus obviating the need of the complex operations such as signal decoding, time-reversal and conjugation at the relay nodes. The numerical results confirm that the proposed scheme provides the same performance in the cooperative diversity and the bit error rate (BER) as the conventional scheme, while reducing the complexity at the relay nodes significantly. Acknowledgment: This work was supported by the National GNSS Research Center program of Defense Acquisition Program Administration and Agency for Defense Development.Keywords: global navigation satellite network, cooperative signaling, data combining, nodes
Procedia PDF Downloads 2791470 Quantifying Uncertainties in an Archetype-Based Building Stock Energy Model by Use of Individual Building Models
Authors: Morten Brøgger, Kim Wittchen
Abstract:
Focus on reducing energy consumption in existing buildings at large scale, e.g. in cities or countries, has been increasing in recent years. In order to reduce energy consumption in existing buildings, political incentive schemes are put in place and large scale investments are made by utility companies. Prioritising these investments requires a comprehensive overview of the energy consumption in the existing building stock, as well as potential energy-savings. However, a building stock comprises thousands of buildings with different characteristics making it difficult to model energy consumption accurately. Moreover, the complexity of the building stock makes it difficult to convey model results to policymakers and other stakeholders. In order to manage the complexity of the building stock, building archetypes are often employed in building stock energy models (BSEMs). Building archetypes are formed by segmenting the building stock according to specific characteristics. Segmenting the building stock according to building type and building age is common, among other things because this information is often easily available. This segmentation makes it easy to convey results to non-experts. However, using a single archetypical building to represent all buildings in a segment of the building stock is associated with loss of detail. Thermal characteristics are aggregated while other characteristics, which could affect the energy efficiency of a building, are disregarded. Thus, using a simplified representation of the building stock could come at the expense of the accuracy of the model. The present study evaluates the accuracy of a conventional archetype-based BSEM that segments the building stock according to building type- and age. The accuracy is evaluated in terms of the archetypes’ ability to accurately emulate the average energy demands of the corresponding buildings they were meant to represent. This is done for the buildings’ energy demands as a whole as well as for relevant sub-demands. Both are evaluated in relation to the type- and the age of the building. This should provide researchers, who use archetypes in BSEMs, with an indication of the expected accuracy of the conventional archetype model, as well as the accuracy lost in specific parts of the calculation, due to use of the archetype method.Keywords: building stock energy modelling, energy-savings, archetype
Procedia PDF Downloads 1531469 A CORDIC Based Design Technique for Efficient Computation of DCT
Authors: Deboraj Muchahary, Amlan Deep Borah Abir J. Mondal, Alak Majumder
Abstract:
A discrete cosine transform (DCT) is described and a technique to compute it using fast Fourier transform (FFT) is developed. In this work, DCT of a finite length sequence is obtained by incorporating CORDIC methodology in radix-2 FFT algorithm. The proposed methodology is simple to comprehend and maintains a regular structure, thereby reducing computational complexity. DCTs are used extensively in the area of digital processing for the purpose of pattern recognition. So the efficient computation of DCT maintaining a transparent design flow is highly solicited.Keywords: DCT, DFT, CORDIC, FFT
Procedia PDF Downloads 4771468 Linearization and Process Standardization of Construction Design Engineering Workflows
Authors: T. R. Sreeram, S. Natarajan, C. Jena
Abstract:
Civil engineering construction is a network of tasks involving varying degree of complexity and streamlining, and standardization is the only way to establish a systemic approach to design. While there are off the shelf tools such as AutoCAD that play a role in the realization of design, the repeatable process in which these tools are deployed often is ignored. The present paper addresses this challenge through a sustainable design process and effective standardizations at all stages in the design workflow. The same is demonstrated through a case study in the context of construction, and further improvement points are highlighted.Keywords: syste, lean, value stream, process improvement
Procedia PDF Downloads 1221467 Approaches to Reduce the Complexity of Mathematical Models for the Operational Optimization of Large-Scale Virtual Power Plants in Public Energy Supply
Authors: Thomas Weber, Nina Strobel, Thomas Kohne, Eberhard Abele
Abstract:
In context of the energy transition in Germany, the importance of so-called virtual power plants in the energy supply continues to increase. The progressive dismantling of the large power plants and the ongoing construction of many new decentralized plants result in great potential for optimization through synergies between the individual plants. These potentials can be exploited by mathematical optimization algorithms to calculate the optimal application planning of decentralized power and heat generators and storage systems. This also includes linear or linear mixed integer optimization. In this paper, procedures for reducing the number of decision variables to be calculated are explained and validated. On the one hand, this includes combining n similar installation types into one aggregated unit. This aggregated unit is described by the same constraints and target function terms as a single plant. This reduces the number of decision variables per time step and the complexity of the problem to be solved by a factor of n. The exact operating mode of the individual plants can then be calculated in a second optimization in such a way that the output of the individual plants corresponds to the calculated output of the aggregated unit. Another way to reduce the number of decision variables in an optimization problem is to reduce the number of time steps to be calculated. This is useful if a high temporal resolution is not necessary for all time steps. For example, the volatility or the forecast quality of environmental parameters may justify a high or low temporal resolution of the optimization. Both approaches are examined for the resulting calculation time as well as for optimality. Several optimization models for virtual power plants (combined heat and power plants, heat storage, power storage, gas turbine) with different numbers of plants are used as a reference for the investigation of both processes with regard to calculation duration and optimality.Keywords: CHP, Energy 4.0, energy storage, MILP, optimization, virtual power plant
Procedia PDF Downloads 1751466 Bounded Rational Heterogeneous Agents in Artificial Stock Markets: Literature Review and Research Direction
Authors: Talal Alsulaiman, Khaldoun Khashanah
Abstract:
In this paper, we provided a literature survey on the artificial stock problem (ASM). The paper began by exploring the complexity of the stock market and the needs for ASM. ASM aims to investigate the link between individual behaviors (micro level) and financial market dynamics (macro level). The variety of patterns at the macro level is a function of the AFM complexity. The financial market system is a complex system where the relationship between the micro and macro level cannot be captured analytically. Computational approaches, such as simulation, are expected to comprehend this connection. Agent-based simulation is a simulation technique commonly used to build AFMs. The paper proceeds by discussing the components of the ASM. We consider the roles of behavioral finance (BF) alongside the traditionally risk-averse assumption in the construction of agent's attributes. Also, the influence of social networks in the developing of agents’ interactions is addressed. Network topologies such as a small world, distance-based, and scale-free networks may be utilized to outline economic collaborations. In addition, the primary methods for developing agents learning and adaptive abilities have been summarized. These incorporated approach such as Genetic Algorithm, Genetic Programming, Artificial neural network and Reinforcement Learning. In addition, the most common statistical properties (the stylized facts) of stock that are used for calibration and validation of ASM are discussed. Besides, we have reviewed the major related previous studies and categorize the utilized approaches as a part of these studies. Finally, research directions and potential research questions are argued. The research directions of ASM may focus on the macro level by analyzing the market dynamic or on the micro level by investigating the wealth distributions of the agents.Keywords: artificial stock markets, market dynamics, bounded rationality, agent based simulation, learning, interaction, social networks
Procedia PDF Downloads 3541465 Bias-Corrected Estimation Methods for Receiver Operating Characteristic Surface
Authors: Khanh To Duc, Monica Chiogna, Gianfranco Adimari
Abstract:
With three diagnostic categories, assessment of the performance of diagnostic tests is achieved by the analysis of the receiver operating characteristic (ROC) surface, which generalizes the ROC curve for binary diagnostic outcomes. The volume under the ROC surface (VUS) is a summary index usually employed for measuring the overall diagnostic accuracy. When the true disease status can be exactly assessed by means of a gold standard (GS) test, unbiased nonparametric estimators of the ROC surface and VUS are easily obtained. In practice, unfortunately, disease status verification via the GS test could be unavailable for all study subjects, due to the expensiveness or invasiveness of the GS test. Thus, often only a subset of patients undergoes disease verification. Statistical evaluations of diagnostic accuracy based only on data from subjects with verified disease status are typically biased. This bias is known as verification bias. Here, we consider the problem of correcting for verification bias when continuous diagnostic tests for three-class disease status are considered. We assume that selection for disease verification does not depend on disease status, given test results and other observed covariates, i.e., we assume that the true disease status, when missing, is missing at random. Under this assumption, we discuss several solutions for ROC surface analysis based on imputation and re-weighting methods. In particular, verification bias-corrected estimators of the ROC surface and of VUS are proposed, namely, full imputation, mean score imputation, inverse probability weighting and semiparametric efficient estimators. Consistency and asymptotic normality of the proposed estimators are established, and their finite sample behavior is investigated by means of Monte Carlo simulation studies. Two illustrations using real datasets are also given.Keywords: imputation, missing at random, inverse probability weighting, ROC surface analysis
Procedia PDF Downloads 4141464 Regulatory Guidelines to Support the Design of Nanosatellite Projects in Mexican Academic Contexts
Authors: Alvaro Armenta-Ramade, Arturo Serrano-Santoyo, Veronica Rojas-Mendizabal, Roberto Conte-Galvan
Abstract:
The availability and affordability of commercial off-the-shell products have brought a major impetus in the development of university projects related to the design, construction and launching of small satellites on a global scale. Universities in emerging economies as well as in least developed countries have been able to develop prototypes of small satellites (cubesats and cansats) with limited budgets. The experience gained in the development of small satellites gives rise to capacity building for designing more complex aerospace systems. This trend has significantly increased the pace and number of aerospace university projects around the world. In the case of Mexico, projects funded by different agencies have been very effective in accelerating the capacity building and technology transfer initiatives in the aerospace ecosystem. However, many of this initiatives have centered their efforts in technology development matters with minimum or no considerations of key regulatory issues related to frequency assignment, management and licensing, as well as launching requirements and measures of mitigation of space debris. These regulatory concerns are fundamental to accomplish successful missions that take into account the complete value chain of an aerospace project. The purpose of this paper is to develop a regulatory framework to support the efforts of educational institutions working on the development of small satellites in Mexico. We base our framework on recommendations from the International Telecommunications Union (ITU), the United Nations Office for Outer Space Affairs (UNOOSA) and other major actors of the Mexican regulatory ecosystem. In order to develop an integrated and cohesive framework, we draw on complexity science to identify the agents, their role and interactions. Our goal is to create a guiding instrument available both in print and online that can also be used in other regions of the worldKeywords: capacity building, complexity science, cubesats, space regulations, small satellites
Procedia PDF Downloads 2581463 Methodology for the Integration of Object Identification Processes in Handling and Logistic Systems
Authors: L. Kiefer, C. Richter, G. Reinhart
Abstract:
The uprising complexity in production systems due to an increasing amount of variants up to customer innovated products leads to requirements that hierarchical control systems are not able to fulfil. Therefore, factory planners can install autonomous manufacturing systems. The fundamental requirement for an autonomous control is the identification of objects within production systems. In this approach an attribute-based identification is focused for avoiding dose-dependent identification costs. Instead of using an identification mark (ID) like a radio frequency identification (RFID)-Tag, an object type is directly identified by its attributes. To facilitate that it’s recommended to include the identification and the corresponding sensors within handling processes, which connect all manufacturing processes and therefore ensure a high identification rate and reduce blind spots. The presented methodology reduces the individual effort to integrate identification processes in handling systems. First, suitable object attributes and sensor systems for object identification in a production environment are defined. By categorising these sensor systems as well as handling systems, it is possible to match them universal within a compatibility matrix. Based on that compatibility further requirements like identification time are analysed, which decide whether the combination of handling and sensor system is well suited for parallel handling and identification within an autonomous control. By analysing a list of more than thousand possible attributes, first investigations have shown, that five main characteristics (weight, form, colour, amount, and position of subattributes as drillings) are sufficient for an integrable identification. This knowledge limits the variety of identification systems and leads to a manageable complexity within the selection process. Besides the procedure, several tools, as an example a sensor pool are presented. These tools include the generated specific expert knowledge and simplify the selection. The primary tool is a pool of preconfigured identification processes depending on the chosen combination of sensor and handling device. By following the defined procedure and using the created tools, even laypeople out of other scientific fields can choose an appropriate combination of handling devices and sensors which enable parallel handling and identification.Keywords: agent systems, autonomous control, handling systems, identification
Procedia PDF Downloads 1751462 Exploring the Applications of Neural Networks in the Adaptive Learning Environment
Authors: Baladitya Swaika, Rahul Khatry
Abstract:
Computer Adaptive Tests (CATs) is one of the most efficient ways for testing the cognitive abilities of students. CATs are based on Item Response Theory (IRT) which is based on item selection and ability estimation using statistical methods of maximum information selection/selection from posterior and maximum-likelihood (ML)/maximum a posteriori (MAP) estimators respectively. This study aims at combining both classical and Bayesian approaches to IRT to create a dataset which is then fed to a neural network which automates the process of ability estimation and then comparing it to traditional CAT models designed using IRT. This study uses python as the base coding language, pymc for statistical modelling of the IRT and scikit-learn for neural network implementations. On creation of the model and on comparison, it is found that the Neural Network based model performs 7-10% worse than the IRT model for score estimations. Although performing poorly, compared to the IRT model, the neural network model can be beneficially used in back-ends for reducing time complexity as the IRT model would have to re-calculate the ability every-time it gets a request whereas the prediction from a neural network could be done in a single step for an existing trained Regressor. This study also proposes a new kind of framework whereby the neural network model could be used to incorporate feature sets, other than the normal IRT feature set and use a neural network’s capacity of learning unknown functions to give rise to better CAT models. Categorical features like test type, etc. could be learnt and incorporated in IRT functions with the help of techniques like logistic regression and can be used to learn functions and expressed as models which may not be trivial to be expressed via equations. This kind of a framework, when implemented would be highly advantageous in psychometrics and cognitive assessments. This study gives a brief overview as to how neural networks can be used in adaptive testing, not only by reducing time-complexity but also by being able to incorporate newer and better datasets which would eventually lead to higher quality testing.Keywords: computer adaptive tests, item response theory, machine learning, neural networks
Procedia PDF Downloads 1731461 Sustainability through Resilience: How Emergency Responders Cope with Stressors
Authors: Sophie Kroeling, Agnetha Schuchardt
Abstract:
Striving for sustainability brings a lot of challenges for different fields of interest, e. g. security or health concerns. In Germany, civil protection is predominantly carried out by emergency responders who perform essential tasks of civil protection. Based on theoretical concepts of different psychological stress theories this contribution focuses on the question, how the resilience of emergency responders can be improved. The goal is to identify resources and successful coping strategies that help to prevent and reduce negative outcomes during or after stressful events. The paper will present results from a qualitative analysis of semi-structured qualitative interviews with 20 emergency responders. These results provide insights into the complexity of coping processes (e. g. controlling the situation, downplaying perceived personal threats through humor) and show the diversity of stressors (like complexity of the disastrous situation, intrusive press and media, or lack of social support within the organization). Self-efficacy expectation was a very important resource for coping with stressful situations. The results served as a starting point for a quantitative survey (that was conducted in March 2017), the development of education and training tools for emergency responders and the improvement of critical incident stress management processes. First results from the quantitative study with more than 700 participants show that, e. g., the emergency responders use social coping within their private social network and also within their aid organization and that both are correlated to resilience. Moreover, missing information, bureaucratic problems and social conflicts within the organization are events that the majority of the participants considered very onerous. Further results from regression analysis will be presented. The proposed paper will combine findings from the qualitative study with the quantitative results, illustrating figures and correlations with respective statements from the interviews. At the end, suggestions for the improvement of the emergency responder’s resilience are given and it is discussed how this can make a contribution to strive for civil security and furthermore a sustainable development.Keywords: civil security, emergency responders, stress, resilience, resources
Procedia PDF Downloads 143