Search results for: learning design
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 17871

Search results for: learning design

11301 Educational Sustainability: Teaching the Next Generation of Educators in Medical Simulation

Authors: Thomas Trouton, Sebastian Tanner, Manvir Sandher

Abstract:

The use of simulation in undergraduate and postgraduate medical curricula is ever-growing, is a useful addition to the traditional apprenticeship model of learning within medical education, and better prepares graduates for the team-based approach to healthcare seen in real-life clinical practice. As a learning tool, however, undergraduate medical students often have little understanding of the theory behind the use of medical simulation and have little experience in planning and delivering their own simulated teaching sessions. We designed and implemented a student-selected component (SSC) as part of the undergraduate medical curriculum at the University of Buckingham Medical School to introduce students to the concepts behind the use of medical simulation in education and allow them to plan and deliver their own simulated medical scenario to their peers. The SSC took place over a 2-week period in the 3rd year of the undergraduate course. There was a mix of lectures, seminars and interactive group work sessions, as well as hands-on experience in the simulation suite, to introduce key concepts related to medical simulation, including technical considerations in simulation, human factors, debriefing and troubleshooting scenarios. We evaluated the success of our SSC using “Net Promotor Scores” (NPS) to assess students’ confidence in planning and facilitating a simulation-based teaching session, as well as leading a debrief session. In all three domains, we showed an increase in the confidence of the students. We also showed an increase in confidence in the management of common medical emergencies as a result of the SSC. Overall, the students who chose our SSC had the opportunity to learn new skills in medical education, with a particular focus on the use of simulation-based teaching, and feedback highlighted that a number of students would take these skills forward in their own practice. We demonstrated an increase in confidence in several domains related to the use of medical simulation in education and have hopefully inspired a new generation of medical educators.

Keywords: simulation, SSC, teaching, medical students

Procedia PDF Downloads 108
11300 Don't Just Guess and Slip: Estimating Bayesian Knowledge Tracing Parameters When Observations Are Scant

Authors: Michael Smalenberger

Abstract:

Intelligent tutoring systems (ITS) are computer-based platforms which can incorporate artificial intelligence to provide step-by-step guidance as students practice problem-solving skills. ITS can replicate and even exceed some benefits of one-on-one tutoring, foster transactivity in collaborative environments, and lead to substantial learning gains when used to supplement the instruction of a teacher or when used as the sole method of instruction. A common facet of many ITS is their use of Bayesian Knowledge Tracing (BKT) to estimate parameters necessary for the implementation of the artificial intelligence component, and for the probability of mastery of a knowledge component relevant to the ITS. While various techniques exist to estimate these parameters and probability of mastery, none directly and reliably ask the user to self-assess these. In this study, 111 undergraduate students used an ITS in a college-level introductory statistics course for which detailed transaction-level observations were recorded, and users were also routinely asked direct questions that would lead to such a self-assessment. Comparisons were made between these self-assessed values and those obtained using commonly used estimation techniques. Our findings show that such self-assessments are particularly relevant at the early stages of ITS usage while transaction level data are scant. Once a user’s transaction level data become available after sufficient ITS usage, these can replace the self-assessments in order to eliminate the identifiability problem in BKT. We discuss how these findings are relevant to the number of exercises necessary to lead to mastery of a knowledge component, the associated implications on learning curves, and its relevance to instruction time.

Keywords: Bayesian Knowledge Tracing, Intelligent Tutoring System, in vivo study, parameter estimation

Procedia PDF Downloads 160
11299 A Full Factorial Analysis of Microhardness Variation in Bead Welds Deposited by the Process Cold Wire Gas Metal Arc Welding (CW-GMAW)

Authors: R. A. Ribeiro, P. D. Angelo Assunção, E. M. Braga

Abstract:

The microhardness in weld beads is a function of the microstructure obtained in the welding process, and this by its time is dependent of the input variables established at the outset of the process. In this study the influence of angle between the plate and the cold wire, the position in which the cold wire is introduced and the rate in which this introduction is made are assessed as input parameters in CW-GMAW process. This paper looks to show that ordinary changes in the frame of CW-GMAW can improve microhardness, which is expected to vary as the input parameters change. To properly correlate the changes in the input parameters to consequent changes in microhardness of the weld bead, a full factorial design was employed. In fact, changes in the operational parameters improved the overall microhardness of the weld bead, which in turns can be an indication of improvement in the resistance to abrasive wear, constituting a cheap way to augment the abrasion wear resistance of welds used for cladding.

Keywords: abrasion, CW-GMAW, full factorial design, microhardness

Procedia PDF Downloads 534
11298 Application of Computational Chemistry for Searching Anticancer Derivatives of 2-Phenazinamines as Bcr-Abl Tyrosine Kinase Inhibitors

Authors: Gajanan M. Sonwane

Abstract:

The computational studies on 2-phenazinamines with their protein targets have been carried out to design compounds with potential anticancer activity. This strategy of designing compounds possessing selectivity over specific tyrosine kinase has been achieved through G-QSAR and molecular docking studies. The objective of this research has been to design newer 2-phenazinamine derivatives as Bcr-Abl tyrosine kinase inhibitors by G-QSAR, molecular docking studies followed by wet-lab studies along with evaluation of their anticancer potential. Computational chemistry was done by using VLife MDS 4.3 and Autodock 4.2 followed by wet-lab experiments for synthesizing 2-phenazinamine derivatives. The chemical structures of ligands in 2D were drawn by employing Chemdraw 2D Ultra 8.0 and were converted into 3D. These were optimized by using a semi-empirical method called MOPAC. The protein structure was retrieved from RCSC protein data bank as a PDB file. The binding interactions of protein and ligands were done by using PYMOL. The molecular properties of the designed compounds were predicted in silico by using Osiris property explorer. The parent compound 2-phenazinamine was synthesized by reduction of 2, 4-dinitro-N-phenyl-benzenamine in the presence of tin chloride followed by cyclization in the presence of nitrobenzene and magnesium sulfate. The derivatization at the amino function of 2-phenazinamine was performed by treating parent compound with various aldehydes in the presence of dicyclohexylcarbodiimide (DCC) and urea to afford 2-(2-chlorophenyl)-3-(phenazine-2-yl) thiazolidine-4-one. Synthesized 39 novel derivatives of 2-phenazinamine and performed antioxidant activity, anti antiproliferative on the bulb of onion and anticancer activity on cell line showing significant competition with marked blockbuster drug imatinib.

Keywords: computer-aided drug design, tyrosin kinases, anticancer, docking

Procedia PDF Downloads 125
11297 Stuck Spaces as Moments of Learning: Uncovering Threshold Concepts in Teacher Candidate Experiences of Teaching in Inclusive Classrooms

Authors: Joy Chadwick

Abstract:

There is no doubt that classrooms of today are more complex and diverse than ever before. Preparing teacher candidates to meet these challenges is essential to ensure the retention of teachers within the profession and to ensure that graduates begin their teaching careers with the knowledge and understanding of how to effectively meet the diversity of students they will encounter. Creating inclusive classrooms requires teachers to have a repertoire of effective instructional skills and strategies. Teachers must also have the mindset to embrace diversity and value the uniqueness of individual students in their care. This qualitative study analyzed teacher candidates' experiences as they completed a fourteen-week teaching practicum while simultaneously completing a university course focused on inclusive pedagogy. The research investigated the challenges and successes teacher candidates had in navigating the translation of theory related to inclusive pedagogy into their teaching practice. Applying threshold concept theory as a framework, the research explored the troublesome concepts, liminal spaces, and transformative experiences as connected to inclusive practices. Threshold concept theory suggests that within all disciplinary fields, there exists particular threshold concepts that serve as gateways or portals into previously inaccessible ways of thinking and practicing. It is in these liminal spaces that conceptual shifts in thinking and understanding and deep learning can occur. The threshold concept framework provided a lens to examine teacher candidate struggles and successes with the inclusive education course content and the application of this content to their practicum experiences. A qualitative research approach was used, which included analyzing twenty-nine course reflective journals and six follow up one-to-one semi structured interviews. The journals and interview transcripts were coded and themed using NVivo software. Threshold concept theory was then applied to the data to uncover the liminal or stuck spaces of learning and the ways in which the teacher candidates navigated those challenging places of teaching. The research also sought to uncover potential transformative shifts in teacher candidate understanding as connected to teaching in an inclusive classroom. The findings suggested that teacher candidates experienced difficulties when they did not feel they had the knowledge, skill, or time to meet the needs of the students in the way they envisioned they should. To navigate the frustration of this thwarted vision, they relied on present and previous course content and experiences, collaborative work with other teacher candidates and their mentor teachers, and a proactive approach to planning for students. Transformational shifts were most evident in their ability to reframe their perceptions of children from a deficit or disability lens to a strength-based belief in the potential of students. It was evident that through their course work and practicum experiences, their beliefs regarding struggling students shifted as they saw the value of embracing neurodiversity, the importance of relationships, and planning for and teaching through a strength-based approach. Research findings have implications for teacher education programs and for understanding threshold concepts theory as connected to practice-based learning experiences.

Keywords: inclusion, inclusive education, liminal space, teacher education, threshold concepts, troublesome knowledge

Procedia PDF Downloads 61
11296 Modeling, Topology Optimization and Experimental Validation of Glass-Transition-Based 4D-Printed Polymeric Structures

Authors: Sara A. Pakvis, Giulia Scalet, Stefania Marconi, Ferdinando Auricchio, Matthijs Langelaar

Abstract:

In recent developments in the field of multi-material additive manufacturing, differences in material properties are exploited to create printed shape-memory structures, which are referred to as 4D-printed structures. New printing techniques allow for the deliberate introduction of prestresses in the specimen during manufacturing, and, in combination with the right design, this enables new functionalities. This research focuses on bi-polymer 4D-printed structures, where the transformation process is based on a heat-induced glass transition in one material lowering its Young’s modulus, combined with an initial prestress in the other material. Upon the decrease in stiffness, the prestress is released, which results in the realization of an essentially pre-programmed deformation. As the design of such functional multi-material structures is crucial but far from trivial, a systematic methodology to find the design of 4D-printed structures is developed, where a finite element model is combined with a density-based topology optimization method to describe the material layout. This modeling approach is verified by a convergence analysis and validated by comparing its numerical results to analytical and published data. Specific aspects that are addressed include the interplay between the definition of the prestress and the material interpolation function used in the density-based topology description, the inclusion of a temperature-dependent stiffness relationship to simulate the glass transition effect, and the importance of the consideration of geometric nonlinearity in the finite element modeling. The efficacy of topology optimization to design 4D-printed structures is explored by applying the methodology to a variety of design problems, both in 2D and 3D settings. Bi-layer designs composed of thermoplastic polymers are printed by means of the fused deposition modeling (FDM) technology. Acrylonitrile butadiene styrene (ABS) polymer undergoes the glass transition transformation, while polyurethane (TPU) polymer is prestressed by means of the 3D-printing process itself. Tests inducing shape transformation in the printed samples through heating are performed to calibrate the prestress and validate the modeling approach by comparing the numerical results to the experimental findings. Using the experimentally obtained prestress values, more complex designs have been generated through topology optimization, and samples have been printed and tested to evaluate their performance. This study demonstrates that by combining topology optimization and 4D-printing concepts, stimuli-responsive structures with specific properties can be designed and realized.

Keywords: 4D-printing, glass transition, shape memory polymer, topology optimization

Procedia PDF Downloads 186
11295 Coupled Field Formulation – A Unified Method for Formulating Structural Mechanics Problems

Authors: Ramprasad Srinivasan

Abstract:

Engineers create inventions and put their ideas in concrete terms to design new products. Design drivers must be established, which requires, among other things, a complete understanding of the product design, load paths, etc. For Aerospace Vehicles, weight/strength ratio, strength, stiffness and stability are the important design drivers. A complex built-up structure is made up of an assemblage of primitive structural forms of arbitrary shape, which include 1D structures like beams and frames, 2D structures like membranes, plate and shell structures, and 3D solid structures. Justification through simulation involves a check for all the quantities of interest, namely stresses, deformation, frequencies, and buckling loads and is normally achieved through the finite element (FE) method. Over the past few decades, Fiber-reinforced composites are fast replacing the traditional metallic structures in the weight-sensitive aerospace and aircraft industries due to their high specific strength, high specific stiffness, anisotropic properties, design freedom for tailoring etc. Composite panel constructions are used in aircraft to design primary structure components like wings, empennage, ailerons, etc., while thin-walled composite beams (TWCB) are used to model slender structures like stiffened panels, helicopter, and wind turbine rotor blades, etc. The TWCB demonstrates many non-classical effects like torsional and constrained warping, transverse shear, coupling effects, heterogeneity, etc., which makes the analysis of composite structures far more complex. Conventional FE formulations to model 1D structures suffer from many limitations like shear locking, particularly in slender beams, lower convergence rates due to material coupling in composites, inability to satisfy, equilibrium in the domain and natural boundary conditions (NBC) etc. For 2D structures, the limitations of conventional displacement-based FE formulations include the inability to satisfy NBC explicitly and many pathological problems such as shear and membrane locking, spurious modes, stress oscillations, lower convergence due to mesh distortion etc. This mandates frequent re-meshing to even achieve an acceptable mesh (satisfy stringent quality metrics) for analysis leading to significant cycle time. Besides, currently, there is a need for separate formulations (u/p) to model incompressible materials, and a single unified formulation is missing in the literature. Hence coupled field formulation (CFF) is a unified formulation proposed by the author for the solution of complex 1D and 2D structures addressing the gaps in the literature mentioned above. The salient features of CFF and its many advantages over other conventional methods shall be presented in this paper.

Keywords: coupled field formulation, kinematic and material coupling, natural boundary condition, locking free formulation

Procedia PDF Downloads 57
11294 Coming Closer to Communities of Practice through Situated Learning: The Case Study of Polish-English, English-Polish Undergraduate BA Level Language for Specific Purposes of Translation Class

Authors: Marta Lisowska

Abstract:

The growing trend of market specialization imposes upon translators the need for proficiency in the working knowledge of specialist discourse. The notion of specialization differs from a broad general category to a highly specialized narrow field. The specialised discourse is used in the channel of communication based upon distinctive features typical for communities of practice whose co-existence is codified and hermetically locked against outsiders. Consequently, any translator deprived of professional discourse competence and social skills is incapable of providing competent translation product from source language into target language. In this paper, we report on research that explores the pedagogical practices aiming to bridge the dichotomy between the professionals and the specialist translators, while accounting for the reality of the world of professional communities entered by undergraduates on two levels: the text-based generic, and the social one. Drawing from the functional social constructivist approach, seen here as situated learning, this paper reports on the case of English-Polish, Polish-English undergraduate BA Level LSP of law translation class run in line with the simulated classroom-based and the reality-based (apprenticeship) approach. This blended method serves the purpose of introducing the young trainees to the professional world. The research provides new insights into how the LSP translation undergraduates become legitimized through discursive and social participation and engagement. The undergraduates, situated peripherally at the outset, experience their own transformation towards becoming members of these professional groups. With subjective evaluation, the trainees take a stance on this dual mode class and development of their skills. Comparing and contrasting their own work done in line with two models of translation teaching: authentic and near-authentic, the undergraduates answer research questions devised by a questionnaire survey The responses take us closer to how students feel about their LSP translation competence development. The major findings show how the trainees perceive the benefits and hardships of their functional translation class. In terms of skills, they related to communication as the most enhanced one; they highly valued the fact of being ‘exposed’ to a variety of texts (cf. multi literalism), team work, learning how to schedule work, IT skills boost and the ability to learn how to work individually. Another finding indicates that students struggled most with specialized language, and co-working with other students. The short-term research shows the momentum when the undergraduate LSP translation trainees entered the path of transformation i.e. gained consciousness of ‘how it is’ to be a participant-translator of real-life communities of practice, gaining pragmatic dint of the social and linguistic skills understood here as discursive competence (text > genre > discourse > professional practice). The undergraduates need to be aware of the work they have to do and challenges they are to face before arriving at the expert level of professional translation competence.

Keywords: communities of practice in LSP translation teaching, learning LSP translation as situated experience, peripheral participation, professional discourse for LSP translation teaching, professional translation competence

Procedia PDF Downloads 86
11293 Action Research-Informed Multiliteracies-Enhanced Pedagogy in an Online English for Academic Purposes Course

Authors: Heejin Song

Abstract:

Employing a critical action research approach that rejects essentialist onto-epistemological orientations to research in English language teaching (ELT) and interrogates the hegemonic relations in the knowledge construction and reconstruction processes, this study illuminates how an action research-informed pedagogical practice can transform the English for academic purposes (EAP) teaching to be more culturally and linguistically inclusive and critically oriented for English language learners’ advancement in academic literacies skills. More specifically, this paper aims to showcase the action research-informed pedagogical innovations that emphasize multilingual learners’ multiliteracies engagement and experiential education-oriented learning to facilitate the development of learners’ academic literacies, intercultural communicative competence, and inclusive global citizenship in the context of Canadian university EAP classrooms. The pedagogical innovations through action research embarked in response to growing discussions surrounding pedagogical possibilities of plurilingualism in ELT and synchronous online teaching. The paper is based on two iterations of action research over the pandemic years between 2020 and 2022. The data includes student work samples, focus group interviews, anonymous surveys, teacher feedback and comments on student work and teaching reflections. The first iteration of the action research focused on the affordances of multimodal expressions in individual learners’ academic endeavors for their literacy skills development through individual online activities such as ‘my language autobiography,’ ‘multimodal expression corner’ and public speeches. While these activities help English language learners enhance their knowledge and skills of English-spoken discourses, these tasks did not necessarily require learners’ team-based collaborative endeavors to complete the assigned tasks. Identifying this area for improvement in the instructional design, the second action research cycle/iteration emphasized collaborative performativity through newly added performance/action-based innovative learning tasks, including ‘situational role-playing’, ‘my cooking show & interview’, and group debates in order to provide learners increased opportunities to communicate with peers who joined the class virtually from different parts of the world and enhance learners’ intercultural competence through various strategic and pragmatic communicative skills to collaboratively achieve their shared goals (i.e., successful completion of the given group tasks). The paper exemplifies instances wherein learners’ unique and diverse linguistic and cultural strengths were amplified, and critical literacies were further developed through learners’ performance-oriented multiliteracies engagement. The study suggests that the action research-informed teaching practice that advocates for collaborative multiliteracies engagement serves to facilitate learners’ activation of their existing linguistic and cultural knowledge and contributes to the development of learners’ academic literacy skills. Importantly, the study illuminates that such action research-informed pedagogical initiatives create an inclusive space for learners to build a strong sense of connectedness as global citizens with increased intercultural awareness in their community of language and cultural practices, and further allow learners to actively participate in the construction of ‘collaborative relations of power’ with their peers.

Keywords: action research, EAP, higher education, multiliteracies

Procedia PDF Downloads 66
11292 An International Curriculum Development for Languages and Technology

Authors: Miguel Nino

Abstract:

When considering the challenges of a changing and demanding globalizing world, it is important to reflect on how university students will be prepared for the realities of internationalization, marketization and intercultural conversation. The present study is an interdisciplinary program designed to respond to the needs of the global community. The proposal bridges the humanities and science through three different fields: Languages, graphic design and computer science, specifically, fundamentals of programming such as python, java script and software animation. Therefore, the goal of the four year program is twofold: First, enable students for intercultural communication between English and other languages such as Spanish, Mandarin, French or German. Second, students will acquire knowledge in practical software and relevant employable skills to collaborate in assisted computer projects that most probable will require essential programing background in interpreted or compiled languages. In order to become inclusive and constructivist, the cognitive linguistics approach is suggested for the three different fields, particularly for languages that rely on the traditional method of repetition. This methodology will help students develop their creativity and encourage them to become independent problem solving individuals, as languages enhance their common ground of interaction for culture and technology. Participants in this course of study will be evaluated in their second language acquisition at the Intermediate-High level. For graphic design and computer science students will apply their creative digital skills, as well as their critical thinking skills learned from the cognitive linguistics approach, to collaborate on a group project design to find solutions for media web design problems or marketing experimentation for a company or the community. It is understood that it will be necessary to apply programming knowledge and skills to deliver the final product. In conclusion, the program equips students with linguistics knowledge and skills to be competent in intercultural communication, where English, the lingua franca, remains the medium for marketing and product delivery. In addition to their employability, students can expand their knowledge and skills in digital humanities, computational linguistics, or increase their portfolio in advertising and marketing. These students will be the global human capital for the competitive globalizing community.

Keywords: curriculum, international, languages, technology

Procedia PDF Downloads 432
11291 Diagnostics of Existing Steel Structures of Winter Sport Halls

Authors: Marcela Karmazínová, Jindrich Melcher, Lubomír Vítek, Petr Cikrle

Abstract:

The paper deals with the diagnostics of steel roof structure of the winter sports stadiums built in 1970 year. The necessity of the diagnostics has been given by the requirement to the evaluation design of this structure, which has been caused by the new situation in the field of the loadings given by the validity of the European Standards in the Czech Republic from 2010 year. Due to these changes in the normative rules, in practice, existing structures are gradually subjected to the evaluation design and depending on its results to the strengthening or reconstruction, respectively. The steel roof is composed of plane truss main girders, purlins and bracings and the roof structure is supported by two arch main girders with the span of L=84 m. The in situ diagnostics of the roof structure was oriented to the following parts: (i) determination and evaluation of the actual material properties of used steel and (ii) verification of the actual dimensions of the structural members. For the solution, the non-destructive methods have been used for in situ measurement. For the indicative determination of steel strengths the modified method based on the determination of Rockwell’s hardness has been used. For the verification of the member’s dimensions (thickness of hollow sections) the ultrasound method has been used. This paper presents the results obtained using these testing methods and their evaluation, from the viewpoint of the usage for the subsequent static assessment and design evaluation of the existing structure. For the comparison, the examples of the similar evaluations realized for steel structures of the stadiums in Olomouc and Jihlava cities are briefly illustrated, too.

Keywords: actual dimensions, destructive methods, diagnostics, existing steel structure, indirect non-destructive methods, Rockwel’s hardness, sport hall, steel strength, ultrasound method.

Procedia PDF Downloads 327
11290 Clean Sky 2 Project LiBAT: Light Battery Pack for High Power Applications in Aviation – Simulation Methods in Early Stage Design

Authors: Jan Dahlhaus, Alejandro Cardenas Miranda, Frederik Scholer, Maximilian Leonhardt, Matthias Moullion, Frank Beutenmuller, Julia Eckhardt, Josef Wasner, Frank Nittel, Sebastian Stoll, Devin Atukalp, Daniel Folgmann, Tobias Mayer, Obrad Dordevic, Paul Riley, Jean-Marc Le Peuvedic

Abstract:

Electrical and hybrid aerospace technologies pose very challenging demands on the battery pack – especially with respect to weight and power. In the Clean Sky 2 research project LiBAT (funded by the EU), the consortium is currently building an ambitious prototype with state-of-the art cells that shows the potential of an intelligent pack design with a high level of integration, especially with respect to thermal management and power electronics. For the latter, innovative multi-level-inverter technology is used to realize the required power converting functions with reduced equipment. In this talk the key approaches and methods of the LiBat project will be presented and central results shown. Special focus will be set on the simulative methods used to support the early design and development stages from an overall system perspective. The applied methods can efficiently handle multiple domains and deal with different time and length scales, thus allowing the analysis and optimization of overall- or sub-system behavior. It will be shown how these simulations provide valuable information and insights for the efficient evaluation of concepts. As a result, the construction and iteration of hardware prototypes has been reduced and development cycles shortened.

Keywords: electric aircraft, battery, Li-ion, multi-level-inverter, Novec

Procedia PDF Downloads 151
11289 Predictive Maintenance: Machine Condition Real-Time Monitoring and Failure Prediction

Authors: Yan Zhang

Abstract:

Predictive maintenance is a technique to predict when an in-service machine will fail so that maintenance can be planned in advance. Analytics-driven predictive maintenance is gaining increasing attention in many industries such as manufacturing, utilities, aerospace, etc., along with the emerging demand of Internet of Things (IoT) applications and the maturity of technologies that support Big Data storage and processing. This study aims to build an end-to-end analytics solution that includes both real-time machine condition monitoring and machine learning based predictive analytics capabilities. The goal is to showcase a general predictive maintenance solution architecture, which suggests how the data generated from field machines can be collected, transmitted, stored, and analyzed. We use a publicly available aircraft engine run-to-failure dataset to illustrate the streaming analytics component and the batch failure prediction component. We outline the contributions of this study from four aspects. First, we compare the predictive maintenance problems from the view of the traditional reliability centered maintenance field, and from the view of the IoT applications. When evolving to the IoT era, predictive maintenance has shifted its focus from ensuring reliable machine operations to improve production/maintenance efficiency via any maintenance related tasks. It covers a variety of topics, including but not limited to: failure prediction, fault forecasting, failure detection and diagnosis, and recommendation of maintenance actions after failure. Second, we review the state-of-art technologies that enable a machine/device to transmit data all the way through the Cloud for storage and advanced analytics. These technologies vary drastically mainly based on the power source and functionality of the devices. For example, a consumer machine such as an elevator uses completely different data transmission protocols comparing to the sensor units in an environmental sensor network. The former may transfer data into the Cloud via WiFi directly. The latter usually uses radio communication inherent the network, and the data is stored in a staging data node before it can be transmitted into the Cloud when necessary. Third, we illustrate show to formulate a machine learning problem to predict machine fault/failures. By showing a step-by-step process of data labeling, feature engineering, model construction and evaluation, we share following experiences: (1) what are the specific data quality issues that have crucial impact on predictive maintenance use cases; (2) how to train and evaluate a model when training data contains inter-dependent records. Four, we review the tools available to build such a data pipeline that digests the data and produce insights. We show the tools we use including data injection, streaming data processing, machine learning model training, and the tool that coordinates/schedules different jobs. In addition, we show the visualization tool that creates rich data visualizations for both real-time insights and prediction results. To conclude, there are two key takeaways from this study. (1) It summarizes the landscape and challenges of predictive maintenance applications. (2) It takes an example in aerospace with publicly available data to illustrate each component in the proposed data pipeline and showcases how the solution can be deployed as a live demo.

Keywords: Internet of Things, machine learning, predictive maintenance, streaming data

Procedia PDF Downloads 374
11288 Data Analysis Tool for Predicting Water Scarcity in Industry

Authors: Tassadit Issaadi Hamitouche, Nicolas Gillard, Jean Petit, Valerie Lavaste, Celine Mayousse

Abstract:

Water is a fundamental resource for the industry. It is taken from the environment either from municipal distribution networks or from various natural water sources such as the sea, ocean, rivers, aquifers, etc. Once used, water is discharged into the environment, reprocessed at the plant or treatment plants. These withdrawals and discharges have a direct impact on natural water resources. These impacts can apply to the quantity of water available, the quality of the water used, or to impacts that are more complex to measure and less direct, such as the health of the population downstream from the watercourse, for example. Based on the analysis of data (meteorological, river characteristics, physicochemical substances), we wish to predict water stress episodes and anticipate prefectoral decrees, which can impact the performance of plants and propose improvement solutions, help industrialists in their choice of location for a new plant, visualize possible interactions between companies to optimize exchanges and encourage the pooling of water treatment solutions, and set up circular economies around the issue of water. The development of a system for the collection, processing, and use of data related to water resources requires the functional constraints specific to the latter to be made explicit. Thus the system will have to be able to store a large amount of data from sensors (which is the main type of data in plants and their environment). In addition, manufacturers need to have 'near-real-time' processing of information in order to be able to make the best decisions (to be rapidly notified of an event that would have a significant impact on water resources). Finally, the visualization of data must be adapted to its temporal and geographical dimensions. In this study, we set up an infrastructure centered on the TICK application stack (for Telegraf, InfluxDB, Chronograf, and Kapacitor), which is a set of loosely coupled but tightly integrated open source projects designed to manage huge amounts of time-stamped information. The software architecture is coupled with the cross-industry standard process for data mining (CRISP-DM) data mining methodology. The robust architecture and the methodology used have demonstrated their effectiveness on the study case of learning the level of a river with a 7-day horizon. The management of water and the activities within the plants -which depend on this resource- should be considerably improved thanks, on the one hand, to the learning that allows the anticipation of periods of water stress, and on the other hand, to the information system that is able to warn decision-makers with alerts created from the formalization of prefectoral decrees.

Keywords: data mining, industry, machine Learning, shortage, water resources

Procedia PDF Downloads 110
11287 Combining Diffusion Maps and Diffusion Models for Enhanced Data Analysis

Authors: Meng Su

Abstract:

High-dimensional data analysis often presents challenges in capturing the complex, nonlinear relationships and manifold structures inherent to the data. This article presents a novel approach that leverages the strengths of two powerful techniques, Diffusion Maps and Diffusion Probabilistic Models (DPMs), to address these challenges. By integrating the dimensionality reduction capability of Diffusion Maps with the data modeling ability of DPMs, the proposed method aims to provide a comprehensive solution for analyzing and generating high-dimensional data. The Diffusion Map technique preserves the nonlinear relationships and manifold structure of the data by mapping it to a lower-dimensional space using the eigenvectors of the graph Laplacian matrix. Meanwhile, DPMs capture the dependencies within the data, enabling effective modeling and generation of new data points in the low-dimensional space. The generated data points can then be mapped back to the original high-dimensional space, ensuring consistency with the underlying manifold structure. Through a detailed example implementation, the article demonstrates the potential of the proposed hybrid approach to achieve more accurate and effective modeling and generation of complex, high-dimensional data. Furthermore, it discusses possible applications in various domains, such as image synthesis, time-series forecasting, and anomaly detection, and outlines future research directions for enhancing the scalability, performance, and integration with other machine learning techniques. By combining the strengths of Diffusion Maps and DPMs, this work paves the way for more advanced and robust data analysis methods.

Keywords: diffusion maps, diffusion probabilistic models (DPMs), manifold learning, high-dimensional data analysis

Procedia PDF Downloads 86
11286 A Review of Fused Deposition Modeling Process: Parameter Optimization, Materials and Design

Authors: Elisaveta Doncheva, Jelena Djokikj, Ognen Tuteski, Bojana Hadjieva

Abstract:

In the past decade, additive manufacturing technology or 3D printing has been promoted as an efficient method for fabricating hybrid composite materials and structures with superior mechanical properties and complex shape and geometry. Fused deposition modeling (FDM) process is commonly used additive manufacturing technique for production of polymer products. Therefore, many studies and experiments are focused on investigating the possibilities for improving the obtained results on product properties as a key factor for expanding the spectrum of their application. This article provides an extensive review on recent research advances in FDM and reports on studies that cover the effects of process parameters, material, and design of the product properties. The paper conclusions provide a clear up-to date information for optimum efficiency and enhancement of the mechanical properties of 3D printed samples and recommends further research work and investigations.

Keywords: additive manufacturing, critical parameters, filament, print orientation, 3D printing

Procedia PDF Downloads 176
11285 Optimum Dewatering Network Design Using Firefly Optimization Algorithm

Authors: S. M. Javad Davoodi, Mojtaba Shourian

Abstract:

Groundwater table close to the ground surface causes major problems in construction and mining operation. One of the methods to control groundwater in such cases is using pumping wells. These pumping wells remove excess water from the site project and lower the water table to a desirable value. Although the efficiency of this method is acceptable, it needs high expenses to apply. It means even small improvement in a design of pumping wells can lead to substantial cost savings. In order to minimize the total cost in the method of pumping wells, a simulation-optimization approach is applied. The proposed model integrates MODFLOW as the simulation model with Firefly as the optimization algorithm. In fact, MODFLOW computes the drawdown due to pumping in an aquifer and the Firefly algorithm defines the optimum value of design parameters which are numbers, pumping rates and layout of the designing wells. The developed Firefly-MODFLOW model is applied to minimize the cost of the dewatering project for the ancient mosque of Kerman city in Iran. Repetitive runs of the Firefly-MODFLOW model indicates that drilling two wells with the total rate of pumping 5503 m3/day is the result of the minimization problem. Results show that implementing the proposed solution leads to at least 1.5 m drawdown in the aquifer beneath mosque region. Also, the subsidence due to groundwater depletion is less than 80 mm. Sensitivity analyses indicate that desirable groundwater depletion has an enormous impact on total cost of the project. Besides, in a hypothetical aquifer decreasing the hydraulic conductivity contributes to decrease in total water extraction for dewatering.

Keywords: groundwater dewatering, pumping wells, simulation-optimization, MODFLOW, firefly algorithm

Procedia PDF Downloads 282
11284 Automated Adaptions of Semantic User- and Service Profile Representations by Learning the User Context

Authors: Nicole Merkle, Stefan Zander

Abstract:

Ambient Assisted Living (AAL) describes a technological and methodological stack of (e.g. formal model-theoretic semantics, rule-based reasoning and machine learning), different aspects regarding the behavior, activities and characteristics of humans. Hence, a semantic representation of the user environment and its relevant elements are required in order to allow assistive agents to recognize situations and deduce appropriate actions. Furthermore, the user and his/her characteristics (e.g. physical, cognitive, preferences) need to be represented with a high degree of expressiveness in order to allow software agents a precise evaluation of the users’ context models. The correct interpretation of these context models highly depends on temporal, spatial circumstances as well as individual user preferences. In most AAL approaches, model representations of real world situations represent the current state of a universe of discourse at a given point in time by neglecting transitions between a set of states. However, the AAL domain currently lacks sufficient approaches that contemplate on the dynamic adaptions of context-related representations. Semantic representations of relevant real-world excerpts (e.g. user activities) help cognitive, rule-based agents to reason and make decisions in order to help users in appropriate tasks and situations. Furthermore, rules and reasoning on semantic models are not sufficient for handling uncertainty and fuzzy situations. A certain situation can require different (re-)actions in order to achieve the best results with respect to the user and his/her needs. But what is the best result? To answer this question, we need to consider that every smart agent requires to achieve an objective, but this objective is mostly defined by domain experts who can also fail in their estimation of what is desired by the user and what not. Hence, a smart agent has to be able to learn from context history data and estimate or predict what is most likely in certain contexts. Furthermore, different agents with contrary objectives can cause collisions as their actions influence the user’s context and constituting conditions in unintended or uncontrolled ways. We present an approach for dynamically updating a semantic model with respect to the current user context that allows flexibility of the software agents and enhances their conformance in order to improve the user experience. The presented approach adapts rules by learning sensor evidence and user actions using probabilistic reasoning approaches, based on given expert knowledge. The semantic domain model consists basically of device-, service- and user profile representations. In this paper, we present how this semantic domain model can be used in order to compute the probability of matching rules and actions. We apply this probability estimation to compare the current domain model representation with the computed one in order to adapt the formal semantic representation. Our approach aims at minimizing the likelihood of unintended interferences in order to eliminate conflicts and unpredictable side-effects by updating pre-defined expert knowledge according to the most probable context representation. This enables agents to adapt to dynamic changes in the environment which enhances the provision of adequate assistance and affects positively the user satisfaction.

Keywords: ambient intelligence, machine learning, semantic web, software agents

Procedia PDF Downloads 266
11283 Tracing the Developmental Repertoire of the Progressive: Evidence from L2 Construction Learning

Authors: Tianqi Wu, Min Wang

Abstract:

Research investigating language acquisition from a constructionist perspective has demonstrated that language is learned as constructions at various linguistic levels, which is related to factors of frequency, semantic prototypicality, and form-meaning contingency. However, previous research on construction learning tended to focus on clause-level constructions such as verb argument constructions but few attempts were made to study morpheme-level constructions such as the progressive construction, which is regarded as a source of acquisition problems for English learners from diverse L1 backgrounds, especially for those whose L1 do not have an equivalent construction such as German and Chinese. To trace the developmental trajectory of Chinese EFL learners’ use of the progressive with respect to verb frequency, verb-progressive contingency, and verbal prototypicality and generality, a learner corpus consisting of three sub-corpora representing three different English proficiency levels was extracted from the Chinese Learners of English Corpora (CLEC). As the reference point, a native speakers’ corpus extracted from the Louvain Corpus of Native English Essays was also established. All the texts were annotated with C7 tagset by part-of-speech tagging software. After annotation all valid progressive hits were retrieved with AntConc 3.4.3 followed by a manual check. Frequency-related data showed that from the lowest to the highest proficiency level, (1) the type token ratio increased steadily from 23.5% to 35.6%, getting closer to 36.4% in the native speakers’ corpus, indicating a wider use of verbs in the progressive; (2) the normalized entropy value rose from 0.776 to 0.876, working towards the target score of 0.886 in native speakers’ corpus, revealing that upper-intermediate learners exhibited a more even distribution and more productive use of verbs in the progressive; (3) activity verbs (i.e., verbs with prototypical progressive meanings like running and singing) dropped from 59% to 34% but non-prototypical verbs such as state verbs (e.g., being and living) and achievement verbs (e.g., dying and finishing) were increasingly used in the progressive. Apart from raw frequency analyses, collostructional analyses were conducted to quantify verb-progressive contingency and to determine what verbs were distinctively associated with the progressive construction. Results were in line with raw frequency findings, which showed that contingency between the progressive and non-prototypical verbs represented by light verbs (e.g., going, doing, making, and coming) increased as English proficiency proceeded. These findings altogether suggested that beginning Chinese EFL learners were less productive in using the progressive construction: they were constrained by a small set of verbs which had concrete and typical progressive meanings (e.g., the activity verbs). But with English proficiency increasing, their use of the progressive began to spread to marginal members such as the light verbs.

Keywords: Construction learning, Corpus-based, Progressives, Prototype

Procedia PDF Downloads 116
11282 Ontology-Driven Knowledge Discovery and Validation from Admission Databases: A Structural Causal Model Approach for Polytechnic Education in Nigeria

Authors: Bernard Igoche Igoche, Olumuyiwa Matthew, Peter Bednar, Alexander Gegov

Abstract:

This study presents an ontology-driven approach for knowledge discovery and validation from admission databases in Nigerian polytechnic institutions. The research aims to address the challenges of extracting meaningful insights from vast amounts of admission data and utilizing them for decision-making and process improvement. The proposed methodology combines the knowledge discovery in databases (KDD) process with a structural causal model (SCM) ontological framework. The admission database of Benue State Polytechnic Ugbokolo (Benpoly) is used as a case study. The KDD process is employed to mine and distill knowledge from the database, while the SCM ontology is designed to identify and validate the important features of the admission process. The SCM validation is performed using the conditional independence test (CIT) criteria, and an algorithm is developed to implement the validation process. The identified features are then used for machine learning (ML) modeling and prediction of admission status. The results demonstrate the adequacy of the SCM ontological framework in representing the admission process and the high predictive accuracies achieved by the ML models, with k-nearest neighbors (KNN) and support vector machine (SVM) achieving 92% accuracy. The study concludes that the proposed ontology-driven approach contributes to the advancement of educational data mining and provides a foundation for future research in this domain.

Keywords: admission databases, educational data mining, machine learning, ontology-driven knowledge discovery, polytechnic education, structural causal model

Procedia PDF Downloads 43
11281 Design of a Telemetry, Tracking, and Command Radio-Frequency Receiver for Small Satellites Based on Commercial Off-The-Shelf Components

Authors: A. Lovascio, A. D’Orazio, V. Centonze

Abstract:

From several years till now the aerospace industry is developing more and more small satellites for Low-Earth Orbit (LEO) missions. Such satellites have a low cost of making and launching since they have a size and weight smaller than other types of satellites. However, because of size limitations, small satellites need integrated electronic equipment based on digital logic. Moreover, the LEOs require telecommunication modules with high throughput to transmit to earth a big amount of data in a short time. In order to meet such requirements, in this paper we propose a Telemetry, Tracking & Command module optimized through the use of the Commercial Off-The-Shelf components. The proposed approach exploits the major flexibility offered by these components in reducing costs and optimizing the performance. The method has been applied in detail for the design of the front-end receiver, which has a low noise figure (1.5 dB) and DC power consumption (smaller than 2 W). Such a performance is particularly attractive since it allows fulfilling the energy budget stringent constraints that are typical for LEO small platforms.

Keywords: COTS, LEO, small-satellite, TT&C

Procedia PDF Downloads 117
11280 Evaluating the Satisfaction of Chinese Consumers toward Influencers at TikTok

Authors: Noriyuki Suyama

Abstract:

The progress and spread of digitalization have led to the provision of a variety of new services. The recent progress in digitization can be attributed to rapid developments in science and technology. First, the research and diffusion of artificial intelligence (AI) has made dramatic progress. Around 2000, the third wave of AI research, which had been underway for about 50 years, arrived. Specifically, machine learning and deep learning were made possible in AI, and the ability of AI to acquire knowledge, define the knowledge, and update its own knowledge in a quantitative manner made the use of big data practical even for commercial PCs. On the other hand, with the spread of social media, information exchange has become more common in our daily lives, and the lending and borrowing of goods and services, in other words, the sharing economy, has become widespread. The scope of this trend is not limited to any industry, and its momentum is growing as the SDGs take root. In addition, the Social Network Service (SNS), a part of social media, has brought about the evolution of the retail business. In the past few years, social network services (SNS) involving users or companies have especially flourished. The People's Republic of China (hereinafter referred to as "China") is a country that is stimulating enormous consumption through its own unique SNS, which is different from the SNS used in developed countries around the world. This paper focuses on the effectiveness and challenges of influencer marketing by focusing on the influence of influencers on users' behavior and satisfaction with Chinese SNSs. Specifically, Conducted was the quantitative survey of Tik Tok users living in China, with the aim of gaining new insights from the analysis and discussions. As a result, we found several important findings and knowledge.

Keywords: customer satisfaction, social networking services, influencer marketing, Chinese consumers’ behavior

Procedia PDF Downloads 82
11279 Noise Barrier Technique as a Way to Improve the Sonic Urban Environment along Existing Roadways Assessment: El-Gish Road Street, Alexandria, Egypt

Authors: Nihal Atif Salim

Abstract:

To improve the quality of life in cities, a variety of interventions are used. Noise is a substantial and important sort of pollution that has a negative impact on the urban environment and human health. According to the complaint survey, it ranks second among environmental contamination complaints (conducted by EEAA in 2019). The most significant source of noise in the city is traffic noise. In order to improve the sound urban environment, many physical techniques are applied. In the local area, noise barriers are considered as one of the most appropriate physical techniques along existing traffic routes. Alexandria is Egypt's second-largest city after Cairo. It is located along the Mediterranean Sea, and El- Gish Road is one of the city's main arteries. It impacts the waterfront promenade that extends along with the city by a high level of traffic noise. The purpose of this paper is to clarify the design considerations for the most appropriate noise barrier type along with the promenade, with the goal of improving the Quality of Life (QOL) and the sonic urban environment specifically. The proposed methodology focuses on how noise affects human perception and the environment. Then it delves into the various physical noise control approaches. After that, the paper discusses sustainable design decisions making. Finally, look into the importance of incorporating sustainability into design decisions making. Three stages will be followed in the case study. The first stage involves doing a site inspection and using specific sound measurement equipment (a noise level meter) to measure the noise level along the promenade at many sites, and the findings will be shown on a noise map. The second step is to inquire about the site's user experience. The third step is to investigate the various types of noise barriers and their effects on QOL along existing routes in order to select the most appropriate type. The goal of this research is to evaluate the suitable design of noise barriers that fulfill environmental and social perceptions while maintaining a balanced approach to the noise issue in order to improve QOL along existing roadways in the local area.

Keywords: noise pollution, sonic urban environment, traffic noise, noise barrier, acoustic sustainability, noise reduction techniques

Procedia PDF Downloads 128
11278 The Effect of Using Mobile Listening Applications on Listening Skills of Iranian Intermediate EFL Learners

Authors: Mahmoud Nabilu

Abstract:

The present study explored the effect of using Mobile listening applications on developing listening skills by Iranian intermediate EFL learners. Fifty male intermediate English learners whose age range was between 15 and 20, participated in the study. The participants were placed in two groups on the basis of their scores on a placement test. Therefore, the participants of the study were homogenized in terms of general proficiency, and groups were assigned as one experimental group and one control group. The experimental group was instructed by the treatment which was using mobile applications to develop their listening skills while the control group received traditional methods. The research data were obtained from the 40-item multiple-choice tests as a pre-test and a post-test. The results of the t-test clearly revealed that the learners in the experimental group performed better in the post-test than the pre-test. This implies that using a mobile application for developing listening skills as a treatment was effective in helping the language learners perform better on post-test. However, a statistically significant difference was found between the post-tests scores of the two groups. The mean of the experimental group was greater compared to the control group. The participants were Iranian and from an Iranian Language Institute, so care should be taken while generalizing the results to the learners of other nationalities. However, in the researcher's view, the findings of this study have valuable implications for teachers and learners, methodologists and syllabus designers, linguists and MALL/CALL (mobile/computer-assisted language learning) experts. Using the result of the present paper is an aim of raising the consciousness of a better technique of developing listening skills in order to make language learning more efficient for the learners.

Keywords: Mobile listening applications, intermediate EFL learners, MALL, CALL

Procedia PDF Downloads 181
11277 Waste Heat Recovery Using Spiral Heat Exchanger

Authors: Parthiban S. R.

Abstract:

Spiral heat exchangers are known as excellent heat exchanger because of far compact and high heat transfer efficiency. An innovative spiral heat exchanger based on polymer materials is designed for waste heat recovery process. Such a design based on polymer film technology provides better corrosion and chemical resistance compared to conventional metal heat exchangers. Due to the smooth surface of polymer film fouling is reduced. A new arrangement for flow of hot flue gas and cold fluid is employed for design, flue gas flows in axial path while the cold fluid flows in a spiral path. Heat load recovery achieved with the presented heat exchanger is in the range of 1.5 kW thermic but potential heat recovery about 3.5 kW might be achievable. To measure the performance of the spiral tube heat exchanger, its model is suitably designed and fabricated so as to perform experimental tests. The paper gives analysis of spiral tube heat exchanger.

Keywords: spiral heat exchanger, polymer based materials, fouling factor, heat load

Procedia PDF Downloads 372
11276 The Effect of Phase Development on Micro-Climate Change of Urban Area

Authors: Tommy Lo

Abstract:

This paper presented the changes in temperature and air ventilation of an inner urban area at different development stages during 2002 to 2012 and the high-rise buildings to be built in 2018. 3D simulation models ENVI-met and Autodesk Falcon were used. The results indicated that replacement of old residence buildings or open space with high-rise buildings will increase the air temperature of inner urban area; the air temperature at the pedestrian level will increase more than that at the upper levels. The temperature of the inner street in future will get higher than that in 2002, 2008 and 2012. It is attributed that heat is trapped in the street canyons as the air permeability at the pedestrian levels is lower. High-rise buildings with massive podium will further reduce the air ventilation in that area. In addition, sufficient separations among buildings is essential in design. High-rise buildings aligned along the waterfront will obstruct the wind flowing into the inner urban area and accelerate the temperature increase both in daytime and night time.

Keywords: micro-climate change, urban design, ENVI-met, construction engineering

Procedia PDF Downloads 270
11275 Ten Minutes Neighbourhood as a Basic PlanningUnit for Happiness in Egypt

Authors: Abeer Elshater

Abstract:

This paper pursues the relationship between the inhabitants’ happiness and the right to the city in an Egyptian neighbourhood status quo. Although the optimum of getting the services comes from ten mints walking in a suitable ambiance, the happiness is not acquired. The research objective is, first, to review the literature that get a guideline of 10 minutes neighbourhoods. Second make a comparative content analysis to recent online articles to the right to the city. Third is to test the concluded principles in Egyptian neighbourhood settings. The idea of ten minutes neighbourhood is manageable. The hypothesis concerns a compliant design. The logic of people who live close to within ten minutes’ walk to essential settings in their area can minimize several problems and maximize a healthy lifestyle. The supposed issue makes the right to the city affect the relationship between ten minutes neighbourhood and citizen happiness. This assumption can be intervention through site observation and oriented questionnaire. The contribution comes from presenting new planning units in away suits the current context of the old cities in MENA region based on ten-minute walking or less distance with a reference to the right to the city. This planning unit can find it way to citizens' happiness.

Keywords: happiness, ten-minute neighbourhood, urban design, well-being

Procedia PDF Downloads 384
11274 Analyzing the Perceptions of Accounting Practitioners regarding Communication Skills of Distance-Learning Graduates

Authors: Carol S. Binnekade, Deon Scott, Christina C. Shuttleworth, Annelien A. Van Rooyen

Abstract:

Higher education institutions are constantly challenged to deliver skilled graduates into the workplace. Employers expect graduates to have the required technical knowledge as well as various pervasive skills. This also applies to accountants who need to know the technical requirements of financial reporting and be able to communicate with individuals, teams and clients at a high level. Accountants need to develop effective business conversational skills and use these skills to communicate up, down and across organizations, taking into consideration cultural and gender diversity. In addition, they need to master business writing and presentation skills. However, providing students with these skills in a distance-learning environment where interaction between students and instructors is limited, is a challenge for academics. The study on which this paper reports, forms part of a larger body of research, which explored the perceptions of accounting practitioners of the communication skills (or lack thereof) of recently qualified accounting students. Feedback (qualitative and quantitative) was obtained from various accounting practitioners in South Africa. Taking into consideration that distance learners communicate mainly with their instructors via email communication and their assignments are submitted using various word processor software, the researchers were of the opinion that the accounting graduates would be capable of communicating effectively once they entered the workplace. However, the research findings, inter alia, suggested that the accounting graduates lacked communication skills and that training was needed to differentiate between business and social communication once they entered the workplace. Recommendations on how these communication challenges may be addressed by higher education institutions are provided.

Keywords: accounting practitioners, communication skills, distance education, pervasive skills

Procedia PDF Downloads 191
11273 Design of an Innovative Geothermal Heat Pump with a PCM Thermal Storage

Authors: Emanuele Bonamente, Andrea Aquino

Abstract:

This study presents an innovative design for geothermal heat pumps with the goal of maximizing the system efficiency (COP - Coefficient of Performance), reducing the soil use (e.g. length/depth of geothermal boreholes) and initial investment costs. Based on experimental data obtained from a two-year monitoring of a working prototype implemented for a commercial building in the city of Perugia, Italy, an upgrade of the system is proposed and the performance is evaluated via CFD simulations. The prototype was designed to include a thermal heat storage (i.e. water), positioned between the boreholes and the heat pump, acting as a flywheel. Results from the monitoring campaign show that the system is still capable of providing the required heating and cooling energy with a reduced geothermal installation (approx. 30% of the standard length). In this paper, an optimization of the system is proposed, re-designing the heat storage to include phase change materials (PCMs). Two stacks of PCMs, characterized by melting temperatures equal to those needed to maximize the system COP for heating and cooling, are disposed within the storage. During the working cycle, the latent heat of the PCMs is used to heat (cool) the water used by the heat pump while the boreholes independently cool (heat) the storage. The new storage is approximately 10 times smaller and can be easily placed close to the heat pump in the technical room. First, a validation of the CFD simulation of the storage is performed against experimental data. The simulation is then used to test possible alternatives of the original design and it is finally exploited to evaluate the PCM-storage performance for two different configurations (i.e. single- and double-loop systems).

Keywords: geothermal heat pump, phase change materials (PCM), energy storage, renewable energies

Procedia PDF Downloads 297
11272 Application of the Standard Deviation in Regulating Design Variation of Urban Solutions Generated through Evolutionary Computation

Authors: Mohammed Makki, Milad Showkatbakhsh, Aiman Tabony

Abstract:

Computational applications of natural evolutionary processes as problem-solving tools have been well established since the mid-20th century. However, their application within architecture and design has only gained ground in recent years, with an increasing number of academics and professionals in the field electing to utilize evolutionary computation to address problems comprised from multiple conflicting objectives with no clear optimal solution. Recent advances in computer science and its consequent constructive influence on the architectural discourse has led to the emergence of multiple algorithmic processes capable of simulating the evolutionary process in nature within an efficient timescale. Many of the developed processes of generating a population of candidate solutions to a design problem through an evolutionary based stochastic search process are often driven through the application of both environmental and architectural parameters. These methods allow for conflicting objectives to be simultaneously, independently, and objectively optimized. This is an essential approach in design problems with a final product that must address the demand of a multitude of individuals with various requirements. However, one of the main challenges encountered through the application of an evolutionary process as a design tool is the ability for the simulation to maintain variation amongst design solutions in the population while simultaneously increasing in fitness. This is most commonly known as the ‘golden rule’ of balancing exploration and exploitation over time; the difficulty of achieving this balance in the simulation is due to the tendency of either variation or optimization being favored as the simulation progresses. In such cases, the generated population of candidate solutions has either optimized very early in the simulation, or has continued to maintain high levels of variation to which an optimal set could not be discerned; thus, providing the user with a solution set that has not evolved efficiently to the objectives outlined in the problem at hand. As such, the experiments presented in this paper seek to achieve the ‘golden rule’ by incorporating a mathematical fitness criterion for the development of an urban tissue comprised from the superblock as its primary architectural element. The mathematical value investigated in the experiments is the standard deviation factor. Traditionally, the standard deviation factor has been used as an analytical value rather than a generative one, conventionally used to measure the distribution of variation within a population by calculating the degree by which the majority of the population deviates from the mean. A higher standard deviation value delineates a higher number of the population is clustered around the mean and thus limited variation within the population, while a lower standard deviation value is due to greater variation within the population and a lack of convergence towards an optimal solution. The results presented will aim to clarify the extent to which the utilization of the standard deviation factor as a fitness criterion can be advantageous to generating fitter individuals in a more efficient timeframe when compared to conventional simulations that only incorporate architectural and environmental parameters.

Keywords: architecture, computation, evolution, standard deviation, urban

Procedia PDF Downloads 121