Search results for: dynamic planning
622 Digital Transformation of Lean Production: Systematic Approach for the Determination of Digitally Pervasive Value Chains
Authors: Peter Burggräf, Matthias Dannapfel, Hanno Voet, Patrick-Benjamin Bök, Jérôme Uelpenich, Julian Hoppe
Abstract:
The increasing digitalization of value chains can help companies to handle rising complexity in their processes and thereby reduce the steadily increasing planning and control effort in order to raise performance limits. Due to technological advances, companies face the challenge of smart value chains for the purpose of improvements in productivity, handling the increasing time and cost pressure and the need of individualized production. Therefore, companies need to ensure quick and flexible decisions to create self-optimizing processes and, consequently, to make their production more efficient. Lean production, as the most commonly used paradigm for complexity reduction, reaches its limits when it comes to variant flexible production and constantly changing market and environmental conditions. To lift performance limits, which are inbuilt in current value chains, new methods and tools must be applied. Digitalization provides the potential to derive these new methods and tools. However, companies lack the experience to harmonize different digital technologies. There is no practicable framework, which instructs the transformation of current value chains into digital pervasive value chains. Current research shows that a connection between lean production and digitalization exists. This link is based on factors such as people, technology and organization. In this paper, the introduced method for the determination of digitally pervasive value chains takes the factors people, technology and organization into account and extends existing approaches by a new dimension. It is the first systematic approach for the digital transformation of lean production and consists of four steps: The first step of ‘target definition’ describes the target situation and defines the depth of the analysis with regards to the inspection area and the level of detail. The second step of ‘analysis of the value chain’ verifies the lean-ability of processes and lies in a special focus on the integration capacity of digital technologies in order to raise the limits of lean production. Furthermore, the ‘digital evaluation process’ ensures the usefulness of digital adaptions regarding their practicability and their integrability into the existing production system. Finally, the method defines actions to be performed based on the evaluation process and in accordance with the target situation. As a result, the validation and optimization of the proposed method in a German company from the electronics industry shows that the digital transformation of current value chains based on lean production achieves a raise of their inbuilt performance limits.Keywords: digitalization, digital transformation, Industrie 4.0, lean production, value chain
Procedia PDF Downloads 313621 Computational Insights Into Allosteric Regulation of Lyn Protein Kinase: Structural Dynamics and Impacts of Cancer-Related Mutations
Authors: Mina Rabipour, Elena Pallaske, Floyd Hassenrück, Rocio Rebollido-Rios
Abstract:
Protein tyrosine kinases, including Lyn kinase of the Src family kinases (SFK), regulate cell proliferation, survival, and differentiation. Lyn kinase has been implicated in various cancers, positioning it as a promising therapeutic target. However, the conserved ATP-binding pocket across SFKs makes developing selective inhibitors challenging. This study aims to address this limitation by exploring the potential for allosteric modulation of Lyn kinase, focusing on how its structural dynamics and specific oncogenic mutations impact its conformation and function. To achieve this, we combined homology modeling, molecular dynamics simulations, and data science techniques to conduct microsecond-length simulations. Our approach allowed a detailed investigation into the interplay between Lyn’s catalytic and regulatory domains, identifying key conformational states involved in allosteric regulation. Additionally, we evaluated the structural effects of Dasatinib, a competitive inhibitor, and ATP binding on Lyn active conformation. Notably, our simulations show that cancer-related mutations, specifically I364L/N and E290D/K, shift Lyn toward an inactive conformation, contrasting with the active state of the wild-type protein. This may suggest how these mutations contribute to aberrant signaling in cancer cells. We conducted a dynamical network analysis to assess residue-residue interactions and the impact of mutations on the Lyn intramolecular network. This revealed significant disruptions due to mutations, especially in regions distant from the ATP-binding site. These disruptions suggest potential allosteric sites as therapeutic targets, offering an alternative strategy for Lyn inhibition with higher specificity and fewer off-target effects compared to ATP-competitive inhibitors. Our findings provide insights into Lyn kinase regulation and highlight allosteric sites as avenues for selective drug development. Targeting these sites may modulate Lyn activity in cancer cells, reducing toxicity and improving outcomes. Furthermore, our computational strategy offers a scalable approach for analyzing other SFK members or kinases with similar properties, facilitating the discovery of selective allosteric modulators and contributing to precise cancer therapies.Keywords: lyn tyrosine kinase, mutation analysis, conformational changes, dynamic network analysis, allosteric modulation, targeted inhibition
Procedia PDF Downloads 17620 Impact of Transitioning to Renewable Energy Sources on Key Performance Indicators and Artificial Intelligence Modules of Data Center
Authors: Ahmed Hossam ElMolla, Mohamed Hatem Saleh, Hamza Mostafa, Lara Mamdouh, Yassin Wael
Abstract:
Artificial intelligence (AI) is reshaping industries, and its potential to revolutionize renewable energy and data center operations is immense. By harnessing AI's capabilities, we can optimize energy consumption, predict fluctuations in renewable energy generation, and improve the efficiency of data center infrastructure. This convergence of technologies promises a future where energy is managed more intelligently, sustainably, and cost-effectively. The integration of AI into renewable energy systems unlocks a wealth of opportunities. Machine learning algorithms can analyze vast amounts of data to forecast weather patterns, solar irradiance, and wind speeds, enabling more accurate energy production planning. AI-powered systems can optimize energy storage and grid management, ensuring a stable power supply even during intermittent renewable generation. Moreover, AI can identify maintenance needs for renewable energy infrastructure, preventing costly breakdowns and maximizing system lifespan. Data centers, which consume substantial amounts of energy, are prime candidates for AI-driven optimization. AI can analyze energy consumption patterns, identify inefficiencies, and recommend adjustments to cooling systems, server utilization, and power distribution. Predictive maintenance using AI can prevent equipment failures, reducing energy waste and downtime. Additionally, AI can optimize data placement and retrieval, minimizing energy consumption associated with data transfer. As AI transforms renewable energy and data center operations, modified Key Performance Indicators (KPIs) will emerge. Traditional metrics like energy efficiency and cost-per-megawatt-hour will continue to be relevant, but additional KPIs focused on AI's impact will be essential. These might include AI-driven cost savings, predictive accuracy of energy generation and consumption, and the reduction of carbon emissions attributed to AI-optimized operations. By tracking these KPIs, organizations can measure the success of their AI initiatives and identify areas for improvement. Ultimately, the synergy between AI, renewable energy, and data centers holds the potential to create a more sustainable and resilient future. By embracing these technologies, we can build smarter, greener, and more efficient systems that benefit both the environment and the economy.Keywords: data center, artificial intelligence, renewable energy, energy efficiency, sustainability, optimization, predictive analytics, energy consumption, energy storage, grid management, data center optimization, key performance indicators, carbon emissions, resiliency
Procedia PDF Downloads 36619 Enhancing Mental Health Services Through Strategic Planning: The East Tennessee State University Counseling Center’s 2024-2028 Plan
Authors: R. M. Kilonzo, S. Bedingfield, K. Smith, K. Hudgins Smith, K. Couper, R. Ratley, Z. Taylor, A. Engelman, M. Renne
Abstract:
Introduction: The mental health needs of university students continue to evolve, necessitating a strategic approach to service delivery. The East Tennessee State University (ETSU) Counseling Center developed its inaugural Strategic Plan (2024-2028) to enhance student mental health services. The plan focuses on improving access, quality of care, and service visibility, aligning with the university’s mission to support academic success and student well-being. Aim: This strategic plan aims to establish a comprehensive framework for delivering high-quality, evidence-based mental health services to ETSU students, addressing current challenges, and anticipating future needs. Methods: The development of the strategic plan was a collaborative effort involving the Counseling Center’s leadership, staff, with technical support from Doctor of Public Health-community and behavioral health intern. Multiple workshops, online/offline reviews, and stakeholder consultations were held to ensure a robust and inclusive process. A SWOT analysis and stakeholder mapping were conducted to identify strengths, weaknesses, opportunities, and challenges. Key performance indicators (KPIs) were set to measure service utilization, satisfaction, and outcomes. Results: The plan resulted in four strategic priorities: service application, visibility/accessibility, safety and satisfaction, and training programs. Key objectives include expanding counseling services, improving service access through outreach, reducing stigma, and increasing peer support programs. The plan also focuses on continuous quality improvement through data-driven assessments and research initiatives. Immediate outcomes include expanded group therapy, enhanced staff training, and increased mental health literacy across campus. Conclusion and Recommendation: The strategic plan provides a roadmap for addressing the mental health needs of ETSU students, with a clear focus on accessibility, inclusivity, and evidence-based practices. Implementing the plan will strengthen the Counseling Center’s capacity to meet the diverse needs of the student population. To ensure sustainability, it is recommended that the center continuously assess student needs, foster partnerships with university and external stakeholders, and advocate for increased funding to expand services and staff capacity.Keywords: strategic plan, university counseling center, mental health, students
Procedia PDF Downloads 21618 A Development of Practice Guidelines for Surgical Safety Management to Reduce Undesirable Incidents from Surgical Services in the Operating Room of Songkhla Hospital, Thailand
Authors: Thitima Plejai
Abstract:
The practice in the operating room has been continually performed according to standards of services; however, undesirable incidents from surgical services are found such as surgical complications in the operating room. This participation action research aimed to develop practice guidelines for surgical safety management to reduce undesirable incidents from surgical services in the operating room of Songkhla Hospital. The target population was all 84 members of the multidisciplinary team who were involved in surgical services in the operating room consisting of 28 surgeons from five branches of surgery, 27 anesthetists and nurse anesthetists, and 29 surgical nurses. The data were collected through in-depth interviews, and non-participatory observations. The research instrument was tested by three experts, and the steps of the development consisted of four cycles, each consisting of assessment, planning, practice, practice reflection, and improvement until every step is practicable. The data were validated through triangulation research method, analyzed through content analysis and statistical analysis with number and percentage. The results of the development of practice guidelines surgical safety management to reduce undesirable incidents from surgical services could be concluded as follows. 1) The multidisciplinary team in surgery participated in the needs assessment for development of practice guidelines for surgical patient safety, and agreed on adapting the WHO Surgical Safety Checklists for use. 2) The WHO Surgical Safety Checklists was implemented, and meetings were held for the multidisciplinary team in surgery and the organizational risk committee to improve the practice guidelines to make them more practicable. 3) The multidisciplinary team consisting of surgeons from five branches of surgery, anesthetists, nurse anesthetists, surgical nurses, and the organizational risk committee announced policy on safety for surgical patients; the organizational risk committee designated the Surgical Safety Checklist as an instrument for surgical patient safety. The results of the safety management found that the surgical team members who could follow 100 percent of the guidelines were: professional nurses who checked patient identity and information before taking the patient to the operating room and kept complete records of data on the patients; surgical nurses who checked readiness of the patient before surgery; nurse anesthetists who assessed readiness before administering anesthetic drugs, and confirmed correctness of the patient; and circulating perioperative nurses who gave confirmation to the surgical team after completion of the surgery. The rates of undesirable incidents (surgical complications rates) before and after the implementation of the surgical safety management were 1.60 percent and 0.66 percent, respectively. The satisfaction of the surgery-related teams towards the use of the guidelines was 89 percent. The practice guidelines for surgical safety management to reduce undesirable incidents were taken as guidelines for surgical safety that the multidisciplinary team involved in the surgical process implemented correctly and in the same direction and clearly reduced undesirable incidents in surgical patients.Keywords: practice guidelines, surgical safety management, reduce undesirable incidents, operating Room
Procedia PDF Downloads 297617 Modeling of the Biodegradation Performance of a Membrane Bioreactor to Enhance Water Reuse in Agri-food Industry - Poultry Slaughterhouse as an Example
Authors: masmoudi Jabri Khaoula, Zitouni Hana, Bousselmi Latifa, Akrout Hanen
Abstract:
Mathematical modeling has become an essential tool for sustainable wastewater management, particularly for the simulation and the optimization of complex processes involved in activated sludge systems. In this context, the activated sludge model (ASM3h) was used for the simulation of a Biological Membrane Reactor (MBR) as it includes the integration of biological wastewater treatment and physical separation by membrane filtration. In this study, the MBR with a useful volume of 12.5 L was fed continuously with poultry slaughterhouse wastewater (PSWW) for 50 days at a feed rate of 2 L/h and for a hydraulic retention time (HRT) of 6.25h. Throughout its operation, High removal efficiency was observed for the removal of organic pollutants in terms of COD with 84% of efficiency. Moreover, the MBR has generated a treated effluent which fits with the limits of discharge into the public sewer according to the Tunisian standards which were set in March 2018. In fact, for the nitrogenous compounds, average concentrations of nitrate and nitrite in the permeat reached 0.26±0.3 mg. L-1 and 2.2±2.53 mg. L-1, respectively. The simulation of the MBR process was performed using SIMBA software v 5.0. The state variables employed in the steady state calibration of the ASM3h were determined using physical and respirometric methods. The model calibration was performed using experimental data obtained during the first 20 days of the MBR operation. Afterwards, kinetic parameters of the model were adjusted and the simulated values of COD, N-NH4+and N- NOx were compared with those reported from the experiment. A good prediction was observed for the COD, N-NH4+and N- NOx concentrations with 467 g COD/m³, 110.2 g N/m³, 3.2 g N/m³ compared to the experimental data which were 436.4 g COD/m³, 114.7 g N/m³ and 3 g N/m³, respectively. For the validation of the model under dynamic simulation, the results of the experiments obtained during the second treatment phase of 30 days were used. It was demonstrated that the model simulated the conditions accurately by yielding a similar pattern on the variation of the COD concentration. On the other hand, an underestimation of the N-NH4+ concentration was observed during the simulation compared to the experimental results and the measured N-NO3 concentrations were lower than the predicted ones, this difference could be explained by the fact that the ASM models were mainly designed for the simulation of biological processes in the activated sludge systems. In addition, more treatment time could be required by the autotrophic bacteria to achieve a complete and stable nitrification. Overall, this study demonstrated the effectiveness of mathematical modeling in the prediction of the performance of the MBR systems with respect to organic pollution, the model can be further improved for the simulation of nutrients removal for a longer treatment period.Keywords: activated sludge model (ASM3h), membrane bioreactor (MBR), poultry slaughter wastewater (PSWW), reuse
Procedia PDF Downloads 60616 Weapon-Being: Weaponized Design and Object-Oriented Ontology in Hypermodern Times
Authors: John Dimopoulos
Abstract:
This proposal attempts a refabrication of Heidegger’s classic thing-being and object-being analysis in order to provide better ontological tools for understanding contemporary culture, technology, and society. In his work, Heidegger sought to understand and comment on the problem of technology in an era of rampant innovation and increased perils for society and the planet. Today we seem to be at another crossroads in this course, coming after postmodernity, during which dreams and dangers of modernity augmented with critical speculations of the post-war era take shape. The new era which we are now living in, referred to as hypermodernity by researchers in various fields such as architecture and cultural theory, is defined by the horizontal implementation of digital technologies, cybernetic networks, and mixed reality. Technology today is rapidly approaching a turning point, namely the point of no return for humanity’s supervision over its creations. The techno-scientific civilization of the 21st century creates a series of problems, progressively more difficult and complex to solve and impossible to ignore, climate change, data safety, cyber depression, and digital stress being some of the most prevalent. Humans often have no other option than to address technology-induced problems with even more technology, as in the case of neuron networks, machine learning, and AI, thus widening the gap between creating technological artifacts and understanding their broad impact and possible future development. As all technical disciplines and particularly design, become enmeshed in a matrix of digital hyper-objects, a conceptual toolbox that allows us to handle the new reality becomes more and more necessary. Weaponized design, prevalent in many fields, such as social and traditional media, urban planning, industrial design, advertising, and the internet in general, hints towards an increase in conflicts. These conflicts between tech companies, stakeholders, and users with implications in politics, work, education, and production as apparent in the cases of Amazon workers’ strikes, Donald Trump’s 2016 campaign, Facebook and Microsoft data scandals, and more are often non-transparent to the wide public’s eye, thus consolidating new elites and technocratic classes and making the public scene less and less democratic. The new category proposed, weapon-being, is outlined in respect to the basic function of reducing complexity, subtracting materials, actants, and parameters, not strictly in favor of a humanistic re-orientation but in a more inclusive ontology of objects and subjects. Utilizing insights of Object-Oriented Ontology (OOO) and its schematization of technological objects, an outline for a radical ontology of technology is approached.Keywords: design, hypermodernity, object-oriented ontology, weapon-being
Procedia PDF Downloads 153615 Mechanism Design and Dynamic Analysis of Active Independent Front Steering System
Authors: Cheng-Chi Yu, Yu-Shiue Wang, Kei-Lin Kuo
Abstract:
Active Independent Front Steering system is a steering system which can according to vehicle driving situation adjusts the relation of steering angle between inner wheel and outer wheel. In low-speed cornering, AIFS sets the steering angles of inner and outer wheel into Ackerman steering geometry to make vehicle has less cornering radius. Besides, AIFS changes the steering geometry to parallel or even anti-Ackerman steering geometry to keep vehicle stability in high-speed cornering. Therefore, based on the analysis of the vehicle steering behavior from different steering geometries, this study develops a new screw type of active independent front steering system to make vehicles best cornering performance at any speeds. The screw type of active independent front steering system keeps the pinion and separates the rack into main rack and second rack. Two racks connect by a screw. Extra screw rotated motion powered by assistant motor through coupler makes second rack move relative to main rack, which can adjust both steering ratio and steering geometry. First of all, this study distinguishes the steering geometry by using Ackerman percentage and utilizes the software of ADAMS/Car to construct diverse steering geometry models. The different steering geometries are compared at low-speed and high-speed cornering, and then control strategies of the active independent front steering systems could be formulated. Secondly, this study applies closed loop equation to analyze tire steering angles and carries out optimization calculations to make the steering geometry from traditional rack and pinion steering system near to Ackerman steering geometry. Steering characteristics of the optimum steering mechanism and motion characteristics of vehicle installed the steering mechanism are verified by ADAMS/Car models of front suspension and full vehicle respectively. By adding dual auxiliary rack and dual motor to the optimum steering mechanism, the active independent front steering system could be developed to achieve the functions of variable steering ratio and variable steering geometry. At last, this study uses ADAMS/Car and Matlab/Simulink to co-simulate the cornering motion of vehicles confirms the vehicle installed the Active Independent Front Steering (AIFS) system has better handling performance than that with Active Independent Steering (AFS) system or with Electric Power Steering (EPS) system. At low-speed cornering, the vehicles with AIFS system and with AFS system have better maneuverability, less cornering radius, than the traditional vehicle with EPS system because that AIFS and AFS systems both provide function of variable steering ratio. However, there is a slight penalty in the motor(s) power consumption. In addition, because of the capability of variable steering geometry, the vehicle with AIFS system has better high-speed cornering stability, trajectory keeping, and even less motor(s) power consumption than that with EPS system and also with AFS system.Keywords: active front steering system, active independent front steering system, steering geometry, steering ratio
Procedia PDF Downloads 190614 Characterization of Herberine Hydrochloride Nanoparticles
Authors: Bao-Fang Wen, Meng-Na Dai, Gao-Pei Zhu, Chen-Xi Zhang, Jing Sun, Xun-Bao Yin, Yu-Han Zhao, Hong-Wei Sun, Wei-Fen Zhang
Abstract:
A drug-loaded nanoparticles containing berberine hydrochloride (BH/FA-CTS-NPs) was prepared. The physicochemical characterizations of BH/FA-CTS-NPs and the inhibitory effect on the HeLa cells were investigated. Folic acid-conjugated chitosan (FA-CTS) was prepared by amino reaction of folic acid active ester and chitosan molecules; BH/FA-CTS-NPs were prepared using ionic cross-linking technique with BH as a model drug. The morphology and particle size were determined by Transmission Electron Microscope (TEM). The average diameters and polydispersity index (PDI) were evaluated by Dynamic Light Scattering (DLS). The interaction between various components and the nanocomplex were characterized by Fourier Transform Infrared Spectroscopy (FT-IR). The entrapment efficiency (EE), drug-loading (DL) and in vitro release were studied by UV spectrophotometer. The effect of cell anti-migratory and anti-invasive actions of BH/FA-CTS-NPs were investigated using MTT assays, wound healing assays, Annexin-V-FITC single staining assays, and flow cytometry, respectively. HeLa nude mice subcutaneously transplanted tumor model was established and treated with different drugs to observe the effect of BH/FA-CTS-NPs in vivo on HeLa bearing tumor. The BH/FA-CTS-NPs prepared in this experiment have a regular shape, uniform particle size, and no aggregation phenomenon. The results of DLS showed that mean particle size, PDI and Zeta potential of BH/FA-CTS NPs were (249.2 ± 3.6) nm, 0.129 ± 0.09, 33.6 ± 2.09, respectively, and the average diameter and PDI were stable in 90 days. The results of FT-IR demonstrated that the characteristic peaks of FA-CTS and BH/FA-CTS-NPs confirmed that FA-CTS cross-linked successfully and BH was encapsulated in NPs. The EE and DL amount were (79.3 ± 3.12) % and (7.24 ± 1.41) %, respectively. The results of in vitro release study indicated that the cumulative release of BH/FA-CTS NPs was (89.48±2.81) % in phosphate-buffered saline (PBS, pH 7.4) within 48h; these results by MTT assays and wund healing assays indicated that BH/FA-CTS NPs not only inhibited the proliferation of HeLa cells in a concentration and time-dependent manner but can induce apoptosis as well. The subcutaneous xenograft tumor formation rate of human cervical cancer cell line HeLa in nude mice was 98% after inoculation for 2 weeks. Compared with BH group and BH/CTS-NPs group, the xenograft tumor growth of BH/FA-CTS-NPs group was obviously slower; the result indicated that BH/FA-CTS-NPs could significantly inhibit the growth of HeLa xenograft tumor. BH/FA-CTS NPs with the sustained release effect could be prepared successfully by the ionic crosslinking method. Considering these properties, block proliferation and impairing the migration of the HeLa cell line, BH/FA-CTS NPs could be an important compound for consideration in the treatment of cervical cancer.Keywords: folic-acid, chitosan, berberine hydrochloride, nanoparticles, cervical cancer
Procedia PDF Downloads 122613 Modeling of Tsunami Propagation and Impact on West Vancouver Island, Canada
Authors: S. Chowdhury, A. Corlett
Abstract:
Large tsunamis strike the British Columbia coast every few hundred years. The Cascadia Subduction Zone, which extends along the Pacific coast from Vancouver Island to Northern California is one of the most seismically active regions in Canada. Significant earthquakes have occurred in this region, including the 1700 Cascade Earthquake with an estimated magnitude of 9.2. Based on geological records, experts have predicted a 'great earthquake' of a similar magnitude within this region may happen any time. This earthquake is expected to generate a large tsunami that could impact the coastal communities on Vancouver Island. Since many of these communities are in remote locations, they are more likely to be vulnerable, as the post-earthquake relief efforts would be impacted by the damage to critical road infrastructures. To assess the coastal vulnerability within these communities, a hydrodynamic model has been developed using MIKE-21 software. We have considered a 500 year probabilistic earthquake design criteria including the subsidence in this model. The bathymetry information was collected from Canadian Hydrographic Services (CHS), and National Oceanic Atmospheric and Administration (NOAA). The arial survey was conducted using a Cessna-172 aircraft for the communities, and then the information was converted to generate a topographic digital elevation map. Both survey information was incorporated into the model, and the domain size of the model was about 1000km x 1300km. This model was calibrated with the tsunami occurred off the west coast of Moresby Island on October 28, 2012. The water levels from the model were compared with two tide gauge stations close to the Vancouver Island and the output from the model indicates the satisfactory result. For this study, the design water level was considered as High Water Level plus the Sea Level Rise for 2100 year. The hourly wind speeds from eight directions were collected from different wind stations and used a 200-year return period wind speed in the model for storm events. The regional model was set for 12 hrs simulation period, which takes more than 16 hrs to complete one simulation using double Xeon-E7 CPU computer plus a K-80 GPU. The boundary information for the local model was generated from the regional model. The local model was developed using a high resolution mesh to estimate the coastal flooding for the communities. It was observed from this study that many communities will be effected by the Cascadia tsunami and the inundation maps were developed for the communities. The infrastructures inside the coastal inundation area were identified. Coastal vulnerability planning and resilient design solutions will be implemented to significantly reduce the risk.Keywords: tsunami, coastal flooding, coastal vulnerable, earthquake, Vancouver, wave propagation
Procedia PDF Downloads 132612 Comparative Analysis of Simulation-Based and Mixed-Integer Linear Programming Approaches for Optimizing Building Modernization Pathways Towards Decarbonization
Authors: Nico Fuchs, Fabian Wüllhorst, Laura Maier, Dirk Müller
Abstract:
The decarbonization of building stocks necessitates the modernization of existing buildings. Key measures for this include reducing energy demands through insulation of the building envelope, replacing heat generators, and installing solar systems. Given limited financial resources, it is impractical to modernize all buildings in a portfolio simultaneously; instead, prioritization of buildings and modernization measures for a given planning horizon is essential. Optimization models for modernization pathways can assist portfolio managers in this prioritization. However, modeling and solving these large-scale optimization problems, often represented as mixed-integer problems (MIP), necessitates simplifying the operation of building energy systems particularly with respect to system dynamics and transient behavior. This raises the question of which level of simplification remains sufficient to accurately account for realistic costs and emissions of building energy systems, ensuring a fair comparison of different modernization measures. This study addresses this issue by comparing a two-stage simulation-based optimization approach with a single-stage mathematical optimization in a mixed-integer linear programming (MILP) formulation. The simulation-based approach serves as a benchmark for realistic energy system operation but requires a restriction of the solution space to discrete choices of modernization measures, such as the sizing of heating systems. After calculating the operation of different energy systems in terms of the resulting final energy demands in simulation models on a first stage, the results serve as input for a second stage MILP optimization, where the design of each building in the portfolio is optimized. In contrast to the simulation-based approach, the MILP-based approach can capture a broader variety of modernization measures due to the efficiency of MILP solvers but necessitates simplifying the building energy system operation. Both approaches are employed to determine the cost-optimal design and dimensioning of several buildings in a portfolio to meet climate targets within limited yearly budgets, resulting in a modernization pathway for the entire portfolio. The comparison reveals that the MILP formulation successfully captures design decisions of building energy systems, such as the selection of heating systems and the modernization of building envelopes. However, the results regarding the optimal dimensioning of heating technologies differ from the results of the two-stage simulation-based approach, as the MILP model tends to overestimate operational efficiency, highlighting the limitations of the MILP approach.Keywords: building energy system optimization, model accuracy in optimization, modernization pathways, building stock decarbonization
Procedia PDF Downloads 36611 The Analysis of Gizmos Online Program as Mathematics Diagnostic Program: A Story from an Indonesian Private School
Authors: Shofiayuningtyas Luftiani
Abstract:
Some private schools in Indonesia started integrating the online program Gizmos in the teaching-learning process. Gizmos was developed to supplement the existing curriculum by integrating it into the instructional programs. The program has some features using an inquiry-based simulation, in which students conduct exploration by using a worksheet while teachers use the teacher guidelines to direct and assess students’ performance In this study, the discussion about Gizmos highlights its features as the assessment media of mathematics learning for secondary school students. The discussion is based on the case study and literature review from the Indonesian context. The purpose of applying Gizmos as an assessment media refers to the diagnostic assessment. As a part of the diagnostic assessment, the teachers review the student exploration sheet, analyze particularly in the students’ difficulties and consider findings in planning future learning process. This assessment becomes important since the teacher needs the data about students’ persistent weaknesses. Additionally, this program also helps to build student’ understanding by its interactive simulation. Currently, the assessment over-emphasizes the students’ answers in the worksheet based on the provided answer keys while students perform their skill in translating the question, doing the simulation and answering the question. Whereas, the assessment should involve the multiple perspectives and sources of students’ performance since teacher should adjust the instructional programs with the complexity of students’ learning needs and styles. Consequently, the approach to improving the assessment components is selected to challenge the current assessment. The purpose of this challenge is to involve not only the cognitive diagnosis but also the analysis of skills and error. Concerning the selected setting for this diagnostic assessment that develops the combination of cognitive diagnosis, skills analysis and error analysis, the teachers should create an assessment rubric. The rubric plays the important role as the guide to provide a set of criteria for the assessment. Without the precise rubric, the teacher potentially ineffectively documents and follows up the data about students at risk of failure. Furthermore, the teachers who employ the program of Gizmos as the diagnostic assessment might encounter some obstacles. Based on the condition of assessment in the selected setting, the obstacles involve the time constrain, the reluctance of higher teaching burden and the students’ behavior. Consequently, the teacher who chooses the Gizmos with those approaches has to plan, implement and evaluate the assessment. The main point of this assessment is not in the result of students’ worksheet. However, the diagnostic assessment has the two-stage process; the process to prompt and effectively follow-up both individual weaknesses and those of the learning process. Ultimately, the discussion of Gizmos as the media of the diagnostic assessment refers to the effort to improve the mathematical learning process.Keywords: diagnostic assessment, error analysis, Gizmos online program, skills analysis
Procedia PDF Downloads 183610 How to Talk about It without Talking about It: Cognitive Processing Therapy Offers Trauma Symptom Relief without Violating Cultural Norms
Authors: Anne Giles
Abstract:
Humans naturally wish they could forget traumatic experiences. To help prevent future harm, however, the human brain has evolved to retain data about experiences of threat, alarm, or violation. When given compassionate support and assistance with thinking helpfully and realistically about traumatic events, most people can adjust to experiencing hardships, albeit with residual sad, unfortunate memories. Persistent, recurrent, intrusive memories, difficulty sleeping, emotion dysregulation, and avoidance of reminders, however, may be symptoms of Post-traumatic Stress Disorder (PTSD). Brain scans show that PTSD affects brain functioning. We currently have no physical means of restoring the system of brain structures and functions involved with PTSD. Medications may ease some symptoms but not others. However, forms of "talk therapy" with cognitive components have been found by researchers to reduce, even resolve, a broad spectrum of trauma symptoms. Many cultures have taboos against talking about hardships. Individuals may present themselves to mental health care professionals with severe, disabling trauma symptoms but, because of cultural norms, be unable to speak about them. In China, for example, relationship expectations may include the belief, "Bad things happening in the family should stay in the family (jiāchǒu bùkě wàiyán 家丑不可外扬)." The concept of "family (jiā 家)" may include partnerships, close and extended families, communities, companies, and the nation itself. In contrast to many trauma therapies, Cognitive Processing Therapy (CPT) for Post-traumatic Stress Disorder asks its participants to focus not on "what" happened but on "why" they think the trauma(s) occurred. The question "why" activates and exercises cognitive functioning. Brain scans of individuals with PTSD reveal executive functioning portions of the brain inadequately active, with emotion centers overly active. CPT conceptualizes PTSD as a network of cognitive distortions that keep an individual "stuck" in this under-functioning and over-functioning dynamic. Through asking participants forms of the question "why," plus offering a protocol for examining answers and relinquishing unhelpful beliefs, CPT assists individuals in consciously reactivating the cognitive, executive functions of their brains, thus restoring normal functioning and reducing distressing trauma symptoms. The culturally sensitive components of CPT that allow people to "talk about it without talking about it" may offer the possibility for worldwide relief from symptoms of trauma.Keywords: cognitive processing therapy (CPT), cultural norms, post-traumatic stress disorder (PTSD), trauma recovery
Procedia PDF Downloads 215609 Evaluating Daylight Performance in an Office Environment in Malaysia, Using Venetian Blind System: Case Study
Authors: Fatemeh Deldarabdolmaleki, Mohamad Fakri Zaky Bin Ja'afar
Abstract:
Having a daylit space together with view results in a pleasant and productive environment for office employees. A daylit space is a space which utilizes daylight as a basic source of illumination to fulfill user’s visual demands and minimizes the electric energy consumption. Malaysian weather is hot and humid all over the year because of its location in the equatorial belt. however, because most of the commercial buildings in Malaysia are air-conditioned, huge glass windows are normally installed in order to keep the physical and visual relation between inside and outside. As a result of climatic situation and mentioned new trend, an ordinary office has huge heat gain, glare, and discomfort for occupants. Balancing occupant’s comfort and energy conservation in a tropical climate is a real challenge. This study concentrates on evaluating a venetian blind system using per pixel analyzing tools based on the suggested cut-out metrics by the literature. Workplace area in a private office room has been selected as a case study. Eight-day measurement experiment was conducted to investigate the effect of different venetian blind angles in an office area under daylight conditions in Serdang, Malaysia. The study goal was to explore daylight comfort of a commercially available venetian blind system, its’ daylight sufficiency and excess (8:00 AM to 5 PM) as well as Glare examination. Recently developed software, analyzing High Dynamic Range Images (HDRI captured by CCD camera), such as radiance based Evalglare and hdrscope help to investigate luminance-based metrics. The main key factors are illuminance and luminance levels, mean and maximum luminance, daylight glare probability (DGP) and luminance ratio of the selected mask regions. The findings show that in most cases, morning session needs artificial lighting in order to achieve daylight comfort. However, in some conditions (e.g. 10° and 40° slat angles) in the second half of day the workplane illuminance level exceeds the maximum of 2000 lx. Generally, a rising trend is discovered toward mean window luminance and the most unpleasant cases occur after 2 P.M. Considering the luminance criteria rating, the uncomfortable conditions occur in the afternoon session. Surprisingly in no blind condition, extreme case of window/task ratio is not common. Studying the daylight glare probability, there is not any DGP value higher than 0.35 in this experiment.Keywords: daylighting, energy simulation, office environment, Venetian blind
Procedia PDF Downloads 260608 Adaptive Environmental Control System Strategy for Cabin Air Quality in Commercial Aircrafts
Authors: Paolo Grasso, Sai Kalyan Yelike, Federico Benzi, Mathieu Le Cam
Abstract:
The cabin air quality (CAQ) in commercial aircraft is of prime interest, especially in the context of the COVID-19 pandemic. Current Environmental Control Systems (ECS) rely on a prescribed fresh airflow per passenger to dilute contaminants. An adaptive ECS strategy is proposed, leveraging air sensing and filtration technologies to ensure a better CAQ. This paper investigates the CAQ level achieved in commercial aircraft’s cabin during various flight scenarios. The modeling and simulation analysis is performed in a Modelica-based environment describing the dynamic behavior of the system. The model includes the following three main systems: cabin, recirculation loop and air-conditioning pack. The cabin model evaluates the thermo-hygrometric conditions and the air quality in the cabin depending on the number of passengers and crew members, the outdoor conditions and the conditions of the air supplied to the cabin. The recirculation loop includes models of the recirculation fan, ordinary and novel filtration technology, mixing chamber and outflow valve. The air-conditioning pack includes models of heat exchangers and turbomachinery needed to condition the hot pressurized air bled from the engine, as well as selected contaminants originated from the outside or bled from the engine. Different ventilation control strategies are modeled and simulated. Currently, a limited understanding of contaminant concentrations in the cabin and the lack of standardized and systematic methods to collect and record data constitute a challenge in establishing a causal relationship between CAQ and passengers' comfort. As a result, contaminants are neither measured nor filtered during flight, and the current sub-optimal way to avoid their accumulation is their dilution with the fresh air flow. However, the use of a prescribed amount of fresh air comes with a cost, making the ECS the most energy-demanding non-propulsive system within an aircraft. In such a context, this study shows that an ECS based on a reduced and adaptive fresh air flow, and relying on air sensing and filtration technologies, provides promising results in terms of CAQ control. The comparative simulation results demonstrate that the proposed adaptive ECS brings substantial improvements to the CAQ in terms of both controlling the asymptotic values of the concentration of the contaminant and in mitigating hazardous scenarios, such as fume events. Original architectures allowing for adaptive control of the inlet air flow rate based on monitored CAQ will change the requirements for filtration systems and redefine the ECS operation.Keywords: cabin air quality, commercial aircraft, environmental control system, ventilation
Procedia PDF Downloads 102607 Nanocellulose Reinforced Biocomposites Based on Wheat Plasticized Starch for Food Packaging
Authors: Belen Montero, Carmen Ramirez, Maite Rico, Rebeca Bouza, Irene Derungs
Abstract:
Starch is a promising polymer for producing biocomposite materials because it is renewable, completely biodegradable and easily available at a low cost. Thermoplastic starches (TPS) can be obtained after the disruption and plasticization of native starch with a plasticizer. In this work, the solvent casting method was used to obtain TPS films from wheat starch plasticized with glycerol and reinforced with nanocellulose (CNC). X-ray diffraction analysis was used to follow the evolution of the crystallinity. The native wheat starch granules have shown a profile corresponding to A-type crystal structures typical for cereal starches. When TPS films are analyzed a high amorphous halo centered on 19º is obtained, indicating the plasticization process is completed. SEM imaging was made in order to analyse the morphology. The image from the raw wheat starch granules shows a bimodal granule size distribution with some granules in large round disk-shape forms (A-type) and the others as smaller spherical particles (B-type). The image from the neat TPS surface shows a continuous surface. No starch aggregates or swollen granules can be seen so, the plasticization process is complete. In the surfaces of reinforced TPS films aggregates are seen as the CNC concentration in the matrix increases. The CNC influence on the mechanical properties of TPS films has been studied by dynamic mechanical analysis. A direct relation exists between the storage modulus values, E’, and the CNC content in reinforced TPS films: higher is the content of nanocellulose in the composite, higher is the value of E’. This reinforcement effect can be explained by the appearance of a strong and crystalline nanoparticle-TPS interphase. Thermal stability of films was analysed by TGA. It has not observed any influence on the behaviour related to the thermal degradation of films with the incorporation of the CNC. Finally, the resistance to the water absorption films was analysed following the standard UNE-EN ISO 1998:483. The percentage of water absorbed by the samples at each time was calculated. The addition of 5 wt % of CNC to the TPS matrix leads to a significant improvement in the moisture resistance of the starch based material decreasing their diffusivity. It has been associated to the formation of a nanocrystal network that prevents swelling of the starch and therefore water absorption and to the high crystallinity of cellulose compared to starch. As a conclusion, the wheat film reinforced with 5 wt % of cellulose nanocrystals seems to be a good alternative for short-life applications into the packaging industry, because of its greatest rigidity, thermal stability and moisture sorption resistance.Keywords: biocomposites, nanocellulose, starch, wheat
Procedia PDF Downloads 212606 High Performance Computing Enhancement of Agent-Based Economic Models
Authors: Amit Gill, Lalith Wijerathne, Sebastian Poledna
Abstract:
This research presents the details of the implementation of high performance computing (HPC) extension of agent-based economic models (ABEMs) to simulate hundreds of millions of heterogeneous agents. ABEMs offer an alternative approach to study the economy as a dynamic system of interacting heterogeneous agents, and are gaining popularity as an alternative to standard economic models. Over the last decade, ABEMs have been increasingly applied to study various problems related to monetary policy, bank regulations, etc. When it comes to predicting the effects of local economic disruptions, like major disasters, changes in policies, exogenous shocks, etc., on the economy of the country or the region, it is pertinent to study how the disruptions cascade through every single economic entity affecting its decisions and interactions, and eventually affect the economic macro parameters. However, such simulations with hundreds of millions of agents are hindered by the lack of HPC enhanced ABEMs. In order to address this, a scalable Distributed Memory Parallel (DMP) implementation of ABEMs has been developed using message passing interface (MPI). A balanced distribution of computational load among MPI-processes (i.e. CPU cores) of computer clusters while taking all the interactions among agents into account is a major challenge for scalable DMP implementations. Economic agents interact on several random graphs, some of which are centralized (e.g. credit networks, etc.) whereas others are dense with random links (e.g. consumption markets, etc.). The agents are partitioned into mutually-exclusive subsets based on a representative employer-employee interaction graph, while the remaining graphs are made available at a minimum communication cost. To minimize the number of communications among MPI processes, real-life solutions like the introduction of recruitment agencies, sales outlets, local banks, and local branches of government in each MPI-process, are adopted. Efficient communication among MPI-processes is achieved by combining MPI derived data types with the new features of the latest MPI functions. Most of the communications are overlapped with computations, thereby significantly reducing the communication overhead. The current implementation is capable of simulating a small open economy. As an example, a single time step of a 1:1 scale model of Austria (i.e. about 9 million inhabitants and 600,000 businesses) can be simulated in 15 seconds. The implementation is further being enhanced to simulate 1:1 model of Euro-zone (i.e. 322 million agents).Keywords: agent-based economic model, high performance computing, MPI-communication, MPI-process
Procedia PDF Downloads 130605 Environmental and Formal Conditions for the Development of Blue-green Infrastructure (BGI) in the Cities of Central Europe on the Example of Poland
Authors: Magdalena Biela, Marta Weber-Siwirska, Edyta Sierka
Abstract:
The current noticed trend in Central European countries, as in other regions of the world, is for people to migrate to cities. As a result, the urban population is to have reached 70% of the total by 2050. Due to this tendency, as well as taking high real estate prices and limited reserves of city green areas into consideration, the greenery and agricultural soil adjacent to cities is are to be devoted to housing projects, while city centres are expected to undergo partial depopulation. Urban heat islands and phenomena such as torrential rains may cause serious damage. They may even endanger the very life and health of the inhabitants. Due to these tangible effects of climate change, residents expect that local government takes action to develop green infrastructure (GI). The main purpose of our research has been to assess the degree of readiness on the part of the local government in Poland to develop BGI. A questionnaire using the CAWI method was prepared, and a survey was carried out. The target group were town hall employees in all 380 powiat cities and towns (380 county centres) in Poland. The form contained 14 questions covering, among others, actions taken to support the development of GI and ways of motivating residents to take such actions. 224 respondents replied to the questions. The results of the research show that 52% of the cities/towns have taken or intend to take measures to favour the development of green spaces. Currently, the installation of green roofs and living walls is are only carried out by 6 Polish cities, and a few more are at the stage of preparing appropriate regulations. The problem of rainwater retention is much more widespread. Among the municipalities declaring any activities for the benefit of GI, approximately 42% have decided to work on this problem. Over 19% of the respondents are planning an increase in the surface occupied by green areas, 14% - the installation of green roofs, and 12% - redevelopment of city greenery. It is optimistic that 67% of the respondents are willing to acquire knowledge about BGI by means of taking part in educational activities both at the national and international levels. There are many ways to help GI development. The most common type of support in the cities and towns surveyed is co-financing (35%), followed by full financing of projects (11%). About 15% of the cities declare only advisory support. Thus, the problem of GI in Central European cities is at the stage of initial development and requires advanced measures and implementation of both proven solutions applied in other European and world countries using the concept of Nature-based Solutions.Keywords: city/town, blue-green infrastructure, green roofs, climate change adaptation
Procedia PDF Downloads 213604 L2 Anxiety, Proficiency, and L2 Willingness to Communicate in the Classroom, Outside the Classroom, and in Digital Setting: Insights from Ethiopian Preparatory Schools
Authors: Merih Welay Welesilassie, Marianne Nikolov
Abstract:
Research into second and foreign language (L2) acquisitions has demonstrated that L2 anxiety, perceived proficiency, and L2 willingness to communicate (L2WTC) profoundly impact language learning outcomes. However, the complex interplay between these variables has yet to be fully explored, as these factors are dynamic and context-specific and can vary across different learners and learning environments. This study, therefore, utilized a cross-sectional quantitative survey research design to scrutinise the causal relationships between L2 anxiety, English proficiency, and L2WTC of 609 Ethiopian preparatory school students. The model for the L2WTC, both inside and outside the classroom, has been expanded to include an additional sub-scale known as the L2WTC in a digital setting. Moreover, in contrast to the commonly recognised debilitative-focused L2 anxiety, the construct of L2 anxiety has been divided into facilitative and debilitative anxiety. This method allows to measure not only the presence or absence of anxiety but also evaluate if anxiety helps or hinders the L2 learning experience. A self-assessment proficiency measure was also developed specifically for Ethiopian high school students. The study treated facilitative and debilitative anxiety as independent variables while considering self-assessed English proficiency and L2WTC in the classroom, outside the classroom, and in digital settings as dependent variables. Additionally, self-assessed English proficiency was used as an independent variable to predict L2WTC in these three settings. The proposed model, including these variables, was tested using structural equation modelling (SEM). According to the descriptive analysis, the mean scores of L2WTC in the three settings were generally low, ranging from 2.30 to 2.84. Debilitative anxiety casts a shadow on the positive aspects of anxiety. Self-assessed English proficiency was also too low. According to SEM, debilitative anxiety displayed a statistically significant negative impact on L2WTC inside the classroom, outside the classroom, in digital settings, and in self-assessed levels of English proficiency. In contrast, facilitative anxiety was found to positively contribute to L2WTC outside the classroom, in digital settings, and in self-assessed English proficiency. Self-assessed English proficiency made a statistically significant and positive contribution to L2WTC within the classroom, outside the classroom, and in digital contexts. L2WTC inside the classroom was found to positively contribute to L2WTC outside the classrooms and in digital contexts. The findings were systematically compared with existing studies, and the pedagogical implications, limitations, and potential avenues for future research were elucidated. The outcomes of the study have the potential to significantly contribute to the advancement of theoretical and empirical knowledge about improving English education, learning, and communication not only in Ethiopia but also in similar EFL contexts, thereby providing valuable insights for educators, researchers, and policymakers.Keywords: debilitative anxiety, facilitative anxiety, L2 willingness to communicate, self-assessed English proficiency
Procedia PDF Downloads 17603 Modern Contraceptives versus Traditional Contraceptives and Abortion: An Ethnography of Fertiliy Control Practices in Burkina Faso
Authors: Seydou Drabo
Abstract:
This paper examines how traditional contraceptives and abortion practices challenges the use of modern contraceptives in Burkina Faso. It demonstrates how fears and ‘superstitions’ interact with knowledge about modern contraceptives methods to determine use in a context where other way of controlling fertility (traditional contraceptives, abortion) are available to women in the public, private and traditional health sectors. Furthermore, these issues come at the time when Burkina Faso is among countries with a high fertility rate which (6.0 in 2010) and a very low used of contraceptives as only 16% of married women of childbearing age were using a contraceptive method in 2010. The country also has a young population since 33 % of the population is between 10-24 years old and this number is expected to increase by 2050, generating fears that a growing population of youth will put excessive pressure on available resources, including access to education, health services, and employment. Despite over two decades of dedicated policy attention, 24% of women of reproductive age (15-49) was estimated to have an unmet need for contraception in 2010. This paper draws on ethnographic fieldwork conducted since march 2016 (The research is still in progress) in Burkina Faso. Data were collected from 25 women (users and non-users of modern contraceptives and /or traditional contraceptives, post abortion care patients), 4 street drugs vendors and 3 traditional healers through formal and informal interviews, as well as direct observation. The findings show that a variety of contraceptives methods and abortion drugs or methods, both traditional and modern circulate and are available to women. Traditional contraceptives called African contraceptives by some of our participants refer to several birth control method including plants decoction, magical ring, waist necklace, a ritual done with a mixture of lay coming from termite mound and menses. Abortion is a practice that is done in secret through the use of abortion drugs or through intra uterine manoeuvres. Modern contraceptives include Oral contraceptive, implants, injectable. Stereotypes about modern contraceptives, having regular menstrual cycles and adopt of natural birth control methods, bad experience with modern contraceptives methods, the side effect of modern contraceptives, irregularity of sexual activities and the availability of emergency contraceptives are among factors that limit their use among women. In addition, a negative perception is built around modern contraceptives seen as the drug of ‘white people’. In general, the information on these drugs circulates in women’s social network (first line of information on contraceptive). Some women prefer using what they call African contraceptives or inducing an abortion over modern contraceptives because of their side effect. Furthermore, the findings show that women practices and attitudes in controlling birth varies throughout different phases of their lives. Beyond global discourses and technical solution, the issue of Family planning is all about social practices.Keywords: abortion, Burkina Faso, contraception, culture, women
Procedia PDF Downloads 210602 Mapping and Mitigation Strategy for Flash Flood Hazards: A Case Study of Bishoftu City
Authors: Berhanu Keno Terfa
Abstract:
Flash floods are among the most dangerous natural disasters that pose a significant threat to human existence. They occur frequently and can cause extensive damage to homes, infrastructure, and ecosystems while also claiming lives. Although flash floods can happen anywhere in the world, their impact is particularly severe in developing countries due to limited financial resources, inadequate drainage systems, substandard housing options, lack of early warning systems, and insufficient preparedness. To address these challenges, a comprehensive study has been undertaken to analyze and map flood inundation using Geographic Information System (GIS) techniques by considering various factors that contribute to flash flood resilience and developing effective mitigation strategies. Key factors considered in the analysis include slope, drainage density, elevation, Curve Number, rainfall patterns, land-use/cover classes, and soil data. These variables were computed using ArcGIS software platforms, and data from the Sentinel-2 satellite image (with a 10-meter resolution) were utilized for land-use/cover classification. Additionally, slope, elevation, and drainage density data were generated from the 12.5-meter resolution of the ALOS Palsar DEM, while other relevant data were obtained from the Ethiopian Meteorological Institute. By integrating and regularizing the collected data through GIS and employing the analytic hierarchy process (AHP) technique, the study successfully delineated flash flood hazard zones (FFHs) and generated a suitable land map for urban agriculture. The FFH model identified four levels of risk in Bishoftu City: very high (2106.4 ha), high (10464.4 ha), moderate (1444.44 ha), and low (0.52 ha), accounting for 15.02%, 74.7%, 10.1%, and 0.004% of the total area, respectively. The results underscore the vulnerability of many residential areas in Bishoftu City, particularly the central areas that have been previously developed. Accurate spatial representation of flood-prone areas and potential agricultural zones is crucial for designing effective flood mitigation and agricultural production plans. The findings of this study emphasize the importance of flood risk mapping in raising public awareness, demonstrating vulnerability, strengthening financial resilience, protecting the environment, and informing policy decisions. Given the susceptibility of Bishoftu City to flash floods, it is recommended that the municipality prioritize urban agriculture adaptation, proper settlement planning, and drainage network design.Keywords: remote sensing, flush flood hazards, Bishoftu, GIS.
Procedia PDF Downloads 38601 The Potential of On-Demand Shuttle Services to Reduce Private Car Use
Authors: B. Mack, K. Tampe-Mai, E. Diesch
Abstract:
Findings of an ongoing discrete choice study of future transport mode choice will be presented. Many urban centers face the triple challenge of having to cope with ever increasing traffic congestion, environmental pollution, and greenhouse gas emission brought about by private car use. In principle, private car use may be diminished by extending public transport systems like bus lines, trams, tubes, and trains. However, there are limits to increasing the (perceived) spatial and temporal flexibility and reducing peak-time crowding of classical public transport systems. An emerging new type of system, publicly or privately operated on-demand shuttle bus services, seem suitable to ameliorate the situation. A fleet of on-demand shuttle busses operates without fixed stops and schedules. It may be deployed efficiently in that each bus picks up passengers whose itineraries may be combined into an optimized route. Crowding may be minimized by limiting the number of seats and the inter-seat distance for each bus. The study is conducted as a discrete choice experiment. The choice between private car, public transport, and shuttle service is registered as a function of several push and pull factors (financial costs, travel time, walking distances, mobility tax/congestion charge, and waiting time/parking space search time). After the completion of the discrete choice items, the study participant is asked to rate the three modes of transport with regard to the pull factors of comfort, safety, privacy, and opportunity to engage in activities like reading or surfing the internet. These ratings are entered as additional predictors into the discrete choice experiment regression model. The study is conducted in the region of Stuttgart in southern Germany. N=1000 participants are being recruited. Participants are between 18 and 69 years of age, hold a driver’s license, and live in the city or the surrounding region of Stuttgart. In the discrete choice experiment, participants are asked to assume they lived within the Stuttgart region, but outside of the city, and were planning the journey from their apartment to their place of work, training, or education during the peak traffic time in the morning. Then, for each item of the discrete choice experiment, they are asked to choose between the transport modes of private car, public transport, and on-demand shuttle in the light of particular values of the push and pull factors studied. The study will provide valuable information on the potential of switching from private car use to the use of on-demand shuttles, but also on the less desirable potential of switching from public transport to on-demand shuttle services. Furthermore, information will be provided on the modulation of these switching potentials by pull and push factors.Keywords: determinants of travel mode choice, on-demand shuttle services, private car use, public transport
Procedia PDF Downloads 184600 Cross-Comparison between Land Surface Temperature from Polar and Geostationary Satellite over Heterogenous Landscape: A Case Study in Hong Kong
Authors: Ibrahim A. Adeniran, Rui F. Zhu, Man S. Wong
Abstract:
Owing to the insufficiency in the spatial representativeness and continuity of in situ temperature measurements from weather stations (WS), the use of temperature measurement from WS for large-range diurnal analysis in heterogenous landscapes has been limited. This has made the accurate estimation of land surface temperature (LST) from remotely sensed data more crucial. Moreover, the study of dynamic interaction between the atmosphere and the physical surface of the Earth could be enhanced at both annual and diurnal scales by using optimal LST data derived from satellite sensors. The tradeoff between the spatial and temporal resolution of LSTs from satellite’s thermal infrared sensors (TIRS) has, however, been a major challenge, especially when high spatiotemporal LST data are recommended. It is well-known from existing literature that polar satellites have the advantage of high spatial resolution, while geostationary satellites have a high temporal resolution. Hence, this study is aimed at designing a framework for the cross-comparison of LST data from polar and geostationary satellites in a heterogeneous landscape. This could help to understand the relationship between the LST estimates from the two satellites and, consequently, their integration in diurnal LST analysis. Landsat-8 satellite data will be used as the representative of the polar satellite due to the availability of its long-term series, while the Himawari-8 satellite will be used as the data source for the geostationary satellite because of its improved TIRS. For the study area, Hong Kong Special Administrative Region (HK SAR) will be selected; this is due to the heterogeneity in the landscape of the region. LST data will be retrieved from both satellites using the Split window algorithm (SWA), and the resulting data will be validated by comparing satellite-derived LST data with temperature data from automatic WS in HK SAR. The LST data from the satellite data will then be separated based on the land use classification in HK SAR using the Global Land Cover by National Mapping Organization version3 (GLCNMO 2013) data. The relationship between LST data from Landsat-8 and Himawari-8 will then be investigated based on the land-use class and over different seasons of the year in order to account for seasonal variation in their relationship. The resulting relationship will be spatially and statistically analyzed and graphically visualized for detailed interpretation. Findings from this study will reveal the relationship between the two satellite data based on the land use classification within the study area and the seasons of the year. While the information provided by this study will help in the optimal combination of LST data from Polar (Landsat-8) and geostationary (Himawari-8) satellites, it will also serve as a roadmap in the annual and diurnal urban heat (UHI) analysis in Hong Kong SAR.Keywords: automatic weather station, Himawari-8, Landsat-8, land surface temperature, land use classification, split window algorithm, urban heat island
Procedia PDF Downloads 75599 Controlled Drug Delivery System for Delivery of Poor Water Soluble Drugs
Authors: Raj Kumar, Prem Felix Siril
Abstract:
The poor aqueous solubility of many pharmaceutical drugs and potential drug candidates is a big challenge in drug development. Nanoformulation of such candidates is one of the major solutions for the delivery of such drugs. We initially developed the evaporation assisted solvent-antisolvent interaction (EASAI) method. EASAI method is use full to prepared nanoparticles of poor water soluble drugs with spherical morphology and particles size below 100 nm. However, to further improve the effect formulation to reduce number of dose and side effect it is important to control the delivery of drugs. However, many drug delivery systems are available. Among the many nano-drug carrier systems, solid lipid nanoparticles (SLNs) have many advantages over the others such as high biocompatibility, stability, non-toxicity and ability to achieve controlled release of drugs and drug targeting. SLNs can be administered through all existing routes due to high biocompatibility of lipids. SLNs are usually composed of lipid, surfactant and drug were encapsulated in lipid matrix. A number of non-steroidal anti-inflammatory drugs (NSAIDs) have poor bioavailability resulting from their poor aqueous solubility. In the present work, SLNs loaded with NSAIDs such as Nabumetone (NBT), Ketoprofen (KP) and Ibuprofen (IBP) were successfully prepared using different lipids and surfactants. We studied and optimized experimental parameters using a number of lipids, surfactants and NSAIDs. The effect of different experimental parameters such as lipid to surfactant ratio, volume of water, temperature, drug concentration and sonication time on the particles size of SLNs during the preparation using hot-melt sonication was studied. It was found that particles size was directly proportional to drug concentration and inversely proportional to surfactant concentration, volume of water added and temperature of water. SLNs prepared at optimized condition were characterized thoroughly by using different techniques such as dynamic light scattering (DLS), field emission scanning electron microscopy (FESEM), transmission electron microscopy (TEM), atomic force microscopy (AFM), X-ray diffraction (XRD) and differential scanning calorimetry and Fourier transform infrared spectroscopy (FTIR). We successfully prepared the SLN of below 220 nm using different lipids and surfactants combination. The drugs KP, NBT and IBP showed 74%, 69% and 53% percentage of entrapment efficiency with drug loading of 2%, 7% and 6% respectively in SLNs of Campul GMS 50K and Gelucire 50/13. In-vitro drug release profile of drug loaded SLNs is shown that nearly 100% of drug was release in 6 h.Keywords: nanoparticles, delivery, solid lipid nanoparticles, hot-melt sonication, poor water soluble drugs, solubility, bioavailability
Procedia PDF Downloads 312598 The Role of Social Media in the Rise of Islamic State in India: An Analytical Overview
Authors: Yasmeen Cheema, Parvinder Singh
Abstract:
The evolution of Islamic State (acronym IS) has an ultimate goal of restoring the caliphate. IS threat to the global security is main concern of international community but has also raised a factual concern for India about the regular radicalization of IS ideology among Indian youth. The incident of joining Arif Ejaz Majeed, an Indian as ‘jihadist’ in IS has set strident alarm in law & enforcement agencies. On 07.03.2017, many people were injured in an Improvised Explosive Device (IED) blast on-board of Bhopal Ujjain Express. One perpetrator of this incident was killed in encounter with police. But, the biggest shock is that the conspiracy was pre-planned and the assailants who carried out the blast were influenced by the ideology perpetrated by the Islamic State. This is the first time name of IS has cropped up in a terror attack in India. It is a red indicator of violent presence of IS in India, which is spreading through social media. The IS have the capacity to influence the younger Muslim generation in India through its brutal and aggressive propaganda videos, social media apps and hatred speeches. It is a well known fact that India is on the radar of IS, as well on its ‘Caliphate Map’. IS uses Twitter, Facebook and other social media platforms constantly. Islamic State has used enticing videos, graphics, and articles on social media and try to influence persons from India & globally that their jihad is worthy. According to arrested perpetrator of IS in different cases in India, the most of Indian youths are victims to the daydreams which are fondly shown by IS. The dreams that the Muslim empire as it was before 1920 can come back with all its power and also that the Caliph and its caliphate can be re-established are shown by the IS. Indian Muslim Youth gets attracted towards these euphemistic ideologies. Islamic State has used social media for disseminating its poisonous ideology, recruitment, operational activities and for future direction of attacks. IS through social media inspired its recruits & lone wolfs to continue to rely on local networks to identify targets and access weaponry and explosives. Recently, a pro-IS media group on its Telegram platform shows Taj Mahal as the target and suggested mode of attack as a Vehicle Born Improvised Explosive Attack (VBIED). Islamic State definitely has the potential to destroy the Indian national security & peace, if timely steps are not taken. No doubt, IS has used social media as a critical mechanism for recruitment, planning and executing of terror attacks. This paper will therefore examine the specific characteristics of social media that have made it such a successful weapon for Islamic State. The rise of IS in India should be viewed as a national crisis and handled at the central level with efficient use of modern technology.Keywords: ideology, India, Islamic State, national security, recruitment, social media, terror attack
Procedia PDF Downloads 231597 The Impact of the Virtual Learning Environment on Teacher's Pedagogy and Student's Learning in Primary School Setting
Authors: Noor Ashikin Omar
Abstract:
The rapid growth and advancement in information and communication technology (ICT) at a global scene has greatly influenced and revolutionised interaction amongst society. The use of ICT has become second nature in managing everyday lives, particularly in the education environment. Traditional learning methods of using blackboards and chalks have been largely improved by the use of ICT devices such as interactive whiteboards and computers in school. This paper aims to explore the impacts of virtual learning environments (VLE) on teacher’s pedagogy and student’s learning in primary school settings. The research was conducted in two phases. Phase one of this study comprised a short interview with the school’s senior assistants to examine issues and challenges faced during planning and implementation of FrogVLE in their respective schools. Phase two involved a survey of a number of questionnaires directed to three major stakeholders; the teachers, students and parents. The survey intended to explore teacher’s and student’s perspective and attitude towards the use of VLE as a teaching and learning medium and as a learning experience as a whole. In addition, the survey from parents provided insights on how they feel towards the use of VLE for their child’s learning. Collectively, the two phases enable improved understanding and provided observations on factors that had affected the implementation of the VLE into primary schools. This study offers the voices of the students which were frequently omitted when addressing innovations as well as teachers who may not always be heard. It is also significant in addressing the importance of teacher’s pedagogy on students’ learning and its effects to enable more effective ICT integration with a student-centred approach. Finally, parental perceptions in the implementation of VLE in supporting their children’s learning have been implicated as having a bearing on educational achievement. The results indicate that the all three stakeholders were positive and highly supportive towards the use of VLE in schools. They were able to understand the benefits of moving towards the modern method of teaching using ICT and accept the change in the education system. However, factors such as condition of ICT facilities at schools and homes as well as inadequate professional development for the teachers in both ICT skills and management skills hindered exploitation of the VLE system in order to fully utilise its benefits. Social influences within different communities and cultures and costs of using the technology also has a significant impact. The findings of this study are important to the Malaysian Ministry of Education because it informs policy makers on the impact of the Virtual Learning Environment (VLE) on teacher’s pedagogy and learning of Malaysian primary school children. The information provided to policy makers allows them to make a sound judgement and enables an informed decision making.Keywords: attitudes towards virtual learning environment (VLE), parental perception, student's learning, teacher's pedagogy
Procedia PDF Downloads 206596 Attachment Theory and Quality of Life: Grief Education and Training
Authors: Jane E. Hill
Abstract:
Quality of life is an important component for many. With that in mind, everyone will experience some type of loss within his or her lifetime. A person can experience loss due to break up, separation, divorce, estrangement, or death. An individual may experience loss of a job, loss of capacity, or loss caused by human or natural-caused disasters. An individual’s response to such a loss is unique to them, and not everyone will seek services to assist them with their grief due to loss. Counseling can promote positive outcomes for clients that are grieving by addressing the client’s personal loss and helping the client process their grief. However, a lack of understanding on the part of counselors of how people grieve may result in negative client outcomes such as poor health, psychological distress, or an increased risk of depression. Education and training in grief counseling can improve counselors’ problem recognition and skills in treatment planning. The purpose of this study was to examine whether the Council for Accreditation of Counseling and Related Educational Programs (CACREP) master’s degree counseling students view themselves as having been adequately trained in grief theories and skills. Many people deal with grief issues that prevent them from having joy or purpose in their lives and that leaves them unable to engage in positive opportunities or relationships. This study examined CACREP-accredited master’s counseling students’ self-reported competency, training, and education in providing grief counseling. The implications for positive social change arising from the research may be to incorporate and promote education and training in grief theories and skills in a majority of counseling programs and to provide motivation to incorporate professional standards for grief training and practice in the mental health counseling field. The theoretical foundation used was modern grief theory based on John Bowlby’s work on Attachment Theory. The overall research question was how competent do master’s-level counselors view themselves regarding the education or training they received in grief theories or counseling skills in their CACREP-accredited studies. The author used a non-experimental, one shot survey comparative quantitative research design. Cicchetti’s Grief Counseling Competency Scale (GCCS) was administered to CACREP master’s-level counseling students enrolled in their practicum or internship experience, which resulted in 153 participants. Using a MANCOVA, there was significance found for relationships between coursework taken and (a) perceived assessment skills (p = .029), (b) perceived treatment skills (p = .025), and (c) perceived conceptual skills and knowledge (p = .003). Results of this study provided insight for CACREP master’s-level counseling programs to explore and discuss curriculum coursework inclusion of education and training in grief theories and skills.Keywords: counselor education and training, grief education and training, grief and loss, quality of life
Procedia PDF Downloads 192595 Gender Policies and Political Culture: An Examination of the Canadian Context
Authors: Chantal Maille
Abstract:
This paper is about gender-based analysis plus (GBA+), an intersectional gender policy used in Canada to assess the impact of policies and programs for men and women from different origins. It looks at Canada’s political culture to explain the nature of its gender policies. GBA+ is defined as an analysis method that makes it possible to assess the eventual effects of policies, programs, services, and other initiatives on women and men of different backgrounds because it takes account of gender and other identity factors. The ‘plus’ in the name serves to emphasize that GBA+ goes beyond gender to include an examination of a wide range of other related identity factors, such as age, education, language, geography, culture, and income. The point of departure for GBA+ is that women and men are not homogeneous populations and gender is never the only factor in defining a person’s identity; rather, it interacts with factors such as ethnic origin, age, disabilities, where the person lives, and other aspects of individual and social identity. GBA+ takes account of these factors and thus challenges notions of similarity or homogeneity within populations of women and men. Comparative analysis based on sex and gender may serve as a gateway to studying a given question, but women, men, girls, and boys do not form homogeneous populations. In the 1990s, intersectionality emerged as a new feminist framework. The popularity of the notion of intersectionality corresponds to a time when, in hindsight, the damage done to minoritized groups by state disengagement policies in concert with global intensification of neoliberalism, and vice versa, can be measured. Although GBA+ constitutes a form of intersectionalization of GBA, it must be understood that the two frameworks do not spring from a similar logic. Intersectionality first emerged as a dynamic analysis of differences between women that was oriented toward change and social justice, whereas GBA is a technique developed by state feminists in a context of analyzing governmental policies and aiming to promote equality between men and women. It can nevertheless be assumed that there might be interest in such a policy and program analysis grid that is decentred from gender and offers enough flexibility to take account of a group of inequalities. In terms of methodology, the research is supported by a qualitative analysis of governmental documents about GBA+ in Canada. Research findings identify links between Canadian gender policies and its political culture. In Canada, diversity has been taken into account as an element at the basis of gendered analysis of public policies since 1995. The GBA+ adopted by the government of Canada conveys an opening to intersectionality and a sensitivity to multiculturalism. The Canadian Multiculturalism Act, adopted 1988, proposes to recognize the fact that multiculturalism is a fundamental characteristic of the Canadian identity and heritage and constitutes an invaluable resource for the future of the country. In conclusion, Canada’s distinct political culture can be associated with the specific nature of its gender policies.Keywords: Canada, gender-based analysis, gender policies, political culture
Procedia PDF Downloads 224594 A Study of Bilingual Development of a Mandarin and English Bilingual Preschool Child from China to Australia
Abstract:
This project aims to trace the developmental patterns of a child's Mandarin and English from China to Australia from age 3; 03 till 5; 06. In childhood bilingual studies, there is an assumption that age 3 is the dividing line between simultaneous bilinguals and sequential bilinguals. Determining similarities and differences between Bilingual First Language Acquisition, Early Second Language Acquisition, and Second Language Acquisition is of great theoretical significance. Studies on Bilingual First Language Acquisition, hereafter, BFLA in the past three decades have shown that the grammatical development of bilingual children progresses through the same developmental trajectories as their monolingual counterparts. Cross-linguistic interaction does not show changes of the basic grammatical knowledge, even in the weaker language. While BFLA studies show consistent results under the conditions of adequate input and meaningful interactional context, the research findings of Early Second Language Acquisition (ESLA) have demonstrated that this cohort proceeds their early English differently from both BFLA and SLA. The different development could be attributed to the age of migration, input pattern, and their Environmental Languages (Lε). In the meantime, the dynamic relationship between the two languages is an issue to invite further attention. The present study attempts to fill this gap. The child in this case study started acquiring L1 Mandarin from birth in China, where the environmental language (Lε) coincided with L1 Mandarin. When she migrated to Australia at 3;06, where the environmental language (Lε) was L2 English, her Mandarin exposure was reduced. On the other hand, she received limited English input starting from 1; 02 in China, where the environmental language (Lε) was L1 Mandarin, a non-English environment. When she relocated to Australia at 3; 06, where the environmental language (Lε) coincided with L2 English, her English exposure significantly increased. The child’s linguistic profile provides an opportunity to explore: (1) What does the child’s English developmental route look like? (2) What does the L1 Mandarin developmental pattern look like in different environmental languages? (3) How do input and environmental language interact in shaping the bilingual child’s linguistic repertoire? In order to answer these questions, two linguistic areas are selected as the focus of the investigation, namely, subject realization and wh-questions. The chosen areas are contrastive in structure but perform the same semantic functions in the two linguistically distant languages and can serve as an ideal testing ground for exploring the developmental path in the two languages. The longitudinal case study adopts a combined approach of qualitative and quantitative analysis. Two years’ Mandarin and English data are examined, and comparisons are made with age-matched monolinguals in each language in CHILDES. To the author’s best knowledge, this study is the first of this kind examining a Mandarin-English bilingual child's bilingual development at a critical age, in different input patterns, and in different environmental languages (Lε). It also expands the scope of the theory of Lε, adding empirical evidence on the relationship between input and Lε in bilingual acquisition.Keywords: bilingual development, age, input, environmental language (Le)
Procedia PDF Downloads 154593 Boussinesq Model for Dam-Break Flow Analysis
Authors: Najibullah M, Soumendra Nath Kuiry
Abstract:
Dams and reservoirs are perceived for their estimable alms to irrigation, water supply, flood control, electricity generation, etc. which civilize the prosperity and wealth of society across the world. Meantime the dam breach could cause devastating flood that can threat to the human lives and properties. Failures of large dams remain fortunately very seldom events. Nevertheless, a number of occurrences have been recorded in the world, corresponding in an average to one to two failures worldwide every year. Some of those accidents have caused catastrophic consequences. So it is decisive to predict the dam break flow for emergency planning and preparedness, as it poses high risk to life and property. To mitigate the adverse impact of dam break, modeling is necessary to gain a good understanding of the temporal and spatial evolution of the dam-break floods. This study will mainly deal with one-dimensional (1D) dam break modeling. Less commonly used in the hydraulic research community, another possible option for modeling the rapidly varied dam-break flows is the extended Boussinesq equations (BEs), which can describe the dynamics of short waves with a reasonable accuracy. Unlike the Shallow Water Equations (SWEs), the BEs taken into account the wave dispersion and non-hydrostatic pressure distribution. To capture the dam-break oscillations accurately it is very much needed of at least fourth-order accurate numerical scheme to discretize the third-order dispersion terms present in the extended BEs. The scope of this work is therefore to develop an 1D fourth-order accurate in both space and time Boussinesq model for dam-break flow analysis by using finite-volume / finite difference scheme. The spatial discretization of the flux and dispersion terms achieved through a combination of finite-volume and finite difference approximations. The flux term, was solved using a finite-volume discretization whereas the bed source and dispersion term, were discretized using centered finite-difference scheme. Time integration achieved in two stages, namely the third-order Adams Basforth predictor stage and the fourth-order Adams Moulton corrector stage. Implementation of the 1D Boussinesq model done using PYTHON 2.7.5. Evaluation of the performance of the developed model predicted as compared with the volume of fluid (VOF) based commercial model ANSYS-CFX. The developed model is used to analyze the risk of cascading dam failures similar to the Panshet dam failure in 1961 that took place in Pune, India. Nevertheless, this model can be used to predict wave overtopping accurately compared to shallow water models for designing coastal protection structures.Keywords: Boussinesq equation, Coastal protection, Dam-break flow, One-dimensional model
Procedia PDF Downloads 232