Search results for: dense masonry unit
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 2733

Search results for: dense masonry unit

1083 The Transient Reactive Power Regulation Capability of SVC for Large Scale WECS Connected to Distribution Networks

Authors: Y. Ates, A. R. Boynuegri, M. Uzunoglu, A. Karakas

Abstract:

The recent interest in alternative and renewable energy systems results in increased installed capacity ratio of such systems in total energy production of the world. Specifically, wind energy conversion systems (WECS) draw significant attention among possible alternative energy options, recently. On the contrary of the positive points of penetrating WECS in all over the world in terms of environment protection, energy independence of the countries, etc., there are significant problems to be solved for the grid connection of large scale WECS. The reactive power regulation, voltage variation suppression, etc. can be presented as major issues to be considered in this regard. Thus, this paper evaluates the application of a Static VAr Compensator (SVC) unit for the reactive power regulation and operation continuity of WECS during a fault condition. The system is modeled employing the IEEE 13 node test system. Thus, it is possible to evaluate the system performance with an overall grid simulation model close to real grid systems. The overall simulation model is developed in MATLAB/Simulink/SimPowerSystems® environments and the obtained results effectively match the target of the provided study.

Keywords: IEEE 13 bus distribution system, reactive power regulation, static VAr compensator, wind energy conversion system

Procedia PDF Downloads 722
1082 Design and Fabrication of a Programmable Stiffness-Sensitive Gripper for Object Handling

Authors: Mehdi Modabberifar, Sanaz Jabary, Mojtaba Ghodsi

Abstract:

Stiffness sensing is an important issue in medical diagnostic, robotics surgery, safe handling, and safe grasping of objects in production lines. Detecting and obtaining the characteristics in dwelling lumps embedded in a soft tissue and safe removing and handling of detected lumps is needed in surgery. Also in industry, grasping and handling an object without damaging in a place where it is not possible to access a human operator is very important. In this paper, a method for object handling is presented. It is based on the use of an intelligent gripper to detect the object stiffness and then setting a programmable force for grasping the object to move it. The main components of this system includes sensors (sensors for measuring force and displacement), electrical (electrical and electronic circuits, tactile data processing and force control system), mechanical (gripper mechanism and driving system for the gripper) and the display unit. The system uses a rotary potentiometer for measuring gripper displacement. A microcontroller using the feedback received by the load cell, mounted on the finger of the gripper, calculates the amount of stiffness, and then commands the gripper motor to apply a certain force on the object. Results of Experiments on some samples with different stiffness show that the gripper works successfully. The gripper can be used in haptic interfaces or robotic systems used for object handling.

Keywords: gripper, haptic, stiffness, robotic

Procedia PDF Downloads 342
1081 Heat Vulnerability Index (HVI) Mapping in Extreme Heat Days Coupled with Air Pollution Using Principal Component Analysis (PCA) Technique: A Case Study of Amiens, France

Authors: Aiman Mazhar Qureshi, Ahmed Rachid

Abstract:

Extreme heat events are emerging human environmental health concerns in dense urban areas due to anthropogenic activities. High spatial and temporal resolution heat maps are important for urban heat adaptation and mitigation, helping to indicate hotspots that are required for the attention of city planners. The Heat Vulnerability Index (HVI) is the important approach used by decision-makers and urban planners to identify heat-vulnerable communities and areas that require heat stress mitigation strategies. Amiens is a medium-sized French city, where the average temperature has been increasing since the year 2000 by +1°C. Extreme heat events are recorded in the month of July for the last three consecutive years, 2018, 2019 and 2020. Poor air quality, especially ground-level ozone, has been observed mainly during the same hot period. In this study, we evaluated the HVI in Amiens during extreme heat days recorded last three years (2018,2019,2020). The Principal Component Analysis (PCA) technique is used for fine-scale vulnerability mapping. The main data we considered for this study to develop the HVI model are (a) socio-economic and demographic data; (b) Air pollution; (c) Land use and cover; (d) Elderly heat-illness; (e) socially vulnerable; (f) Remote sensing data (Land surface temperature (LST), mean elevation, NDVI and NDWI). The output maps identified the hot zones through comprehensive GIS analysis. The resultant map shows that high HVI exists in three typical areas: (1) where the population density is quite high and the vegetation cover is small (2) the artificial surfaces (built-in areas) (3) industrial zones that release thermal energy and ground-level ozone while those with low HVI are located in natural landscapes such as rivers and grasslands. The study also illustrates the system theory with a causal diagram after data analysis where anthropogenic activities and air pollution appear in correspondence with extreme heat events in the city. Our suggested index can be a useful tool to guide urban planners and municipalities, decision-makers and public health professionals in targeting areas at high risk of extreme heat and air pollution for future interventions adaptation and mitigation measures.

Keywords: heat vulnerability index, heat mapping, heat health-illness, remote sensing, urban heat mitigation

Procedia PDF Downloads 128
1080 Haemocompatibility of Surface Modified AISI 316L Austenitic Stainless Steel Tested in Artificial Plasma

Authors: W. Walke, J. Przondziono, K. Nowińska

Abstract:

The study comprises evaluation of suitability of passive layer created on the surface of AISI 316L stainless steel for products that are intended to have contact with blood. For that purpose, prior to and after chemical passivation, samples were subject to 7 day exposure in artificial plasma at the temperature of T=37°C. Next, tests of metallic ions infiltration from the surface to the solution were performed. The tests were performed with application of spectrometer JY 2000, by Yobin – Yvon, employing Inductively Coupled Plasma Atomic Emission Spectrometry (ICP-AES). In order to characterize physical and chemical features of electrochemical processes taking place during exposure of samples to artificial plasma, tests with application of electrochemical impedance spectroscopy were suggested. The tests were performed with application of measuring unit equipped with potentiostat PGSTAT 302n with an attachment for impedance tests FRA2. Measurements were made in the environment simulating human blood at the temperature of T=37°C. Performed tests proved that application of chemical passivation process for AISI 316L stainless steel used for production of goods intended to have contact with blood is well-grounded and useful in order to improve safety of their usage.

Keywords: AISI 316L stainless steel, chemical passivation, artificial plasma, ions infiltration, EIS

Procedia PDF Downloads 250
1079 A Quantum Leap: Developing Quantum Semi-Structured Complex Numbers to Solve the “Division by Zero” Problem

Authors: Peter Jean-Paul, Shanaz Wahid

Abstract:

The problem of division by zero can be stated as: “what is the value of 0 x 1/0?” This expression has been considered undefined by mathematicians because it can have two equally valid solutions either 0 or 1. Recently semi-structured complex number set was invented to solve “division by zero”. However, whilst the number set had some merits it was considered to have a poor theoretical foundation and did not provide a quality solution to “division by zero”. Moreover, the set lacked consistency in simple algebraic calculations producing contradictory results when dividing by zero. To overcome these issues this research starts by treating the expression " 0 x 1/0" as a quantum mechanical system that produces two tangled results 0 and 1. Dirac Notation (a tool from quantum mechanics) was then used to redefine the unstructured unit p in semi-structured complex numbers so that p represents the superposition of two results (0 and 1) and collapses into a single value when used in algebraic expressions. In the process, this paper describes a new number set called Quantum Semi-structured Complex Numbers that provides a valid solution to the problem of “division by zero”. This research shows that this new set (1) forms a “Field”, (2) can produce consistent results when solving division by zero problems, (3) can be used to accurately describe systems whose mathematical descriptions involve division by zero. This research served to provide a firm foundation for Quantum Semi-structured Complex Numbers and support their practical use.

Keywords: division by zero, semi-structured complex numbers, quantum mechanics, Hilbert space, Euclidean space

Procedia PDF Downloads 143
1078 Integrating Cooperative Education Experience into Engineering Curriculum: An Approach

Authors: Robin Lok-Wang Ma

Abstract:

The Center/Unit for Industry Engagement and Collaboration, as well as Internship, play a significant role at university. In general, the Center serves as the official interface between the industry and the School or Department to cultivate students’ early exposure to professional experience. The missions of the Center are not limited to provide a communication channel and collaborative platform for the industries and the university but also to assist students to build up their career paths early while still in the university. In recent years, a cooperative education experience (commonly known as a co-op) has been strongly advocated for students to make the school-to-work transition. The nature of the co-op program is not only consistent with the internships/final year design projects, but it is also more industrial-oriented with academic support from faculty at the university. The purpose of this paper is to describe an approach to how cooperative education experience can be integrated into Engineering Curriculum. It provides a mutual understanding and exchange of ideas for the approach between the university and the industry. A suggested format in terms of timeline, duration, selection of candidates, students, and companies’ expectations for the co-op program is described. Also, feedbacks from employers/industries show that a longer-term co-op program is well suited for students compared with a short-term internship. To this end, it provides a new insight into collaboration and/or partnership between the university and the industries to prepare professional work-ready graduates.

Keywords: cooperative education, industry, engagement, collaboration

Procedia PDF Downloads 84
1077 High Performance Computing Enhancement of Agent-Based Economic Models

Authors: Amit Gill, Lalith Wijerathne, Sebastian Poledna

Abstract:

This research presents the details of the implementation of high performance computing (HPC) extension of agent-based economic models (ABEMs) to simulate hundreds of millions of heterogeneous agents. ABEMs offer an alternative approach to study the economy as a dynamic system of interacting heterogeneous agents, and are gaining popularity as an alternative to standard economic models. Over the last decade, ABEMs have been increasingly applied to study various problems related to monetary policy, bank regulations, etc. When it comes to predicting the effects of local economic disruptions, like major disasters, changes in policies, exogenous shocks, etc., on the economy of the country or the region, it is pertinent to study how the disruptions cascade through every single economic entity affecting its decisions and interactions, and eventually affect the economic macro parameters. However, such simulations with hundreds of millions of agents are hindered by the lack of HPC enhanced ABEMs. In order to address this, a scalable Distributed Memory Parallel (DMP) implementation of ABEMs has been developed using message passing interface (MPI). A balanced distribution of computational load among MPI-processes (i.e. CPU cores) of computer clusters while taking all the interactions among agents into account is a major challenge for scalable DMP implementations. Economic agents interact on several random graphs, some of which are centralized (e.g. credit networks, etc.) whereas others are dense with random links (e.g. consumption markets, etc.). The agents are partitioned into mutually-exclusive subsets based on a representative employer-employee interaction graph, while the remaining graphs are made available at a minimum communication cost. To minimize the number of communications among MPI processes, real-life solutions like the introduction of recruitment agencies, sales outlets, local banks, and local branches of government in each MPI-process, are adopted. Efficient communication among MPI-processes is achieved by combining MPI derived data types with the new features of the latest MPI functions. Most of the communications are overlapped with computations, thereby significantly reducing the communication overhead. The current implementation is capable of simulating a small open economy. As an example, a single time step of a 1:1 scale model of Austria (i.e. about 9 million inhabitants and 600,000 businesses) can be simulated in 15 seconds. The implementation is further being enhanced to simulate 1:1 model of Euro-zone (i.e. 322 million agents).

Keywords: agent-based economic model, high performance computing, MPI-communication, MPI-process

Procedia PDF Downloads 113
1076 A Comparison of Tsunami Impact to Sydney Harbour, Australia at Different Tidal Stages

Authors: Olivia A. Wilson, Hannah E. Power, Murray Kendall

Abstract:

Sydney Harbour is an iconic location with a dense population and low-lying development. On the east coast of Australia, facing the Pacific Ocean, it is exposed to several tsunamigenic trenches. This paper presents a component of the most detailed assessment of the potential for earthquake-generated tsunami impact on Sydney Harbour to date. Models in this study use dynamic tides to account for tide-tsunami interaction. Sydney Harbour’s tidal range is 1.5 m, and the spring tides from January 2015 that are used in the modelling for this study are close to the full tidal range. The tsunami wave trains modelled include hypothetical tsunami generated from earthquakes of magnitude 7.5, 8.0, 8.5, and 9.0 MW from the Puysegur and New Hebrides trenches as well as representations of the historical 1960 Chilean and 2011 Tohoku events. All wave trains are modelled for the peak wave to coincide with both a low tide and a high tide. A single wave train, representing a 9.0 MW earthquake at the Puysegur trench, is modelled for peak waves to coincide with every hour across a 12-hour tidal phase. Using the hydrodynamic model ANUGA, results are compared according to the impact parameters of inundation area, depth variation and current speeds. Results show that both maximum inundation area and depth variation are tide dependent. Maximum inundation area increases when coincident with a higher tide, however, hazardous inundation is only observed for the larger waves modelled: NH90high and P90high. The maximum and minimum depths are deeper on higher tides and shallower on lower tides. The difference between maximum and minimum depths varies across different tidal phases although the differences are slight. Maximum current speeds are shown to be a significant hazard for Sydney Harbour; however, they do not show consistent patterns according to tide-tsunami phasing. The maximum current speed hazard is shown to be greater in specific locations such as Spit Bridge, a narrow channel with extensive marine infrastructure. The results presented for Sydney Harbour are novel, and the conclusions are consistent with previous modelling efforts in the greater area. It is shown that tide must be a consideration for both tsunami modelling and emergency management planning. Modelling with peak tsunami waves coinciding with a high tide would be a conservative approach; however, it must be considered that maximum current speeds may be higher on other tides.

Keywords: emergency management, sydney, tide-tsunami interaction, tsunami impact

Procedia PDF Downloads 227
1075 Challenges of Teaching Physical Education to Students With Special Needs in Regular School Settings

Authors: Christine Okello

Abstract:

Physical Education (PE) curriculum provides school age students to explore issues that are likely to impact on health, safety, and well-being. The current curriculum includes the physical activity component, intended to improve physical fitness, social skills as well as building confidence. While this viewpoint is vital, there are challenges and stigma attached when specific issues are either ignored, inadequately addressed, or not seen to be important. The department stipulates that students attend a school that is closest to their home, to access available government transportation to and from school. Equivalently, parents of students with a disability decide where their children attend school. A choice between a regular classroom, mainstream Special Unit classroom, or a School for Specific Purposes (SSP). Parents who take their children to regular schools may be oblivious of the details of the curriculum. Physical Education outcomes does not stipulate the extent to which a student must perform or expected to perform. It is therefore due to the classroom teacher to adjust their teaching goals or outcomes to suit all students in their classroom. A student who can run a hundred meters race in 20 seconds may belong in the same classroom as a student in a wheelchair. While these students are challenged because of a lack of performance, teachers are challenged to effectively teach successful PE lessons, and on the other hand students without a disability may not be able to attain their optimum. This paper will identify areas of need, address the challenges, and explore a possible solution.

Keywords: special needs, disability, challenges, physical education

Procedia PDF Downloads 49
1074 The Malfatti’s Problem in Reuleaux Triangle

Authors: Ching-Shoei Chiang

Abstract:

The Malfatti’s Problem is to ask for fitting 3 circles into a right triangle such that they are tangent to each other, and each circle is also tangent to a pair of the triangle’s side. This problem has been extended to any triangle (called general Malfatti’s Problem). Furthermore, the problem has been extended to have 1+2+…+n circles, we call it extended general Malfatti’s problem, these circles whose tangency graph, using the center of circles as vertices and the edge connect two circles center if these two circles tangent to each other, has the structure as Pascal’s triangle, and the exterior circles of these circles tangent to three sides of the triangle. In the extended general Malfatti’s problem, there are closed-form solutions for n=1, 2, and the problem becomes complex when n is greater than 2. In solving extended general Malfatti’s problem (n>2), we initially give values to the radii of all circles. From the tangency graph and current radii, we can compute angle value between two vectors. These vectors are from the center of the circle to the tangency points with surrounding elements, and these surrounding elements can be the boundary of the triangle or other circles. For each circle C, there are vectors from its center c to its tangency point with its neighbors (count clockwise) pi, i=0, 1,2,..,n. We add all angles between cpi to cp(i+1) mod (n+1), i=0,1,..,n, call it sumangle(C) for circle C. Using sumangle(C), we can reduce/enlarge the radii for all circles in next iteration, until sumangle(C) is equal to 2πfor all circles. With a similar idea, this paper proposed an algorithm to find the radii of circles whose tangency has the structure of Pascal’s triangle, and the exterior circles of these circles are tangent to the unit Realeaux Triangle.

Keywords: Malfatti’s problem, geometric constraint solver, computer-aided geometric design, circle packing, data visualization

Procedia PDF Downloads 116
1073 Transesterification of Refined Palm Oil to Biodiesel in a Continuous Spinning Disc Reactor

Authors: Weerinda Appamana, Jirapong Keawkoon, Yamonporn Pacthong, Jirathiti Chitsanguansuk, Yanyong Sookklay

Abstract:

In the present work, spinning disc reactor has been used for the intensification of synthesis of biodiesel from refined palm oil (RPO) based on the transesterification reaction. Experiments have been performed using different spinning disc surface and under varying operating parameters viz. molar ratio of oil to methanol (over the range of 1:4.5–1:9), rotational speed (over the range of 500–2,000 rpm), total flow rate (over the range of 260-520 ml/min), and KOH catalyst loading of 1.50% by weight of oil. Maximum FAME (fatty acid methyl esters) yield (97.5 %) of biodiesel from RPO was obtained at oil to methanol ratio of 1:6, temperature of 60 °C, and rotational speed of 1500 rpm and flow rate of 520 mL/min using groove disc at KOH catalyst loading of 1.5 wt%. Also, higher yield efficiency (biodiesel produced per unit energy consumed) was obtained for using the spinning disc reactor based approach as compared to the ultrasound hydrodynamic cavitation and conventional mechanical stirrer reactors. It obviously offers a significant reduction in the reaction time for the transesterification, especially when compared with the reaction time of 90 minutes required for the conventional mechanical stirrer. It can be concluded that the spinning disk reactor is a promising alternative method for continuous biodiesel production.

Keywords: spinning disc reactor, biodiesel, process intensification, yield efficiency

Procedia PDF Downloads 144
1072 Innovative Fabric Integrated Thermal Storage Systems and Applications

Authors: Ahmed Elsayed, Andrew Shea, Nicolas Kelly, John Allison

Abstract:

In northern European climates, domestic space heating and hot water represents a significant proportion of total primary total primary energy use and meeting these demands from a national electricity grid network supplied by renewable energy sources provides an opportunity for a significant reduction in EU CO2 emissions. However, in order to adapt to the intermittent nature of renewable energy generation and to avoid co-incident peak electricity usage from consumers that may exceed current capacity, the demand for heat must be decoupled from its generation. Storage of heat within the fabric of dwellings for use some hours, or days, later provides a route to complete decoupling of demand from supply and facilitates the greatly increased use of renewable energy generation into a local or national electricity network. The integration of thermal energy storage into the building fabric for retrieval at a later time requires much evaluation of the many competing thermal, physical, and practical considerations such as the profile and magnitude of heat demand, the duration of storage, charging and discharging rate, storage media, space allocation, etc. In this paper, the authors report investigations of thermal storage in building fabric using concrete material and present an evaluation of several factors that impact upon performance including heating pipe layout, heating fluid flow velocity, storage geometry, thermo-physical material properties, and also present an investigation of alternative storage materials and alternative heat transfer fluids. Reducing the heating pipe spacing from 200 mm to 100 mm enhances the stored energy by 25% and high-performance Vacuum Insulation results in heat loss flux of less than 3 W/m2, compared to 22 W/m2 for the more conventional EPS insulation. Dense concrete achieved the greatest storage capacity, relative to medium and light-weight alternatives, although a material thickness of 100 mm required more than 5 hours to charge fully. Layers of 25 mm and 50 mm thickness can be charged in 2 hours, or less, facilitating a fast response that could, aggregated across multiple dwellings, provide significant and valuable reduction in demand from grid-generated electricity in expected periods of high demand and potentially eliminate the need for additional new generating capacity from conventional sources such as gas, coal, or nuclear.

Keywords: fabric integrated thermal storage, FITS, demand side management, energy storage, load shifting, renewable energy integration

Procedia PDF Downloads 159
1071 Structural Invertibility and Optimal Sensor Node Placement for Error and Input Reconstruction in Dynamic Systems

Authors: Maik Kschischo, Dominik Kahl, Philipp Wendland, Andreas Weber

Abstract:

Understanding and modelling of real-world complex dynamic systems in biology, engineering and other fields is often made difficult by incomplete knowledge about the interactions between systems states and by unknown disturbances to the system. In fact, most real-world dynamic networks are open systems receiving unknown inputs from their environment. To understand a system and to estimate the state dynamics, these inputs need to be reconstructed from output measurements. Reconstructing the input of a dynamic system from its measured outputs is an ill-posed problem if only a limited number of states is directly measurable. A first requirement for solving this problem is the invertibility of the input-output map. In our work, we exploit the fact that invertibility of a dynamic system is a structural property, which depends only on the network topology. Therefore, it is possible to check for invertibility using a structural invertibility algorithm which counts the number of node disjoint paths linking inputs and outputs. The algorithm is efficient enough, even for large networks up to a million nodes. To understand structural features influencing the invertibility of a complex dynamic network, we analyze synthetic and real networks using the structural invertibility algorithm. We find that invertibility largely depends on the degree distribution and that dense random networks are easier to invert than sparse inhomogeneous networks. We show that real networks are often very difficult to invert unless the sensor nodes are carefully chosen. To overcome this problem, we present a sensor node placement algorithm to achieve invertibility with a minimum set of measured states. This greedy algorithm is very fast and also guaranteed to find an optimal sensor node-set if it exists. Our results provide a practical approach to experimental design for open, dynamic systems. Since invertibility is a necessary condition for unknown input observers and data assimilation filters to work, it can be used as a preprocessing step to check, whether these input reconstruction algorithms can be successful. If not, we can suggest additional measurements providing sufficient information for input reconstruction. Invertibility is also important for systems design and model building. Dynamic models are always incomplete, and synthetic systems act in an environment, where they receive inputs or even attack signals from their exterior. Being able to monitor these inputs is an important design requirement, which can be achieved by our algorithms for invertibility analysis and sensor node placement.

Keywords: data-driven dynamic systems, inversion of dynamic systems, observability, experimental design, sensor node placement

Procedia PDF Downloads 131
1070 Production of Hydrogen and Carbon Monoxide Fuel Gas From Pine Needles

Authors: Despina Vamvuka, Despina Pentari

Abstract:

Forestry wastes are readily available in large quantities around the world. Based on European Green Deal for the deployment of renewable and decarbonized energy by 2050, as well as global energy crisis, energy recovery from such wastes reducing greenhouse gas emissions is very attractive. Gasification has superior environmental performance to combustion, producing a clean fuel gas utilized in internal combustion engines, gas turbines, solid oxide fuel cells, or for synthesis of liquid bio-fuels and value-added chemicals. In this work, pine needles, which are abundantly found in Mediterranean countries, were gasified by either steam or carbon dioxide via a two-step process to improve reactivity and eliminate tar, employing a fixed bed unit and a thermal analysis system. Solid, liquid and gaseous products from the whole process were characterized and their energy potential was determined. Thermal behaviour, reactivity, conversion and energy recovery were examined. The gasification process took place above 650°C. At 950°C conversion and energy recovery were 77% dry and 2 under a flow of steam and 85% dry and 2.9 under a flow of carbon dioxide, respectively. Organic matter was almost completely converted to syngas, the yield of which varied between 89% and 99%. The higher heating values of biochar, bio-oil and pyrolysis gas were 27.8 MJ/kg, 33.5 MJ/kg and 13.6 MJ/m3. Upon steam or carbon dioxide gasification, the higher heating value of syngas produced was 11.5 MJ/m3 and 12.7 MJ/m3, respectively.

Keywords: gasification, biomass, steam, carbon dioxide

Procedia PDF Downloads 83
1069 A Conceptual Study for Investigating the Creation of Energy and Understanding the Properties of Nothing

Authors: Mahmoud Reza Hosseini

Abstract:

The universe is in a continuous expansion process, resulting in the reduction of its density and temperature. Also, by extrapolating back from its current state, the universe at its early times is studied, known as the big bang theory. According to this theory, moments after creation, the universe was an extremely hot and dense environment. However, its rapid expansion due to nuclear fusion led to a reduction in its temperature and density. This is evidenced through the cosmic microwave background and the universe structure at a large scale. However, extrapolating back further from this early state reaches singularity, which cannot be explained by modern physics, and the big bang theory is no longer valid. In addition, one can expect a nonuniform energy distribution across the universe from a sudden expansion. However, highly accurate measurements reveal an equal temperature mapping across the universe, which is contradictory to the big bang principles. To resolve this issue, it is believed that cosmic inflation occurred at the very early stages of the birth of the universe. According to the cosmic inflation theory, the elements which formed the universe underwent a phase of exponential growth due to the existence of a large cosmological constant. The inflation phase allows the uniform distribution of energy so that an equal maximum temperature can be achieved across the early universe. Also, the evidence of quantum fluctuations of this stage provides a means for studying the types of imperfections the universe would begin with. Although well-established theories such as cosmic inflation and the big bang together provide a comprehensive picture of the early universe and how it evolved into its current state, they are unable to address the singularity paradox at the time of universe creation. Therefore, a practical model capable of describing how the universe was initiated is needed. This research series aims at addressing the singularity issue by introducing a state of energy called a "neutral state," possessing an energy level that is referred to as the "base energy." The governing principles of base energy are discussed in detail in our second paper in the series "A Conceptual Study for Addressing the Singularity of the Emerging Universe," which is discussed in detail. To establish a complete picture, the origin of the base energy should be identified and studied. In this research paper, the mechanism which led to the emergence of this natural state and its corresponding base energy is proposed. In addition, the effect of the base energy in the space-time fabric is discussed. Finally, the possible role of the base energy in quantization and energy exchange is investigated. Therefore, the proposed concept in this research series provides a road map for enhancing our understating of the universe's creation from nothing and its evolution and discusses the possibility of base energy as one of the main building blocks of this universe.

Keywords: big bang, cosmic inflation, birth of universe, energy creation, universe evolution

Procedia PDF Downloads 77
1068 Using Soil Texture Field Observations as Ordinal Qualitative Variables for Digital Soil Mapping

Authors: Anne C. Richer-De-Forges, Dominique Arrouays, Songchao Chen, Mercedes Roman Dobarco

Abstract:

Most of the digital soil mapping (DSM) products rely on machine learning (ML) prediction models and/or the use or pedotransfer functions (PTF) in which calibration data come from soil analyses performed in labs. However, many other observations (often qualitative, nominal, or ordinal) could be used as proxies of lab measurements or as input data for ML of PTF predictions. DSM and ML are briefly described with some examples taken from the literature. Then, we explore the potential of an ordinal qualitative variable, i.e., the hand-feel soil texture (HFST) estimating the mineral particle distribution (PSD): % of clay (0-2µm), silt (2-50µm) and sand (50-2000µm) in 15 classes. The PSD can also be measured by lab measurements (LAST) to determine the exact proportion of these particle-sizes. However, due to cost constraints, HFST are much more numerous and spatially dense than LAST. Soil texture (ST) is a very important soil parameter to map as it is controlling many of the soil properties and functions. Therefore, comes an essential question: is it possible to use HFST as a proxy of LAST for calibration and/or validation of DSM predictions of ST? To answer this question, the first step is to compare HFST with LAST on a representative set where both information are available. This comparison was made on ca 17,400 samples representative of a French region (34,000 km2). The accuracy of HFST was assessed, and each HFST class was characterized by a probability distribution function (PDF) of its LAST values. This enables to randomly replace HFST observations by LAST values while respecting the PDF previously calculated and results in a very large increase of observations available for the calibration or validation of PTF and ML predictions. Some preliminary results are shown. First, the comparison between HFST classes and LAST analyses showed that accuracies could be considered very good when compared to other studies. The causes of some inconsistencies were explored and most of them were well explained by other soil characteristics. Then we show some examples applying these relationships and the increase of data to several issues related to DSM. The first issue is: do the PDF functions that were established enable to use HSFT class observations to improve the LAST soil texture prediction? For this objective, we replaced all HFST for topsoil by values from the PDF 100 time replicates). Results were promising for the PTF we tested (a PTF predicting soil water holding capacity). For the question related to the ML prediction of LAST soil texture on the region, we did the same kind of replacement, but we implemented a 10-fold cross-validation using points where we had LAST values. We obtained only preliminary results but they were rather promising. Then we show another example illustrating the potential of using HFST as validation data. As in numerous countries, the HFST observations are very numerous; these promising results pave the way to an important improvement of DSM products in all the countries of the world.

Keywords: digital soil mapping, improvement of digital soil mapping predictions, potential of using hand-feel soil texture, soil texture prediction

Procedia PDF Downloads 205
1067 Survey of Neonatologists’ Burnout on a Neonatal Surgical Unit: Audit Study from Cairo University Specialized Pediatric Hospital

Authors: Mahmoud Tarek, Alaa Obeida, Mai Magdy, Khalid Hussein, Aly Shalaby

Abstract:

Background: More doctors are complaining of burnout than before, Burnout is a state of physical and mental exhaustion caused by the doctor’s lifestyle, unfortunately, Medical errors are also more likely in those suffering from burnout and these may result in malpractice suits. Methodology: It is a retrospective audit of burnout response on all neonatologists over a 9 months period. We gathered data using burnout questionnaire, it was obtained from 23 physicians, the physicians divided into 5 categories according to the final score of the 28 questions in the questionnaire. Category 1 with score from 28-38 with almost no work stress, category 2 with score (38-50) who express a low amount of job related stress, category 3 with score (51-70) with moderate amount of stress, category 4 with score (71-90) those express a high amount of job stress and begun to burnout, category 5 with score (91 and above) who are under a dangerous amount of stress and advanced stage of burnout. Results: 33 neonatologists have received the questionnaire, 23 responses were sent back with a response rate of 69.6%. The results showed that 61% of physicians fall in category 4, 31% of the physician in category 5, while 8% of physicians equally distributed between category 2 and 3 (4% each of them). On the other hand, there is no physician present in category 1. Conclusion: Burnout is prevalent in SNICUs, So interventions to minimize burnout prevalence may be of greater importance as this may be reflected indirectly on medical conditions of the patients and physicians, efforts should be done to decrease this high rate of burnout.

Keywords: Cairo, work overload, exhaustion, surgery, neonatal ICU

Procedia PDF Downloads 187
1066 Income-Consumption Relationships in Pakistan (1980-2011): A Cointegration Approach

Authors: Himayatullah Khan, Alena Fedorova

Abstract:

The present paper analyses the income-consumption relationships in Pakistan using annual time series data from 1980-81 to 2010-1. The paper uses the Augmented Dickey-Fuller test to check the unit root and stationarity in these two time series. The paper finds that the two time series are nonstationary but stationary at their first difference levels. The Augmented Engle-Granger test and the Cointegrating Regression Durbin-Watson test imply that the two time series of consumption and income are cointegrated and that long-run marginal propensity to consume is 0.88 which is given by the estimated (static) equilibrium relation. The paper also used the error correction mechanism to find out to model dynamic relationship. The purpose of the ECM is to indicate the speed of adjustment from the short-run equilibrium to the long-run equilibrium state. The results show that MPC is equal to 0.93 and is highly significant. The coefficient of Engle-Granger residuals is negative but insignificant. Statistically, the equilibrium error term is zero, which suggests that consumption adjusts to changes in GDP in the same period. The short-run changes in GDP have a positive impact on short-run changes in consumption. The paper concludes that we may interpret 0.93 as the short-run MPC. The pair-wise Granger Causality test shows that both GDP and consumption Granger cause each other.

Keywords: cointegrating regression, Augmented Dickey Fuller test, Augmented Engle-Granger test, Granger causality, error correction mechanism

Procedia PDF Downloads 397
1065 A Framework for Event-Based Monitoring of Business Processes in the Supply Chain Management of Industry 4.0

Authors: Johannes Atug, Andreas Radke, Mitchell Tseng, Gunther Reinhart

Abstract:

In modern supply chains, large numbers of SKU (Stock-Keeping-Unit) need to be timely managed, and any delays in noticing disruptions of items often limit the ability to defer the impact on customer order fulfillment. However, in supply chains of IoT-connected enterprises, the ERP (Enterprise-Resource-Planning), the MES (Manufacturing-Execution-System) and the SCADA (Supervisory-Control-and-Data-Acquisition) systems generate large amounts of data, which generally glean much earlier notice of deviations in the business process steps. That is, analyzing these streams of data with process mining techniques allows the monitoring of the supply chain business processes and thus identification of items that deviate from the standard order fulfillment process. In this paper, a framework to enable event-based SCM (Supply-Chain-Management) processes including an overview of core enabling technologies are presented, which is based on the RAMI (Reference-Architecture-Model for Industrie 4.0) architecture. The application of this framework in the industry is presented, and implications for SCM in industry 4.0 and further research are outlined.

Keywords: cyber-physical production systems, event-based monitoring, supply chain management, RAMI (Reference-Architecture-Model for Industrie 4.0)

Procedia PDF Downloads 215
1064 Agrowastes to Edible Hydrogels through Bio Nanotechnology Interventions: Bioactive from Mandarin Peels

Authors: Niharika Kaushal, Minni Singh

Abstract:

Citrus fruits contain an abundance of phytochemicals that can promote health. A substantial amount of agrowaste is produced from the juice processing industries, primarily peels and seeds. This leftover agrowaste is a reservoir of nutraceuticals, particularly bioflavonoids which render it antioxidant and potentially anticancerous. It is, therefore, favorable to utilize this biomass and contribute towards sustainability in a manner that value-added products may be derived from them, nutraceuticals, in this study. However, the pre-systemic metabolism of flavonoids in the gastric phase limits the effectiveness of these bioflavonoids derived from mandarin biomass. In this study, ‘kinnow’ mandarin (Citrus nobilis X Citrus deliciosa) biomass was explored for its flavonoid profile. This work entails supercritical fluid extraction and identification of bioflavonoids from mandarin biomass. Furthermore, to overcome the limitations of these flavonoids in the gastrointestinal tract, a double-layered vehicular mechanism comprising the fabrication of nanoconjugates and edible hydrogels was adopted. Total flavonoids in the mandarin peel extract were estimated by the aluminum chloride complexation method and were found to be 47.3±1.06 mg/ml rutin equivalents as total flavonoids. Mass spectral analysis revealed the abundance of polymethoxyflavones (PMFs), nobiletin and tangeretin as the major flavonoids in the extract, followed by hesperetin and naringenin. Furthermore, the antioxidant potential was analyzed by the 2,2-diphenyl-1-picrylhydrazyl (DPPH) method, which showed an IC50 of 0.55μg/ml. Nanoconjugates were fabricated via the solvent evaporation method, which was further impregnated into hydrogels. Additionally, the release characteristics of nanoconjugate-laden hydrogels in a simulated gastrointestinal environment were studied. The PLGA-PMFs nanoconjugates exhibited a particle size between 200-250nm having a smooth and spherical shape as revealed by FE-SEM. The impregnated alginate hydrogels offered a dense network that ensured the holding of PLGA-PMF nanoconjugates, as confirmed by Cryo-SEM images. Rheological studies revealed the shear-thinning behavior of hydrogels and their high resistance to deformation. Gastrointestinal studies showed a negligible 4.0% release of flavonoids in the gastric phase, followed by a sustained release over the next hours in the intestinal environment. Therefore, based on the enormous potential of recovering nutraceuticals from agro-processing wastes, further augmented by nanotechnological interventions for enhancing the bioefficacy of these compounds, lays the foundation for exploring the path towards the development of value-added products, thereby contributing towards the sustainable use of agrowaste.

Keywords: agrowaste, gastrointestinal, hydrogel, nutraceuticals

Procedia PDF Downloads 76
1063 Numerical Modeling of hybrid Photovoltaic-Thermoelectric Solar Unit by Applying Various Cross-Sections of Cooling Ducts

Authors: Ziba Khalili, Mohsen Sheikholeslami, Ladan Momayez

Abstract:

Combining the photovoltaic/thermal (PVT) systems with a thermoelectric (TE) module can raise energy yields since the TE module boosts the system's energy conversion efficiency. In the current study, a PVT system integrated with a TE module was designed and simulated in ANSYS Fluent 19.2. A copper heat transfer tube (HTT) was employed for cooling the photovoltaic (PV) cells. Four different shapes of HTT cross-section, i.e., circular, square, elliptical, and triangular, with equal cross-section areas were investigated. Also, the influence of Cu-Al2O3/water hybrid nanofluid (0.024% volume concentration), fluid inlet velocity (uᵢ ), and amount of solar radiation (G), on the PV temperature (Tₚᵥ) and system performance were investigated. The ambient temperature (Tₐ), wind speed (u𝓌), and fluid inlet temperature (Tᵢ), were considered to be 25°C, 1 m/s, and 27°C, respectively. According to the obtained data, the triangular case had the greatest impact on reducing the compared to other cases. In the triangular case, examination of the effect of hybrid nanofluid showed that the use of hybrid nanofluid at 800 W/m2 led to a reduction of the TPV by 0.6% compared to water, at 0.19 m/s. Moreover, the thermal efficiency ( ) and the overall electrical efficiency (nₜ) of the system improved by 0.93% and 0.22%, respectively, at 0.19 m/s. In a triangular case where G and were 800 W/m2 and 19 m/s, respectively, the highest amount of, thermal power (Eₜ), and, were obtained as 72.76%, 130.84 W and 12.03%, respectively.

Keywords: electrical performance, photovoltaic/thermal, thermoelectric, hybrid nanofluid, thermal efficiency

Procedia PDF Downloads 56
1062 Cladding Technology for Metal-Hybrid Composites with Network-Structure

Authors: Ha-Guk Jeong, Jong-Beom Lee

Abstract:

Cladding process is very typical technology for manufacturing composite materials by the hydrostatic extrusion. Because there is no friction between the metal and the container, it can be easily obtained in uniform flow during the deformation. The general manufacturing process for a metal-matrix composite in the solid state, mixing metal powders and ceramic powders with a suited volume ratio, prior to be compressed or extruded at the cold or hot condition in a can. Since through a plurality of unit processing steps of dispersing the materials having a large difference in their characteristics and physical mixing, the process is complicated and leads to non-uniform dispersion of ceramics. It is difficult and hard to reach a uniform ideal property in the coherence problems at the interface between the metal and the ceramic reinforcements. Metal hybrid composites, which presented in this report, are manufactured through the traditional plastic deformation processes like hydrostatic extrusion, caliber-rolling, and drawing. By the previous process, the realization of uniform macro and microstructure is surely possible. In this study, as a constituent material, aluminum, copper, and titanium have been used, according to the component ratio, excellent characteristics of each material were possible to produce a metal hybrid composite that appears to maximize. MgB₂ superconductor wire also fabricated via the same process. It will be introduced to their unique artistic and thermal characteristics.

Keywords: cladding process, metal-hybrid composites, hydrostatic extrusion, electronic/thermal characteristics

Procedia PDF Downloads 162
1061 Application of Supervised Deep Learning-based Machine Learning to Manage Smart Homes

Authors: Ahmed Al-Adaileh

Abstract:

Renewable energy sources, domestic storage systems, controllable loads and machine learning technologies will be key components of future smart homes management systems. An energy management scheme that uses a Deep Learning (DL) approach to support the smart home management systems, which consist of a standalone photovoltaic system, storage unit, heating ventilation air-conditioning system and a set of conventional and smart appliances, is presented. The objective of the proposed scheme is to apply DL-based machine learning to predict various running parameters within a smart home's environment to achieve maximum comfort levels for occupants, reduced electricity bills, and less dependency on the public grid. The problem is using Reinforcement learning, where decisions are taken based on applying the Continuous-time Markov Decision Process. The main contribution of this research is the proposed framework that applies DL to enhance the system's supervised dataset to offer unlimited chances to effectively support smart home systems. A case study involving a set of conventional and smart appliances with dedicated processing units in an inhabited building can demonstrate the validity of the proposed framework. A visualization graph can show "before" and "after" results.

Keywords: smart homes systems, machine learning, deep learning, Markov Decision Process

Procedia PDF Downloads 176
1060 Harnessing of Electricity from Distillery Effluent and Simultaneous Effluent Treatment by Microbial Fuel Cell

Authors: Hanish Mohammed, C. H. Muthukumar Muthuchamy

Abstract:

The advancement in the science and technology has made it possible to convert electrical energy into any desired form. It has given electrical energy a place of pride in the modern world. The survival of industrial undertakings and our social structure depends primarily upon low cost and uninterrupted supply of electrical energy. Microbial fuel cell (MFC) is a promising and emerging technique for sustainable bioelectricity generation and wastewater treatment. MFCs are devices which are capable of converting organic matter to electricity/hydrogen with help of microorganisms. Different kinds of wastewater could be used in this technique, distillery effluent is one of the most troublesome and complex and strong organic effluent with high chemical oxygen demand of 1,53,846 mg/L. A single cell MFC unit was designed and fabricated for the distillery effluent treatment and to generate electricity. Due to the high COD value of the distillery effluent helped in the production of energy for 74 days. The highest voltage got from the fuel cell is 206 mV on the 30th day. A maximum power density obtained from the MFC was 9.8 mW, treatment efficiency was evaluated in terms of COD removal and other parameters. COD removal efficiencies were around 68.5 % and other parameters such as Total Hardness (81.5%), turbidity (70 %), chloride (66%), phosphate (79.5%), Nitrate (77%) and sulphate (71%). MFC using distillery effluent is a promising new unexplored substrate for the power generation and sustainable treatment technique through harnessing of bioelectricity.

Keywords: microbial fuel cell (MFC), bioelectricity, distillery effluent, wastewater treatment

Procedia PDF Downloads 199
1059 Intelligent Indoor Localization Using WLAN Fingerprinting

Authors: Gideon C. Joseph

Abstract:

The ability to localize mobile devices is quite important, as some applications may require location information of these devices to operate or deliver better services to the users. Although there are several ways of acquiring location data of mobile devices, the WLAN fingerprinting approach has been considered in this work. This approach uses the Received Signal Strength Indicator (RSSI) measurement as a function of the position of the mobile device. RSSI is a quantitative technique of describing the radio frequency power carried by a signal. RSSI may be used to determine RF link quality and is very useful in dense traffic scenarios where interference is of major concern, for example, indoor environments. This research aims to design a system that can predict the location of a mobile device, when supplied with the mobile’s RSSIs. The developed system takes as input the RSSIs relating to the mobile device, and outputs parameters that describe the location of the device such as the longitude, latitude, floor, and building. The relationship between the Received Signal Strengths (RSSs) of mobile devices and their corresponding locations is meant to be modelled; hence, subsequent locations of mobile devices can be predicted using the developed model. It is obvious that describing mathematical relationships between the RSSIs measurements and localization parameters is one option to modelling the problem, but the complexity of such an approach is a serious turn-off. In contrast, we propose an intelligent system that can learn the mapping of such RSSIs measurements to the localization parameters to be predicted. The system is capable of upgrading its performance as more experiential knowledge is acquired. The most appealing consideration to using such a system for this task is that complicated mathematical analysis and theoretical frameworks are excluded or not needed; the intelligent system on its own learns the underlying relationship in the supplied data (RSSI levels) that corresponds to the localization parameters. These localization parameters to be predicted are of two different tasks: Longitude and latitude of mobile devices are real values (regression problem), while the floor and building of the mobile devices are of integer values or categorical (classification problem). This research work presents artificial neural network based intelligent systems to model the relationship between the RSSIs predictors and the mobile device localization parameters. The designed systems were trained and validated on the collected WLAN fingerprint database. The trained networks were then tested with another supplied database to obtain the performance of trained systems on achieved Mean Absolute Error (MAE) and error rates for the regression and classification tasks involved therein.

Keywords: indoor localization, WLAN fingerprinting, neural networks, classification, regression

Procedia PDF Downloads 327
1058 Potentials of Additive Manufacturing: An Approach to Increase the Flexibility of Production Systems

Authors: A. Luft, S. Bremen, N. Balc

Abstract:

The task of flexibility planning and design, just like factory planning, for example, is to create the long-term systemic framework that constitutes the restriction for short-term operational management. This is a strategic challenge since, due to the decision defect character of the underlying flexibility problem, multiple types of flexibility need to be considered over the course of various scenarios, production programs, and production system configurations. In this context, an evaluation model has been developed that integrates both conventional and additive resources on a basic task level and allows the quantification of flexibility enhancement in terms of mix and volume flexibility, complexity reduction, and machine capacity. The model helps companies to decide in early decision-making processes about the potential gains of implementing additive manufacturing technologies on a strategic level. For companies, it is essential to consider both additive and conventional manufacturing beyond pure unit costs. It is necessary to achieve an integrative view of manufacturing that incorporates both additive and conventional manufacturing resources and quantifies their potential with regard to flexibility and manufacturing complexity. This also requires a structured process for the strategic production systems design that spans the design of various scenarios and allows for multi-dimensional and comparative analysis. A respective guideline for the planning of additive resources on a strategic level is being laid out in this paper.

Keywords: additive manufacturing, production system design, flexibility enhancement, strategic guideline

Procedia PDF Downloads 105
1057 Comparative Histological, Immunohistochemical and Biochemical Study on the Effect of Vit. C, Vit. E, Gallic Acid and Silymarin on Carbon Tetrachloride Model of Liver Fibrosis in Rats

Authors: Safaa S. Hassan, Mohammed H. Elbakry, Safwat A. Mangoura, Zainab M. Omar

Abstract:

Background: Liver fibrosis is the main reason for increased mortality in chronic liver disease. It has no standard treatment. Antioxidants from a variety of sources are capable of slowing or preventing oxidation of other molecules. Aim: to evaluate the hepatoprotective effect of vit. C, vit. E and gallic acid in comparison to silymarin in the rat model of carbon tetrachloride induced liver fibrosis and their possible mechanisms of action. Material& Methods: A total number of 60 adult male albino rats 160-200gm were divided into six equal groups; received subcutaneous (s.c) injection for 8 weeks. Group I: as control. Group II: received 1.5 mL/kg of CCL4 .Group III: CCL4 and co- treatment with silymarin 100mg/kg p.o. daily. Group IV: CCL4 and co-treatment with vit. C 50mg/kg p.o. daily. Group V: CCL4 and co-treatment with vit. E 200mg/kg. p.o. Group VI: CCL4 and co-treatment with Gallic acid 100mg/kg. p.o. daily. Liver was processed for histological and immunohistochemical examination. Levels of AST, ALT, ALP, reduced GSH, MDA, SOD and hydroxyproline concentration were measured and evaluated statistically. Results: Light and electron microscopic examination of liver of group II exhibited foci of altered cells with dense nuclei and vacuolated, granular cytoplasm, mononuclear cell infiltration in portal areas, profuse collagen fiber deposits were found around portal tract, more intense staining α-SMA-positive cells occupied most of the liver fibrosis tissue, electron lucent areas in the cytoplasm of the hepatocytes, margination of nuclear chromatin. Treatment by any of the antioxidants variably reduced the hepatic structural changes induced by CCL4. Biochemical analysis showed that carbon tetrachloride significantly increased the levels of serum AST, ALT, ALP, hepatic malondialdehyde and hydroxyproline content. Moreover, it decreased the activities of superoxide dismutase and glutathione. Treatment with silymarin, gallic acid, vit. C and vit. E decreased significantly the AST, ALT, and ALP levels in plasma, MDA and hydroxyproline and increased the activities of SOD and glutathione in liver tissue. The effect of administration of CCl4 was improved with the used antioxidants in variable degrees. The most efficient antioxidant was silymarin followed by gallic acid and vit. C then vit. E. It is possibly due to their antioxidant effect, free radical scavenging properties and the reduction of oxidant dependent activation and proliferation of HSCs. Conclusion: So these antioxidants can be a promising drugs candidate for ameliorating liver fibrosis better than the use of the drugs and their side effects.

Keywords: antioxidant, ccl4, gallic acid, liver fibrosis

Procedia PDF Downloads 255
1056 Influence of Mothers’ Knowledge, Attitude and Behavior on Diet and Physical Activity of Their Pre-School Children: A Cross-Sectional Study from a Semi-Urban Area of Nepal

Authors: Natalia Oli, Abhinav Vaidya, Katja Pahkala, Gabriele Eiben, Alexandra Krettek

Abstract:

The nutritional transition towards a high fat and energy dense diet, decreasing physical activity level, and poor cardiovascular health knowledge contributes to a rising burden of cardiovascular diseases in Nepal. Dietary and physical activity behaviors are formed early in life and influenced by family, particularly by mothers in the social context of Nepal. The purpose of this study was to explore knowledge, attitude and behavior of mothers regarding diet and physical activity of their pre-school children. Cross-sectional study was conducted in the semi-urban area of Duwakot and Jhaukhel communities near the capital Kathmandu. Between August and November 2014, nine trained enumerators interviewed all mothers having children aged 2 to 7 years in their homes. Questionnaire contained information about mothers’ socio-demographic characteristics; their knowledge, attitude, and behavior regarding diet and physical activity as well as their children’s diet and physical activity. Knowledge, attitude and behavior responses were scored. SPSS version 22.0 was used for data analyses. Out of the 1,052 eligible mothers, 962 consented to participate in the study. The mean age was 28.9 ± 4.5 years. The majority of them (73%) were housewives. Mothers with higher education and income had higher knowledge, attitude, and behavior scores (All p < 0.001) whereas housewives and farmers had low knowledge score (p < 0.001). They, along with laborers, also exhibited lower attitude (p<0.001) and behavior scores (p < 0.001). Children’s diet score increased with mothers’ level of education (p <0.001) and income (p=0.041). Their physical activity score, however, declined with increasing level of their mothers’ education (p < 0.001) and income (p < 0.001). Children’s overall behavior score correlated poorly with mothers’ knowledge (r = 0.009, p=0.003), attitude (r =0.012, p=0.001), and behavior (r = 0.007, p= 0.008). Such poor correlation can be due to existence of the barriers among mothers. Mothers reported such barriers as expensive healthy food, difficulty to give up favorite food, taste preference of others family members and lack of knowledge on healthy food. Barriers for physical activity were lack of leisure time, lack of parks and playgrounds, being busy by caring for children and old people, feeling lazy and embarrassed in front of others. Additionally, among the facilitators for healthy lifestyle, mentioned by mothers, were better information, family eating healthy food and supporting physical activity, advice of medical personnel regarding healthy lifestyle and own ill health. The study demonstrated poor correlation of mothers’ knowledge and attitude with children’s behavior regarding diet and physical activity. Hence improving mothers’ knowledge or attitude may not be enough to improve dietary and physical activity habits of their children. Barriers and facilitators that affect mothers’ practices towards their children should also be addressed due to future intervention.

Keywords: attitude, behavior, diet, knowledge, mothers, physical activity

Procedia PDF Downloads 271
1055 Effect of Shrinkage on Heat and Mass Transfer Parameters of Solar Dried Potato Samples of Variable Diameter

Authors: Kshanaprava Dhalsamant, Punyadarshini P. Tripathy, Shanker L. Shrivastava

Abstract:

Potato is chosen as the food product for carrying out the natural convection mixed-mode solar drying experiments since they are easily available and globally consumed. The convective heat and mass transfer coefficients along with effective diffusivity were calculated considering both shrinkage and without shrinkage for the potato cylinders of different geometry (8, 10 and 13 mm diameters and a constant length of 50 mm). The convective heat transfer coefficient (hc) without considering shrinkage effect were 24.28, 18.69, 15.89 W/m2˚C and hc considering shrinkage effect were 37.81, 29.21, 25.72 W/m2˚C for 8, 10 and 13 mm diameter samples respectively. Similarly, the effective diffusivity (Deff) without considering shrinkage effect were 3.20×10-9, 4.82×10-9, 2.48×10-8 m2/s and Deff considering shrinkage effect were 1.68×10-9, 2.56×10-9, 1.34×10-8 m2/s for 8, 10 and 13 mm diameter samples respectively and the mass transfer coefficient (hm) without considering the shrinkage effect were 5.16×10-7, 2.93×10-7, 2.59×10-7 m/s and hm considering shrinkage effect were 3.71×10-7, 2.04×10-7, 1.80×10-7 m/s for 8, 10 and 13 mm diameter samples respectively. Increased values of hc were obtained by considering shrinkage effect in all diameter samples because shrinkage results in decreasing diameter with time achieving in enhanced rate of water loss. The average values of Deff determined without considering the shrinkage effect were found to be almost double that with shrinkage effect. The reduction in hm values is due to the fact that with increasing sample diameter, the exposed surface area per unit mass decreases, resulting in a slower moisture removal. It is worth noting that considering shrinkage effect led to overestimation of hc values in the range of 55.72-61.86% and neglecting the shrinkage effect in the mass transfer analysis, the values of Deff and hm are overestimated in the range of 85.02-90.27% and 39.11-45.11%, respectively, for the range of sample diameter investigated in the present study.

Keywords: shrinkage, convective heat transfer coefficient, effectivive diffusivity, convective mass transfer coefficient

Procedia PDF Downloads 239
1054 Student Perceptions on Administrative Support in the Delivering of Open Distance Learning Programmes – A Case Study

Authors: E. J. Spamer, J. M. Van Zyl, MHA Combrinck

Abstract:

The Unit for Open Distance Learning (UODL) at the North-West University (NWU), South Africa was established in 2013 with its main function to deliver open distance learning (ODL) programmes to approximately 30 000 students from the Faculties of Education Sciences, Health Sciences, Theology and Arts and Culture. Quality operational and administrative processes are key components in the delivery of these programmes and they need to function optimally for students to be successful in their studies. Operational and administrative processes include aspects such as applications, registration, dissemination of study material, availability of electronic platforms, the management of assessment, and the dissemination of important information. To be able to ensure and enhance quality during these processes, it is vital to determine students’ perceptions with regards to these mentioned processes. A questionnaire was available online and also distributed to the 63 tuition centres. The purpose of this research was to determine the perceptions of ODL students from NWU regarding operational and administrative processes. 1903 students completed and submitted the questionnaire. The data was quantitatively analysed and discussed. Results indicated that the majority of students are satisfied with the operational and administrative processes; however, the results also indicated some areas that need improvement. The data gathered is important to identify strengths and areas for improvement and form part of a bigger strategy of qualitative assurance at the UODL.

Keywords: administrative support, ODL programmes, quantitative study, students' perceptions

Procedia PDF Downloads 254