Search results for: modular prototype
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 828

Search results for: modular prototype

168 Optimum Turbomachine Preliminary Selection for Power Regeneration in Vapor Compression Cool Production Plants

Authors: Sayyed Benyamin Alavi, Giovanni Cerri, Leila Chennaoui, Ambra Giovannelli, Stefano Mazzoni

Abstract:

Primary energy consumption and emissions of pollutants (including CO2) sustainability call to search methodologies to lower power absorption for unit of a given product. Cool production plants based on vapour compression are widely used for many applications: air conditioning, food conservation, domestic refrigerators and freezers, special industrial processes, etc. In the field of cool production, the amount of Yearly Consumed Primary Energy is enormous, thus, saving some percentage of it, leads to big worldwide impact in the energy consumption and related energy sustainability. Among various techniques to reduce power required by a Vapour Compression Cool Production Plant (VCCPP), the technique based on Power Regeneration by means of Internal Direct Cycle (IDC) will be considered in this paper. Power produced by IDC reduces power need for unit of produced Cool Power by the VCCPP. The paper contains basic concepts that lead to develop IDCs and the proposed options to use the IDC Power. Among various selections for using turbo machines, Best Economically Available Technologies (BEATs) have been explored. Based on vehicle engine turbochargers, they have been taken into consideration for this application. According to BEAT Database and similarity rules, the best turbo machine selection leads to the minimum nominal power required by VCCPP Main Compressor. Results obtained installing the prototype in “ad hoc” designed test bench will be discussed and compared with the expected performance. Forecasts for the upgrading VCCPP, various applications will be given and discussed. 4-6% saving is expected for air conditioning cooling plants and 15-22% is expected for cryogenic plants.

Keywords: Refrigeration Plant, Vapour Pressure Amplifier, Compressor, Expander, Turbine, Turbomachinery Selection, Power Saving

Procedia PDF Downloads 409
167 A Prototype of an Information and Communication Technology Based Intervention Tool for Children with Dyslexia

Authors: Rajlakshmi Guha, Sajjad Ansari, Shazia Nasreen, Hirak Banerjee, Jiaul Paik

Abstract:

Dyslexia is a neurocognitive disorder, affecting around fifteen percent of the Indian population. The symptoms include difficulty in reading alphabet, words, and sentences. This can be difficult at the phonemic or recognition level and may further affect lexical structures. Therapeutic intervention of dyslexic children post assessment is generally done by special educators and psychologists through one on one interaction. Considering the large number of children affected and the scarcity of experts, access to care is limited in India. Moreover, unavailability of resources and timely communication with caregivers add on to the problem of proper intervention. With the development of Educational Technology and its use in India, access to information and care has been improved in such a large and diverse country. In this context, this paper proposes an ICT enabled home-based intervention program for dyslexic children which would support the child, and provide an interactive interface between expert, parents, and students. The paper discusses the details of the database design and system layout of the program. Along with, it also highlights the development of different technical aids required to build out personalized android applications for the Indian dyslexic population. These technical aids include speech database creation for children, automatic speech recognition system, serious game development, and color coded fonts. The paper also emphasizes the games developed to assist the dyslexic child on cognitive training primarily for attention, working memory, and spatial reasoning. In addition, it talks about the specific elements of the interactive intervention tool that makes it effective for home based intervention of dyslexia.

Keywords: Android applications, cognitive training, dyslexia, intervention

Procedia PDF Downloads 276
166 Chassis Level Control Using Proportional Integrated Derivative Control, Fuzzy Logic and Deep Learning

Authors: Atakan Aral Ormancı, Tuğçe Arslantaş, Murat Özcü

Abstract:

This study presents the design and implementation of an experimental chassis-level system for various control applications. Specifically, the height level of the chassis is controlled using proportional integrated derivative, fuzzy logic, and deep learning control methods. Real-time data obtained from height and pressure sensors installed in a 6x2 truck chassis, in combination with pulse-width modulation signal values, are utilized during the tests. A prototype pneumatic system of a 6x2 truck is added to the setup, which enables the Smart Pneumatic Actuators to function as if they were in a real-world setting. To obtain real-time signal data from height sensors, an Arduino Nano is utilized, while a Raspberry Pi processes the data using Matlab/Simulink and provides the correct output signals to control the Smart Pneumatic Actuator in the truck chassis. The objective of this research is to optimize the time it takes for the chassis to level down and up under various loads. To achieve this, proportional integrated derivative control, fuzzy logic control, and deep learning techniques are applied to the system. The results show that the deep learning method is superior in optimizing time for a non-linear system. Fuzzy logic control with a triangular membership function as the rule base achieves better outcomes than proportional integrated derivative control. Traditional proportional integrated derivative control improves the time it takes to level the chassis down and up compared to an uncontrolled system. The findings highlight the superiority of deep learning techniques in optimizing the time for a non-linear system, and the potential of fuzzy logic control. The proposed approach and the experimental results provide a valuable contribution to the field of control, automation, and systems engineering.

Keywords: automotive, chassis level control, control systems, pneumatic system control

Procedia PDF Downloads 58
165 Digi-Buddy: A Smart Cane with Artificial Intelligence and Real-Time Assistance

Authors: Amaladhithyan Krishnamoorthy, Ruvaitha Banu

Abstract:

Vision is considered as the most important sense in humans, without which leading a normal can be often difficult. There are many existing smart canes for visually impaired with obstacle detection using ultrasonic transducer to help them navigate. Though the basic smart cane increases the safety of the users, it does not help in filling the void of visual loss. This paper introduces the concept of Digi-Buddy which is an evolved smart cane for visually impaired. The cane consists for several modules, apart from the basic obstacle detection features; the Digi-Buddy assists the user by capturing video/images and streams them to the server using a wide-angled camera, which then detects the objects using Deep Convolutional Neural Network. In addition to determining what the particular image/object is, the distance of the object is assessed by the ultrasonic transducer. The sound generation application, modelled with the help of Natural Language Processing is used to convert the processed images/object into audio. The object detected is signified by its name which is transmitted to the user with the help of Bluetooth hear phones. The object detection is extended to facial recognition which maps the faces of the person the user meets in the database of face images and alerts the user about the person. One of other crucial function consists of an automatic-intimation-alarm which is triggered when the user is in an emergency. If the user recovers within a set time, a button is provisioned in the cane to stop the alarm. Else an automatic intimation is sent to friends and family about the whereabouts of the user using GPS. In addition to safety and security by the existing smart canes, the proposed concept devices to be implemented as a prototype helping visually-impaired visualize their surroundings through audio more in an amicable way.

Keywords: artificial intelligence, facial recognition, natural language processing, internet of things

Procedia PDF Downloads 330
164 Performance Analysis of Organic Rankine Cycle Technology to Exploit Low-Grade Waste Heat to Power Generation in Indian Industry

Authors: Bipul Krishna Saha, Basab Chakraborty, Ashish Alex Sam, Parthasarathi Ghosh

Abstract:

The demand for energy is cumulatively increasing with time.  Since the availability of conventional energy resources is dying out gradually, significant interest is being laid on searching for alternate energy resources and minimizing the wastage of energy in various fields.  In such perspective, low-grade waste heat from several industrial sources can be reused to generate electricity. The present work is to further the adoption of the Organic Rankine Cycle (ORC) technology in Indian industrial sector.  The present paper focuses on extending the previously reported idea to the next level through a comparative review with three different working fluids using practical data from an Indian industrial plant. For comprehensive study in the simulation platform of Aspen Hysys®, v8.6, the waste heat data has been collected from a current coke oven gas plant in India.  A parametric analysis of non-regenerative ORC and regenerative ORC is executed using the working fluids R-123, R-11 and R-21 for subcritical ORC system.  The primary goal is to determine the optimal working fluid considering various system parameters like turbine work output, obtained system efficiency, irreversibility rate and second law efficiency under applied multiple heat source temperature (160 °C- 180 °C).  Selection of the turbo-expanders is one of the most crucial tasks for low-temperature applications in ORC system. The present work is an attempt to make suitable recommendation for the appropriate configuration of the turbine. In a nutshell, this study justifies the proficiency of integrating the ORC technology in Indian perspective and also finds the appropriate parameter of all components integrated in ORC system for building up an ORC prototype.

Keywords: organic Rankine cycle, regenerative organic Rankine cycle, waste heat recovery, Indian industry

Procedia PDF Downloads 355
163 Design of New Baby Food Product Using Whey

Authors: Henri El Zakhem, Anthony Dahdah, Lara Frangieh, Jessica Koura

Abstract:

Nowadays, the removal of whey produced in the dairy processes has been the most important problem in the dairy industry. Every year, about 47% of the 115 million tons of whey produced world-wide are disposed in the environment. Whey is a nutritious liquid, containing whey proteins (β-lactoglobulin, α-lactalbumin, immunoglobulin-G, proteose pepton), lactose, vitamins (B5, B2, C, and B6), minerals (Calcium, Magnesium, Phosphorous, Potassium, Chloride, and Sodium), and trace elements (Zinc, Iron, Iodine, and Copper). The first objective was to increase the economical and commercial value of whey which is considered as by-product. The second objective of this study was to formulate a new baby food with good nutritional, sensory and storage properties and acceptable to consumers using the cheese whey. The creation of the new product must pass through the following stages: idea stage, development stage which includes the business planning and the product development prototype, packaging stage, production stage, test marketing stage, quality control/sanitation. Three types of whey-based food were selected and prepared by mixing whey and apple, whey and banana as well as whey, apple, and banana.To compile with the recommended dietary allowances (RDA) and adequate intakes (AI) for vitamins and minerals, each sample is formed from 114g of sliced and smashed fruits mixed with 8 mL of whey. Mixtures are heated to 72oC for 15 seconds, and filled in pasteurized jars. Jars were conserved at 4oC. Following the experimental part, sensory evaluation made by an experienced panel took place. Hedonic tests results show that the mixture of whey, apple, and banana has the most delicious and sweetness taste followed by the mixture of whey and banana, and finally the mixture of whey and apple. This study was concluded with a managerial and engineering study that reveals that the project is economically profitable to be executed in Lebanon.

Keywords: baby food, by-product, cheese whey, formulation

Procedia PDF Downloads 253
162 STML: Service Type-Checking Markup Language for Services of Web Components

Authors: Saqib Rasool, Adnan N. Mian

Abstract:

Web components are introduced as the latest standard of HTML5 for writing modular web interfaces for ensuring maintainability through the isolated scope of web components. Reusability can also be achieved by sharing plug-and-play web components that can be used as off-the-shelf components by other developers. A web component encapsulates all the required HTML, CSS and JavaScript code as a standalone package which must be imported for integrating a web component within an existing web interface. It is then followed by the integration of web component with the web services for dynamically populating its content. Since web components are reusable as off-the-shelf components, these must be equipped with some mechanism for ensuring their proper integration with web services. The consistency of a service behavior can be verified through type-checking. This is one of the popular solutions for improving the quality of code in many programming languages. However, HTML does not provide type checking as it is a markup language and not a programming language. The contribution of this work is to introduce a new extension of HTML called Service Type-checking Markup Language (STML) for adding support of type checking in HTML for JSON based REST services. STML can be used for defining the expected data types of response from JSON based REST services which will be used for populating the content within HTML elements of a web component. Although JSON has five data types viz. string, number, boolean, object and array but STML is made to supports only string, number and object. This is because of the fact that both object and array are considered as string, when populated in HTML elements. In order to define the data type of any HTML element, developer just needs to add the custom STML attributes of st-string, st-number and st-boolean for string, number and boolean respectively. These all annotations of STML are used by the developer who is writing a web component and it enables the other developers to use automated type-checking for ensuring the proper integration of their REST services with the same web component. Two utilities have been written for developers who are using STML based web components. One of these utilities is used for automated type-checking during the development phase. It uses the browser console for showing the error description if integrated web service is not returning the response with expected data type. The other utility is a Gulp based command line utility for removing the STML attributes before going in production. This ensures the delivery of STML free web pages in the production environment. Both of these utilities have been tested to perform type checking of REST services through STML based web components and results have confirmed the feasibility of evaluating service behavior only through HTML. Currently, STML is designed for automated type-checking of integrated REST services but it can be extended to introduce a complete service testing suite based on HTML only, and it will transform STML from Service Type-checking Markup Language to Service Testing Markup Language.

Keywords: REST, STML, type checking, web component

Procedia PDF Downloads 233
161 A Benchmark System for Testing Medium Voltage Direct Current (MVDC-CB) Robustness Utilizing Real Time Digital Simulation and Hardware-In-Loop Theory

Authors: Ali Kadivar, Kaveh Niayesh

Abstract:

The integration of green energy resources is a major focus, and the role of Medium Voltage Direct Current (MVDC) systems is exponentially expanding. However, the protection of MVDC systems against DC faults is a challenge that can have consequences on reliable and safe grid operation. This challenge reveals the need for MVDC circuit breakers (MVDC CB), which are in infancies of their improvement. Therefore will be a lack of MVDC CBs standards, including thresholds for acceptable power losses and operation speed. To establish a baseline for comparison purposes, a benchmark system for testing future MVDC CBs is vital. The literatures just give the timing sequence of each switch and the emphasis is on the topology, without in-depth study on the control algorithm of DCCB, as the circuit breaker control system is not yet systematic. A digital testing benchmark is designed for the Proof-of-concept of simulation studies using software models. It can validate studies based on real-time digital simulators and Transient Network Analyzer (TNA) models. The proposed experimental setup utilizes data accusation from the accurate sensors installed on the tested MVDC CB and through general purpose input/outputs (GPIO) from the microcontroller and PC Prototype studies in the laboratory-based models utilizing Hardware-in-the-Loop (HIL) equipment connected to real-time digital simulators is achieved. The improved control algorithm of the circuit breaker can reduce the peak fault current and avoid arc resignation, helping the coordination of DCCB in relay protection. Moreover, several research gaps are identified regarding case studies and evaluation approaches.

Keywords: DC circuit breaker, hardware-in-the-loop, real time digital simulation, testing benchmark

Procedia PDF Downloads 54
160 Isolation Enhancement of Compact Dual-Band Printed Multiple Input Multiple Output Antenna for WLAN Applications

Authors: Adham M. Salah, Tariq A. Nagem, Raed A. Abd-Alhameed, James M. Noras

Abstract:

Recently, the demand for wireless communications systems to cover more than one frequency band (multi-band) with high data rate has been increased for both fixed and mobile services. Multiple Input Multiple Output (MIMO) technology is one of the significant solutions for attaining these requirements and to achieve the maximum channel capacity of the wireless communications systems. The main issue associated with MIMO antennas especially in portable devices is the compact space between the radiating elements which leads to limit the physical separation between them. This issue exacerbates the performance of the MIMO antennas by increasing the mutual coupling between the radiating elements. In other words, the mutual coupling will be stronger if the radiating elements of the MIMO antenna are closer. This paper presents a low–profile dual-band (2×1) MIMO antenna that works at 2.4GHz, 5.3GHz and 5.8GHz for wireless local area networks (WLAN) applications. A neutralization line (NL) technique for enhancing the isolation has been used by introducing a strip line with a length of λg/4 at the isolation frequency (2.4GHz) between the radiating elements. The overall dimensions of the antenna are 33.5 x 36 x 1.6 mm³. The fabricated prototype shows a good agreement between the simulated and measured results. The antenna impedance bandwidths are 2.38–2.75 GHz and 4.4–6 GHz for the lower and upper band respectively; the reflection coefficient and mutual coupling are better than -25 dB in both lower and higher bands. The MIMO antenna performance characteristics are reported in terms of the scattering parameters, envelope correlation coefficient (ECC), total active reflection coefficient, capacity loss, antenna gain, and radiation patterns. Analysis of these characteristics indicates that the design is appropriate for the WLAN terminal applications.

Keywords: ECC, neutralization line, MIMO antenna, multi-band, mutual coupling, WLAN

Procedia PDF Downloads 116
159 The Colouration of Additive-Manufactured Polymer

Authors: Abisuga Oluwayemisi Adebola, Kerri Akiwowo, Deon de Beer, Kobus Van Der Walt

Abstract:

The convergence of additive manufacturing (AM) and traditional textile dyeing techniques has initiated innovative possibilities for improving the visual application and customization potential of 3D-printed polymer objects. Textile dyeing techniques have progressed to transform fabrics with vibrant colours and complex patterns over centuries. The layer-by-layer deposition characteristic of AM necessitates adaptations in dye application methods to ensure even colour penetration across complex surfaces. Compatibility between dye formulations and polymer matrices influences colour uptake and stability, demanding careful selection and testing of dyes for optimal results. This study investigates the development interaction between these areas, revealing the challenges and opportunities of applying textile dyeing methods to colour 3D-printed polymer materials. The method explores three innovative approaches to colour the 3D-printed polymer object: (a) Additive Manufacturing of a Prototype, (b) the traditional dyebath method, and (c) the contemporary digital sublimation technique. The results show that the layer lines inherent to AM interact with dyes differently and affect the visual outcome compared to traditional textile fibers. Skillful manipulation of textile dyeing methods and dye type used for this research reduced the appearance of these lines to achieve consistency and desirable colour outcomes. In conclusion, integrating textile dyeing techniques into colouring 3D-printed polymer materials connects historical craftsmanship with innovative manufacturing. Overcoming challenges of colour distribution, compatibility, and layer line management requires a holistic approach that blends the technical consistency of AM with the artistic sensitivity of textile dyeing. Hence, applying textile dyeing methods to 3D-printed polymers opens new dimensions of aesthetic and functional possibilities.

Keywords: polymer, 3D-printing, sublimation, textile, dyeing, additive manufacturing

Procedia PDF Downloads 51
158 Exploring Affordable Care Practs in Nigeria’s Health Insurance Discourse

Authors: Emmanuel Chinaguh, Kehinde Adeosun

Abstract:

Nigerians die untimely, with 55.75 years of life expectancy, which is 17.45 below the world average of 73.2 (Worldometer, 2020). This is due, among other factors, to the country's limited access to high-quality healthcare. To increase access to good and affordable healthcare services, the National Health Insurance Authority (NHIA) Bill 2022 – which repealed the National Health Insurance Scheme Act 2004 – was passed into law. Applying Jacob Mey’s (2001) pragmatics act (pract) theory, this study explores how NHIA seeks to actualise these healthcare goals by characterising the general situational prototype or pragmemes and pragmatic acts in institutional communications. Data was sourced from the NHIA operational guidelines, which has 147 pages and four sections, and shared posters on NHIA Nigeria Twitter Handle with 14,200 followers. Digital humanities tools, like AntConc and Voyant, were engaged in the data analysis for text encoding and data visualisation. This study identifies these discourse tokens in the data: advertisement and programmes, standards and accreditation, records and information, and offences and penalties. Advertisement and programmes pract facilitating, propagating, prospecting, advising and informing; standards and accreditation, and records and information pract stating, informing and instructing; and offences and penalties pract stating and sanctioning. These practs combined to advance the goals of affordable care and universal accessibility to quality healthcare services. The pragmatic acts were marked by these pragmatic tools: shared situational knowledge (SSK), relevance (REL), reference (REF) and inference (INF). This paper adds to the understanding of health insurance discourse in Nigeria as a mediated social practice that promotes the health of Nigerians.

Keywords: affordable care, NHIA, Nigeria’s health insurance discourse, pragmatic acts.

Procedia PDF Downloads 61
157 Enhancing Institutional Roles and Managerial Instruments for Irrigation Modernization in Sudan: The Case of Gezira Scheme

Authors: Mohamed Ahmed Abdelmawla

Abstract:

Calling to achieve Millennium Development Goals (MDGs) engaged with agriculture, i.e. poverty alleviation targets, human resources involved in agricultural sectors with special emphasis on irrigation must receive wealth of practical experience and training. Increased food production, including staple food, is needed to overcome the present and future threats to food security. This should happen within a framework of sustainable management of natural resources, elimination of unsustainable methods of production and poverty reduction (i.e. axes of modernization). A didactic tool to confirm the task of wise and maximum utility is the best management and accurate measurement, as major requisites for modernization process. The key component to modernization as a warranted goal is adhering great attention to management and measurement issues via capacity building. As such, this paper stressed the issues of discharge management and measurement by Field Outlet Pipes (FOP) for selected ones within the Gezira Scheme, where randomly nine FOPs were selected as representative locations. These FOPs extended along the Gezira Main Canal at Kilo 57 areas in the South up to Kilo 194 in the North. The following steps were followed during the field data collection and measurements: For each selected FOP, a 90 v- notch thin plate weir was placed in such away that the water was directed to pass only through the notch. An optical survey level was used to measure the water head of the notch and FOP. Both calculated discharge rates as measured by the v – notch, denoted as [Qc], and the adopted discharges given by (MOIWR), denoted as [Qa], are tackled for the average of three replicated readings undertaken at each location. The study revealed that the FOP overestimates and sometimes underestimates the discharges. This is attributed to the fact that the original design specifications were not fulfilled or met at present conditions where water is allowed to flow day and night with high head fluctuation, knowing that the FOP is non modular structure, i.e. the flow depends on both levels upstream and downstream and confirmed by the results of this study. It is convenient and formative to quantify the discharge in FOP with weirs or Parshall flumes. Cropping calendar should be clearly determined and agreed upon before the beginning of the season in accordance and consistency with the Sudan Gezira Board (SGB) and Ministry of Irrigation and Water Resources. As such, the water indenting should be based on actual Crop Water Requirements (CWRs), not on rules of thumb (420 m3/feddan, irrespective of crop or time of season).

Keywords: management, measurement, MDGs, modernization

Procedia PDF Downloads 234
156 Application of Electrochromic Glazing for Reducing Peak Cooling Loads

Authors: Ranojoy Dutta

Abstract:

HVAC equipment capacity has a direct impact on occupant comfort and energy consumption of a building. Glazing gains, especially in buildings with high window area, can be a significant contributor to the total peak load on the HVAC system, leading to over-sized systems that mostly operate at poor part load efficiency. In addition, radiant temperature, which largely drives occupant comfort in glazed perimeter zones, is often not effectively controlled despite the HVAC being designed to meet the air temperature set-point. This is due to short wave solar radiation transmitted through windows, that is not sensed by the thermostat until much later when the thermal mass in the room releases the absorbed solar heat to the indoor air. The implication of this phenomenon is increased cooling energy despite poor occupant comfort. EC glazing can significantly eliminate direct solar transmission through windows, reducing both the space cooling loads for the building and improving comfort for occupants near glazing. This paper will review the exact mechanism of how EC glazing would reduce the peak load under design day conditions, leading to reduced cooling capacity vs regular high-performance glazing. Since glazing heat transfer only affects the sensible load, system sizing will be evaluated both with and without the availability of a DOAS to isolate the downsizing potential of the primary cooling equipment when outdoor air is conditioned separately. Given the dynamic nature of glazing gains due to the sun’s movement, effective peak load mitigation with EC requires an automated control system that can predict solar movement and radiation levels so that the right tint state with the appropriate SHGC is utilized at any given time for a given façade orientation. Such an automated EC product will be evaluated for a prototype commercial office model situated in four distinct climate zones.

Keywords: electrochromic glazing, peak sizing, thermal comfort, glazing load

Procedia PDF Downloads 110
155 Laboratory Analysis of Stormwater Runoff Hydraulic and Pollutant Removal Performance of Pervious Concrete Based on Seashell By-Products

Authors: Jean-Jacques Randrianarimanana, Nassim Sebaibi, Mohamed Boutouil

Abstract:

In order to solve problems associated with stormwater runoff in urban areas and their effects on natural and artificial water bodies, the integration of new technical solutions to the rainwater drainage becomes even more essential. Permeable pavement systems are one of the most widely used techniques. This paper presents a laboratory analysis of stormwater runoff hydraulic and pollutant removal performance of permeable pavement system using pervious pavements based on seashell products. The laboratory prototype is a square column of 25 cm of side and consists of the surface in pervious concrete, a bedding of 3 cm in height, a geotextile and a subbase layer of 50 cm in height. A series of constant simulated rain events using semi-synthetic runoff which varied in intensity and duration were carried out. The initial vertical saturated hydraulic conductivity of the entire pervious pavement system was 0.25 cm/s (148 L/m2/min). The hydraulic functioning was influenced by both the inlet flow rate value and the test duration. The total water losses including evaporation ranged between 9% to 20% for all hydraulic experiments. The temporal and vertical variability of the pollutant removal efficiency (PRE) of the system were studied for total suspended solids (TSS). The results showed that the PRE along the vertical profile was influenced by the size of the suspended solids, and the pervious paver has the highest capacity to trap pollutant than the other porous layers of the permeable pavement system after the geotextile. The TSS removal efficiency was about 80% for the entire system. The first-flush effect of TSS was observed, but it appeared only at the beginning (2 to 6 min) of the experiments. It has been shown that the PPS can capture first-flush. The project in which this study is integrated aims to contribute to both the valorization of shellfish waste and the sustainable management of rainwater.

Keywords: hydraulic, pervious concrete, pollutant removal efficiency, seashell by-products, stormwater runoff

Procedia PDF Downloads 196
154 High-Frequency Acoustic Microscopy Imaging of Pellet/Cladding Interface in Nuclear Fuel Rods

Authors: H. Saikouk, D. Laux, Emmanuel Le Clézio, B. Lacroix, K. Audic, R. Largenton, E. Federici, G. Despaux

Abstract:

Pressurized Water Reactor (PWR) fuel rods are made of ceramic pellets (e.g. UO2 or (U,Pu) O2) assembled in a zirconium cladding tube. By design, an initial gap exists between these two elements. During irradiation, they both undergo transformations leading progressively to the closure of this gap. A local and non destructive examination of the pellet/cladding interface could constitute a useful help to identify the zones where the two materials are in contact, particularly at high burnups when a strong chemical bonding occurs under nominal operating conditions in PWR fuel rods. The evolution of the pellet/cladding bonding during irradiation is also an area of interest. In this context, the Institute of Electronic and Systems (IES- UMR CNRS 5214), in collaboration with the Alternative Energies and Atomic Energy Commission (CEA), is developing a high frequency acoustic microscope adapted to the control and imaging of the pellet/cladding interface with high resolution. Because the geometrical, chemical and mechanical nature of the contact interface is neither axially nor radially homogeneous, 2D images of this interface need to be acquired via this ultrasonic system with a highly performing processing signal and by means of controlled displacement of the sample rod along both its axis and its circumference. Modeling the multi-layer system (water, cladding, fuel etc.) is necessary in this present study and aims to take into account all the parameters that have an influence on the resolution of the acquired images. The first prototype of this microscope and the first results of the visualization of the inner face of the cladding will be presented in a poster in order to highlight the potentials of the system, whose final objective is to be introduced in the existing bench MEGAFOX dedicated to the non-destructive examination of irradiated fuel rods at LECA-STAR facility in CEA-Cadarache.

Keywords: high-frequency acoustic microscopy, multi-layer model, non-destructive testing, nuclear fuel rod, pellet/cladding interface, signal processing

Procedia PDF Downloads 169
153 Role of Symbolism in the Journey towards Spirituality: A Case Study of Mosque Architecture in Bahrain

Authors: Ayesha Agha Shah

Abstract:

The purpose of a mosque or a place of worship is to build a spiritual relation with God. If the sense of spirituality is not achieved, then sacred architecture appears to be lacking depth. Form and space play a significant role to enhance the architectural quality to impart a divine feel to a place. To achieve this divine feeling, form and space, and unity of opposites, either abstract or symbolic can be employed. It is challenging to imbue the emptiness of a space with qualitative experience. Mosque architecture mostly entails traditional forms and design typology. This approach for Muslim worship produces distinct landmarks in the urban neighborhoods of Muslim societies, while creating a great sense of spirituality. The universal symbolic characters in the mosque architecture had prototype geometrical forms for a long time in history. However, modern mosques have deviated from this approach to employ different built elements and symbolism, which are often hard to be identified as related to mosques or even as Islamic. This research aims to explore the sense of spirituality in modern mosques and questions whether the modification of geometrical features produce spirituality in the same manner. The research also seeks to investigate the role of ‘geometry’ in the modern mosque architecture. The research employs the analytical study of some modern mosque examples in the Kingdom of Bahrain, reflecting on the geometry and symbolism adopted in the new mosque architecture design. It buttresses the analysis by the engagement of people’s perceptions derived using a survey of opinions. The research expects to see the significance of geometrical architectural elements in the mosque designs. It will find answers to the questions such as; what is the role of the form of the mosque, interior spaces and the effect of the modified symbolic features in the modern mosque design? How can the symbolic geometry, forms and spaces of a mosque invite a believer to leave the worldly environment behind and move towards spirituality?

Keywords: geometry, mosque architecture, spirituality, symbolism

Procedia PDF Downloads 93
152 Development of an Atmospheric Radioxenon Detection System for Nuclear Explosion Monitoring

Authors: V. Thomas, O. Delaune, W. Hennig, S. Hoover

Abstract:

Measurement of radioactive isotopes of atmospheric xenon is used to detect, locate and identify any confined nuclear tests as part of the Comprehensive Nuclear Test-Ban Treaty (CTBT). In this context, the Alternative Energies and French Atomic Energy Commission (CEA) has developed a fixed device to continuously measure the concentration of these fission products, the SPALAX process. During its atmospheric transport, the radioactive xenon will undergo a significant dilution between the source point and the measurement station. Regarding the distance between fixed stations located all over the globe, the typical volume activities measured are near 1 mBq m⁻³. To avoid the constraints induced by atmospheric dilution, the development of a mobile detection system is in progress; this system will allow on-site measurements in order to confirm or infringe a suspicious measurement detected by a fixed station. Furthermore, this system will use beta/gamma coincidence measurement technique in order to drastically reduce environmental background (which masks such activities). The detector prototype consists of a gas cell surrounded by two large silicon wafers, coupled with two square NaI(Tl) detectors. The gas cell has a sample volume of 30 cm³ and the silicon wafers are 500 µm thick with an active surface area of 3600 mm². In order to minimize leakage current, each wafer has been segmented into four independent silicon pixels. This cell is sandwiched between two low background NaI(Tl) detectors (70x70x40 mm³ crystal). The expected Minimal Detectable Concentration (MDC) for each radio-xenon is in the order of 1-10 mBq m⁻³. Three 4-channels digital acquisition modules (Pixie-NET) are used to process all the signals. Time synchronization is ensured by a dedicated PTP-network, using the IEEE 1588 Precision Time Protocol. We would like to present this system from its simulation to the laboratory tests.

Keywords: beta/gamma coincidence technique, low level measurement, radioxenon, silicon pixels

Procedia PDF Downloads 113
151 An Artificial Intelligence Framework to Forecast Air Quality

Authors: Richard Ren

Abstract:

Air pollution is a serious danger to international well-being and economies - it will kill an estimated 7 million people every year, costing world economies $2.6 trillion by 2060 due to sick days, healthcare costs, and reduced productivity. In the United States alone, 60,000 premature deaths are caused by poor air quality. For this reason, there is a crucial need to develop effective methods to forecast air quality, which can mitigate air pollution’s detrimental public health effects and associated costs by helping people plan ahead and avoid exposure. The goal of this study is to propose an artificial intelligence framework for predicting future air quality based on timing variables (i.e. season, weekday/weekend), future weather forecasts, as well as past pollutant and air quality measurements. The proposed framework utilizes multiple machine learning algorithms (logistic regression, random forest, neural network) with different specifications and averages the results of the three top-performing models to eliminate inaccuracies, weaknesses, and biases from any one individual model. Over time, the proposed framework uses new data to self-adjust model parameters and increase prediction accuracy. To demonstrate its applicability, a prototype of this framework was created to forecast air quality in Los Angeles, California using datasets from the RP4 weather data repository and EPA pollutant measurement data. The results showed good agreement between the framework’s predictions and real-life observations, with an overall 92% model accuracy. The combined model is able to predict more accurately than any of the individual models, and it is able to reliably forecast season-based variations in air quality levels. Top air quality predictor variables were identified through the measurement of mean decrease in accuracy. This study proposed and demonstrated the efficacy of a comprehensive air quality prediction framework leveraging multiple machine learning algorithms to overcome individual algorithm shortcomings. Future enhancements should focus on expanding and testing a greater variety of modeling techniques within the proposed framework, testing the framework in different locations, and developing a platform to automatically publish future predictions in the form of a web or mobile application. Accurate predictions from this artificial intelligence framework can in turn be used to save and improve lives by allowing individuals to protect their health and allowing governments to implement effective pollution control measures.Air pollution is a serious danger to international wellbeing and economies - it will kill an estimated 7 million people every year, costing world economies $2.6 trillion by 2060 due to sick days, healthcare costs, and reduced productivity. In the United States alone, 60,000 premature deaths are caused by poor air quality. For this reason, there is a crucial need to develop effective methods to forecast air quality, which can mitigate air pollution’s detrimental public health effects and associated costs by helping people plan ahead and avoid exposure. The goal of this study is to propose an artificial intelligence framework for predicting future air quality based on timing variables (i.e. season, weekday/weekend), future weather forecasts, as well as past pollutant and air quality measurements. The proposed framework utilizes multiple machine learning algorithms (logistic regression, random forest, neural network) with different specifications and averages the results of the three top-performing models to eliminate inaccuracies, weaknesses, and biases from any one individual model. Over time, the proposed framework uses new data to self-adjust model parameters and increase prediction accuracy. To demonstrate its applicability, a prototype of this framework was created to forecast air quality in Los Angeles, California using datasets from the RP4 weather data repository and EPA pollutant measurement data. The results showed good agreement between the framework’s predictions and real-life observations, with an overall 92% model accuracy. The combined model is able to predict more accurately than any of the individual models, and it is able to reliably forecast season-based variations in air quality levels. Top air quality predictor variables were identified through the measurement of mean decrease in accuracy. This study proposed and demonstrated the efficacy of a comprehensive air quality prediction framework leveraging multiple machine learning algorithms to overcome individual algorithm shortcomings. Future enhancements should focus on expanding and testing a greater variety of modeling techniques within the proposed framework, testing the framework in different locations, and developing a platform to automatically publish future predictions in the form of a web or mobile application. Accurate predictions from this artificial intelligence framework can in turn be used to save and improve lives by allowing individuals to protect their health and allowing governments to implement effective pollution control measures.Air pollution is a serious danger to international wellbeing and economies - it will kill an estimated 7 million people every year, costing world economies $2.6 trillion by 2060 due to sick days, healthcare costs, and reduced productivity. In the United States alone, 60,000 premature deaths are caused by poor air quality. For this reason, there is a crucial need to develop effective methods to forecast air quality, which can mitigate air pollution’s detrimental public health effects and associated costs by helping people plan ahead and avoid exposure. The goal of this study is to propose an artificial intelligence framework for predicting future air quality based on timing variables (i.e. season, weekday/weekend), future weather forecasts, as well as past pollutant and air quality measurements. The proposed framework utilizes multiple machine learning algorithms (logistic regression, random forest, neural network) with different specifications and averages the results of the three top-performing models to eliminate inaccuracies, weaknesses, and biases from any one individual model. Over time, the proposed framework uses new data to self-adjust model parameters and increase prediction accuracy. To demonstrate its applicability, a prototype of this framework was created to forecast air quality in Los Angeles, California using datasets from the RP4 weather data repository and EPA pollutant measurement data. The results showed good agreement between the framework’s predictions and real-life observations, with an overall 92% model accuracy. The combined model is able to predict more accurately than any of the individual models, and it is able to reliably forecast season-based variations in air quality levels. Top air quality predictor variables were identified through the measurement of mean decrease in accuracy. This study proposed and demonstrated the efficacy of a comprehensive air quality prediction framework leveraging multiple machine learning algorithms to overcome individual algorithm shortcomings. Future enhancements should focus on expanding and testing a greater variety of modeling techniques within the proposed framework, testing the framework in different locations, and developing a platform to automatically publish future predictions in the form of a web or mobile application. Accurate predictions from this artificial intelligence framework can in turn be used to save and improve lives by allowing individuals to protect their health and allowing governments to implement effective pollution control measures.

Keywords: air quality prediction, air pollution, artificial intelligence, machine learning algorithms

Procedia PDF Downloads 97
150 Evaluation of the Gasification Process for the Generation of Syngas Using Solid Waste at the Autónoma de Colombia University

Authors: Yeraldin Galindo, Soraida Mora

Abstract:

Solid urban waste represents one of the largest sources of global environmental pollution due to the large quantities of these that are produced every day; thus, the elimination of such waste is a major problem for the environmental authorities who must look for alternatives to reduce the volume of waste with the possibility of obtaining an energy recovery. At the Autónoma de Colombia University, approximately 423.27 kg/d of solid waste are generated mainly paper, cardboard, and plastic. A large amount of these solid wastes has as final disposition the sanitary landfill of the city, wasting the energy potential that these could have, this, added to the emissions generated by the collection and transport of the same, has as consequence the increase of atmospheric pollutants. One of the alternative process used in the last years to generate electrical energy from solid waste such as paper, cardboard, plastic and, mainly, organic waste or biomass to replace the use of fossil fuels is the gasification. This is a thermal conversion process of biomass. The objective of it is to generate a combustible gas as the result of a series of chemical reactions propitiated by the addition of heat and the reaction agents. This project was developed with the intention of giving an energetic use to the waste (paper, cardboard, and plastic) produced inside the university, using them to generate a synthesis gas with a gasifier prototype. The gas produced was evaluated to determine their benefits in terms of electricity generation or raw material for the chemical industry. In this process, air was used as gasifying agent. The characterization of the synthesis gas was carried out by a gas chromatography carried out by the Chemical Engineering Laboratory of the National University of Colombia. Taking into account the results obtained, it was concluded that the gas generated is of acceptable quality in terms of the concentration of its components, but it is a gas of low calorific value. For this reason, the syngas generated in this project is not viable for the production of electrical energy but for the production of methanol transformed by the Fischer-Tropsch cycle.

Keywords: alternative energies, gasification, gasifying agent, solid urban waste, syngas

Procedia PDF Downloads 236
149 Preliminary Composite Overwrapped Pressure Vessel Design for Hydrogen Storage Using Netting Analysis and American Society of Mechanical Engineers Section X

Authors: Natasha Botha, Gary Corderely, Helen M. Inglis

Abstract:

With the move to cleaner energy applications the transport industry is working towards on-board hydrogen, or compressed natural gas-fuelled vehicles. A popular method for storage is to use composite overwrapped pressure vessels (COPV) because of their high strength to weight ratios. The proper design of these COPVs are according to international standards; this study aims to provide a preliminary design for a 350 Bar Type IV COPV (i.e. a polymer liner with a composite overwrap). Netting analysis, a popular analytical approach, is used as a first step to generate an initial design concept for the composite winding. This design is further improved upon by following the American Society of Mechanical Engineers (ASME) Boiler and Pressure Vessel standards, Section X: Fibre-reinforced composite pressure vessels. A design program based on these two approaches is developed using Python. A numerical model of a burst test simulation is developed based on the two approaches and compared. The results indicate that the netting analysis provides a good preliminary design, while the ASME-based design is more robust and accurate as it includes a better approximation of the material behaviour. Netting analysis is an easy method to follow when considering an initial concept design for the composite winding when not all the material characteristics are known. Once these characteristics have been fully defined with experimental testing, an ASME-based design should always be followed to ensure that all designs conform to international standards and practices. Future work entails more detailed numerical testing of the design for improvement, this will include the boss design. Once finalised prototype manufacturing and experimental testing will be conducted, and the results used to improve on the COPV design.

Keywords: composite overwrapped pressure vessel, netting analysis, design, American Society of Mechanical Engineers section x, fiber-reinforced, hydrogen storage

Procedia PDF Downloads 218
148 Cost Effective Microfabrication Technique for Lab on Chip (LOC) Devices Using Epoxy Polymers

Authors: Charmi Chande, Ravindra Phadke

Abstract:

Microfluidics devices are fabricated by using multiple fabrication methods. Photolithography is one of the common methods wherein SU8 is widely used for making master which in turn is used for making working chip by the process of soft lithography. The high-aspect ratio features of SU-8 makes it suitable to be used as micro moulds for injection moulding, hot embossing, and moulds to form polydimethylsiloxane (PDMS) structures for bioMEMS (Microelectromechanical systems) applications. But due to high cost, difficulty in procuring and need for clean room, restricts the use of this polymer especially in developing countries and small research labs. ‘Bisphenol –A’ based polymers in mixture with curing agent are used in various industries like Paints and coatings, Adhesives, Electrical systems and electronics, Industrial tooling and composites. We present the novel use of ‘Bisphenol – A’ based polymer in fabricating micro channels for Lab On Chip(LOC) devices. The present paper describes the prototype for production of microfluidics chips using range of ‘Bisphenol-A’ based polymers viz. GY 250, ATUL B11, DER 331, DER 330 in mixture with cationic photo initiators. All the steps of chip production were carried out using an inexpensive approach that uses low cost chemicals and equipment. This even excludes the need of clean room. The produced chips using all above mentioned polymers were validated with respect to height and the chip giving least height was selected for further experimentation. The lowest height achieved was 7 micrometers by GY250. The cost of the master fabricated was $ 0.20 and working chip was $. 0.22. The best working chip was used for morphological identification and profiling of microorganisms from environmental samples like soil, marine water and salt water pan sites. The current chip can be adapted for various microbiological screening experiments like biochemical based microbial identification, studying uncultivable microorganisms at single cell/community level.

Keywords: bisphenol–A based epoxy, cationic photoinitiators, microfabrication, photolithography

Procedia PDF Downloads 267
147 Intentions and Willingness of Marketing Professionals to Adopt Neuromarketing

Authors: Anka Gorgiev, Chris Martin, Nikolaos Dimitriadis, Dimitrios V. Nikolaidis

Abstract:

This paper is part of a doctoral research study aimed to identify behavioral indicators for the existence of the new marketing paradigm. Neuromarketing is becoming a growing trend in the marketing industry worldwide and it is capturing a lot of interest among the members of academia and the practitioner community. However, it is still not very clear how big of an impact neuromarketing might have in the following years. In an effort to get closer to an answer, this study investigates behavioral intentions and willingness to adopt neuromarketing and its practices by the marketing professionals, including academics, practitioners, students, researchers, experts and journal editors. The participants in the study include marketing professionals at different levels of neuromarketing fluency with residency in the United States of America and the South East Europe. The total of 19 participants participated in the interviews, all of whom belong to more than one group of marketing professionals. The authors use qualitative research approach and open-ended interview questions specifically developed to assess ideas, beliefs and opinions that marketing professionals hold towards neuromarketing. In constructing the interview questions, the authors have used the theory of planned behavior, the prototype willingness model and the technology acceptance model as a theoretical framework. Previous studies have not explicitly investigated the behavioral intentions of marketing professionals to engage in neuromarketing behavior, which is described here as a tendency to apply neuromarketing assumptions and tools in usual marketing practices. This study suggests that the marketing professionals believe that neuromarketing can contribute to the business in a positive way and outlines the main advantages and disadvantages of adopting neuromarketing as identified by the participants. In addition, the study reveals an emerging image of an exemplar company that is perceived to be using neuromarketing, including the most common characteristics and attributes. These findings are believed to be crucial in facilitating a way for neuromarketing field to have a broader impact than it currently does by recognizing and understanding the limitations that such exemplars imply and how that has an effect on the decision-making of marketing professionals.

Keywords: behavioral intentions, marketing paradigm, neuromarketing adoption, theory of planned behavior

Procedia PDF Downloads 159
146 USBware: A Trusted and Multidisciplinary Framework for Enhanced Detection of USB-Based Attacks

Authors: Nir Nissim, Ran Yahalom, Tomer Lancewiki, Yuval Elovici, Boaz Lerner

Abstract:

Background: Attackers increasingly take advantage of innocent users who tend to use USB devices casually, assuming these devices benign when in fact they may carry an embedded malicious behavior or hidden malware. USB devices have many properties and capabilities that have become the subject of malicious operations. Many of the recent attacks targeting individuals, and especially organizations, utilize popular and widely used USB devices, such as mice, keyboards, flash drives, printers, and smartphones. However, current detection tools, techniques, and solutions generally fail to detect both the known and unknown attacks launched via USB devices. Significance: We propose USBWARE, a project that focuses on the vulnerabilities of USB devices and centers on the development of a comprehensive detection framework that relies upon a crucial attack repository. USBWARE will allow researchers and companies to better understand the vulnerabilities and attacks associated with USB devices as well as providing a comprehensive platform for developing detection solutions. Methodology: The framework of USBWARE is aimed at accurate detection of both known and unknown USB-based attacks by a process that efficiently enhances the framework's detection capabilities over time. The framework will integrate two main security approaches in order to enhance the detection of USB-based attacks associated with a variety of USB devices. The first approach is aimed at the detection of known attacks and their variants, whereas the second approach focuses on the detection of unknown attacks. USBWARE will consist of six independent but complimentary detection modules, each detecting attacks based on a different approach or discipline. These modules include novel ideas and algorithms inspired from or already developed within our team's domains of expertise, including cyber security, electrical and signal processing, machine learning, and computational biology. The establishment and maintenance of the USBWARE’s dynamic and up-to-date attack repository will strengthen the capabilities of the USBWARE detection framework. The attack repository’s infrastructure will enable researchers to record, document, create, and simulate existing and new USB-based attacks. This data will be used to maintain the detection framework’s updatability by incorporating knowledge regarding new attacks. Based on our experience in the cyber security domain, we aim to design the USBWARE framework so that it will have several characteristics that are crucial for this type of cyber-security detection solution. Specifically, the USBWARE framework should be: Novel, Multidisciplinary, Trusted, Lightweight, Extendable, Modular and Updatable and Adaptable. Major Findings: Based on our initial survey, we have already found more than 23 types of USB-based attacks, divided into six major categories. Our preliminary evaluation and proof of concepts showed that our detection modules can be used for efficient detection of several basic known USB attacks. Further research, development, and enhancements are required so that USBWARE will be capable to cover all of the major known USB attacks and to detect unknown attacks. Conclusion: USBWARE is a crucial detection framework that must be further enhanced and developed.

Keywords: USB, device, cyber security, attack, detection

Procedia PDF Downloads 376
145 Modeling of Conjugate Heat Transfer including Radiation in a Kerosene/Air Certification Burner

Authors: Lancelot Boulet, Pierre Benard, Ghislain Lartigue, Vincent Moureau, Nicolas Chauvet, Sheddia Didorally

Abstract:

International aeronautic standards demand a fire certification for engines that demonstrate their resistance. This demonstration relies on tests performed with prototype engines in the late stages of the development. Hardest tests require to place a kerosene standardized flame in front of the engine casing during a given time with imposed temperature and heat flux. The purpose of this work is to provide a better characterization of a kerosene/air certification burner in order to minimize the risks of test failure. A first Large-Eddy Simulation (LES) study of the certification burner permitted to model and simulate this burner, including both adiabatic and Conjugate Heat Transfer (CHT) computations. Carried out on unstructured grids with 40 million tetrahedral cells, using the finite-volume YALES2 code, spray combustion, forced convection on walls and conduction in the solid parts of the burner were coupled to achieve a detailed description of heat transfer. It highlighted the fact that conduction inside the solid has a real impact on the flame topology and the combustion regime. However, in the absence of radiative heat transfer, unrealistic temperature of the equipment was obtained. The aim of the present study is to include the radiative heat transfer in order to reach the same temperature given by experimental measurements. First, various test-cases are conducted to validate the coupling between the different heat solvers. Then, adiabatic case, CHT case, as well as CHT including radiative transfer are studied and compared. The LES model is finally applied to investigate the heat transfer in a flame impaction configuration. The aim is to progress on fire test modeling so as to reach a good confidence level as far as success of the certification test is concerned.

Keywords: conjugate heat transfer, fire resistance test, large-eddy simulation, radiative transfer, turbulent combustion

Procedia PDF Downloads 206
144 Dielectric Study of Ethanol Water Mixtures at Different Concentration Using Hollow Channel Cantilever Platform

Authors: Maryam S. Ghoraishi, John E. Hawk, Thomas Thundat

Abstract:

Understanding liquid properties in small scale has become important in recent decades as immerging new microelectromechanical systems (MEMS) devices have been widely used for micro pumps, drug delivery, and many other laboratory-on-microchips analysis. Often in microfluidic devices, fluids are transported electrokinetically. Therefore, extensive knowledge of fluid flow, heat transport, electrokinetics and electrochemistry are key to successful lab on a chip design. Among different microfluidic devices, recently developed hollow channel cantilever offers an ideal platform to study different fluid properties simultaneously without drastic decrease in quality factor which normally occurs when traditional cantilevers operate in the liquid phase. Using hollow channel cantilever, we monitor changes in density and viscosity of liquid while simultaneously investigating dielectric properties of alcohol water binary mixtures. Considerable research has been conducted on alcohol-water mixtures since such a mixture is a typical prototype for biomolecules, Micelle formation, and structural stability of proteins (to name a few). Here we show that hollow channel cantilever can be employed to investigate dielectric properties of ethanol/water mixtures in different concentrations. We study dynamic amplitude shifts of hollow channel cantilever oscillation at different concentrations of ethanol/water for different voltages. Our results show how interactions between solute and solvent, and possibly cluster formation, could change dielectric properties and dipole reorientation of the mixture, as well as the resulting force on the hollow cantilever. For comparison, we also examine higher conductivity ionic mixtures of sodium sulfate solution under the same conditions as low conductivity ethanol/water mixtures. We will show the results from systematic investigation of solvent effects on dielectric properties of the binary mixture. We will also address the question of resolution limits in dielectric study of analyte molecules imposed by solvent concentrations.

Keywords: dielectric constant, cantilever sensors, ethanol water mixtures, low frequency

Procedia PDF Downloads 179
143 An Enzyme Technology - Metnin™ - Enables the Full Replacement of Fossil-Based Polymers by Lignin in Polymeric Composites

Authors: Joana Antunes, Thomas Levée, Barbara Radovani, Anu Suonpää, Paulina Saloranta, Liji Sobhana, Petri Ihalainen

Abstract:

Lignin is an important component in the exploitation of lignocellulosic biomass. It has been shown that within the next years, the yield of added-value lignin-based chemicals and materials will generate renewable alternatives to oil-based products (e.g. polymeric composites, resins and adhesives) and enhance the economic feasibility of biorefineries. In this paper, a novel technology for lignin valorisation (METNIN™) is presented. METNIN™ is based on the oxidative action of an alkaliphilic enzyme in aqueous alkaline conditions (pH 10-11) at mild temperature (40-50 °C) combined with a cascading membrane operation, yielding a collection of lignin fractions (from oligomeric down to mixture of tri-, di- and monomeric units) with distinct molecular weight distribution, low polydispersity and favourable physicochemical properties. The alkaline process conditions ensure the high processibility of crude lignin in an aqueous environment and the efficiency of the enzyme, yielding better compatibility of lignin towards targeted applications. The application of a selected lignin fraction produced by METNIN™ as a suitable lignopolyol to completely replace a commercial polyol in polyurethane rigid foam formulations is presented as a prototype. Liquid lignopolyols with a high lignin content were prepared by oxypropylation and their full utilization in the polyurethane rigid foam formulation was successfully demonstrated. Moreover, selected technical specifications of different foam demonstrators were determined, including closed cell count, water uptake and compression characteristics. These specifications are within industrial standards for rigid foam applications. The lignin loading in the lignopolyol was a major factor determining the properties of the foam. In addition to polyurethane foam demonstrators, other examples of lignin-based products related to resins and sizing applications will be presented.

Keywords: enzyme, lignin valorisation, polyol, polyurethane foam

Procedia PDF Downloads 133
142 A Formal Microlectic Framework for Biological Circularchy

Authors: Ellis D. Cooper

Abstract:

“Circularchy” is supposed to be an adjustable formal framework with enough expressive power to articulate biological theory about Earthly Life in the sense of multi-scale biological autonomy constrained by non-equilibrium thermodynamics. “Formal framework” means specifically a multi-sorted first-order-theorywithequality (for each sort). Philosophically, such a theory is one kind of “microlect,” which means a “way of speaking” (or, more generally, a “way of behaving”) for overtly expressing a “mental model” of some “referent.” Other kinds of microlect include “natural microlect,” “diagrammatic microlect,” and “behavioral microlect,” with examples such as “political theory,” “Euclidean geometry,” and “dance choreography,” respectively. These are all describable in terms of a vocabulary conforming to grammar. As aspects of human culture, they are possibly reminiscent of Ernst Cassirer’s idea of “symbolic form;” as vocabularies, they are akin to Richard Rorty’s idea of “final vocabulary” for expressing a mental model of one’s life. A formal microlect is presented by stipulating sorts, variables, calculations, predicates, and postulates. Calculations (a.k.a., “terms”) may be composed to form more complicated calculations; predicates (a.k.a., “relations”) may be logically combined to form more complicated predicates; and statements (a.k.a., “sentences”) are grammatically correct expressions which are true or false. Conclusions are statements derived using logical rules of deduction from postulates, other assumed statements, or previously derived conclusions. A circularchy is a formal microlect constituted by two or more sub-microlects, each with its distinct stipulations of sorts, variables, calculations, predicates, and postulates. Within a sub-microlect some postulates or conclusions are equations which are statements that declare equality of specified calculations. An equational bond between an equation in one sub-microlect and an equation in either the same sub-microlect or in another sub-microlect is a predicate that declares equality of symbols occurring in a side of one equation with symbols occurring in a side of the other equation. Briefly, a circularchy is a network of equational bonds between sub-microlects. A circularchy is solvable if there exist solutions for all equations that satisfy all equational bonds. If a circularchy is not solvable, then a challenge would be to discover the obstruction to solvability and then conjecture what adjustments might remove the obstruction. Adjustment means changes in stipulated ingredients (sorts, etc.) of sub-microlects, or changes in equational bonds between sub-microlects, or introduction of new sub-microlects and new equational bonds. A circularchy is modular insofar as each sub-microlect is a node in a network of equation bonds. Solvability of a circularchy may be conjectured. Efforts to prove solvability may be thwarted by a counter-example or may lead to the construction of a solution. An automated theorem-proof assistant would likely be necessary for investigating a substantial circularchy, such as one purported to represent Earthly Life. Such investigations (chains of statements) would be concurrent with and no substitute for simulations (chains of numbers).

Keywords: autonomy, first-order theory, mathematics, thermodynamics

Procedia PDF Downloads 201
141 Infusing Social Business Skills into the Curriculum of Higher Learning Institutions with Special Reference to Albukhari International University

Authors: Abdi Omar Shuriye

Abstract:

A social business is a business designed to address socio-economic problems to enhance the welfare of the communities involved. Lately, social business, with its focus on innovative ideas, is capturing the interest of educational institutions, governments, and non-governmental organizations. Social business uses a business model to achieve a social goal, and in the last few decades, the idea of imbuing social business into the education system of higher learning institutions has spurred much excitement. This is due to the belief that it will lead to job creation and increased social resilience. One of the higher learning institutions which have invested immensely in the idea is Albukhari International University; it is a private education institution, on a state-of-the-art campus, providing an advantageous learning ecosystem. The niche area of this institution is social business, and it graduates job creators, not job seekers; this Malaysian institution is unique and one of its kind. The objective of this paper is to develop a work plan, direction, and milestone as well as the focus area for the infusion of social business into higher learning institutions with special reference to Al-Bukhari International University. The purpose is to develop a prototype and model full-scale to enable higher learning education institutions to construct the desired curriculum fermented with social business. With this model, major predicaments faced by these institutions could be overcome. The paper sets forth an educational plan and will spell out the basic tenets of social business, focusing on the nature and implementational aspects of the curriculum. It will also evaluate the mechanisms applied by these educational institutions. Currently, since research in this area remains scarce, institutions adopt the process of experimenting with various methods to find the best way to reach the desired result on the matter. The author is of the opinion that social business in education is the main tool to educate holistic future leaders; hence educational institutions should inspire students in the classroom to start up their own businesses by adopting creative and proactive teaching methods. This proposed model is a contribution in that direction.

Keywords: social business, curriculum, skills, university

Procedia PDF Downloads 69
140 Principles to Design Urbanism in Cinema; An Aesthetic Study on Identity and Representation of a City in a Movie

Authors: Dorsa Moayedi

Abstract:

‘The Cities’ and Cinema have a history going as far back as silent films; however, the standards of picturing a city in a film are somewhat vague. ‘Genius Loci’ of a city can be easily described with parameters that architects have detected; nevertheless, the genius loci of an ‘urban movie’ is untouched. Cities have been among the provocative matters that pushed filmmakers to ponder upon them and to picture them along with their urban identity thoroughly in their artworks, though the impacts of the urban life on the plot and characters is neglected, and so a city in a movie is usually restricted to ‘the place where the story happens’. Cities and urban life are among those that are in constant change and ongoing expansion; therefore, they are always fresh and ready to challenge people with their existence. Thus, the relationship between the city and cinema is metamorphic, though it could be defined and explored. The dominant research on the idea of urbanism has been conducted by outstanding scholars of architecture, like Christian Norberg-Schulz, and the studies on Cinema have been done by theorists of cinema, like Christian Metz, who have mastered defining their own realm; still, the idea to mingle the domains to reach a unified theory which could be applied to ‘urban movies’ is barely worked on. In this research, we have sought mutual grounds to discuss ‘urbanism in cinema,’ the grounds that cinema could benefit from and get to a more accurate audio-visual representation of a city, in accordance with the ideas of Christopher Alexander and the term he coined ‘The Timeless Way of Building.’ We concentrate on movies that are dependent on urban life, mainly those that possess the names of cities, like ‘Nashville (1975), Manhattan (1979), Fargo (1996), Midnight in Paris (2011) or Roma (2018), according to the ideas of urban design and narratives of cinema. Contrary to what has often been assumed, cinema and architecture could be defined in line with similar parameters, and architectural terms could be applied to the research done on movies. Our findings indicate that the theories of Christopher Alexander can best fit the paradigm to study an ‘Urban Movie’, definitions of a timeless building, elaborate on the characteristics of a design that could be applied to definitions of an urban movie, and set a prototype for further filmmaking regarding the urban life.

Keywords: city, urbanism, urban movies, identity, representation

Procedia PDF Downloads 41
139 Irradion: Portable Small Animal Imaging and Irradiation Unit

Authors: Josef Uher, Jana Boháčová, Richard Kadeřábek

Abstract:

In this paper, we present a multi-robot imaging and irradiation research platform referred to as Irradion, with full capabilities of portable arbitrary path computed tomography (CT). Irradion is an imaging and irradiation unit entirely based on robotic arms for research on cancer treatment with ion beams on small animals (mice or rats). The platform comprises two subsystems that combine several imaging modalities, such as 2D X-ray imaging, CT, and particle tracking, with precise positioning of a small animal for imaging and irradiation. Computed Tomography: The CT subsystem of the Irradion platform is equipped with two 6-joint robotic arms that position a photon counting detector and an X-ray tube independently and freely around the scanned specimen and allow image acquisition utilizing computed tomography. Irradiation measures nearly all conventional 2D and 3D trajectories of X-ray imaging with precisely calibrated and repeatable geometrical accuracy leading to a spatial resolution of up to 50 µm. In addition, the photon counting detectors allow X-ray photon energy discrimination, which can suppress scattered radiation, thus improving image contrast. It can also measure absorption spectra and recognize different materials (tissue) types. X-ray video recording and real-time imaging options can be applied for studies of dynamic processes, including in vivo specimens. Moreover, Irradion opens the door to exploring new 2D and 3D X-ray imaging approaches. We demonstrate in this publication various novel scan trajectories and their benefits. Proton Imaging and Particle Tracking: The Irradion platform allows combining several imaging modules with any required number of robots. The proton tracking module comprises another two robots, each holding particle tracking detectors with position, energy, and time-sensitive sensors Timepix3. Timepix3 detectors can track particles entering and exiting the specimen and allow accurate guiding of photon/ion beams for irradiation. In addition, quantifying the energy losses before and after the specimen brings essential information for precise irradiation planning and verification. Work on the small animal research platform Irradion involved advanced software and hardware development that will offer researchers a novel way to investigate new approaches in (i) radiotherapy, (ii) spectral CT, (iii) arbitrary path CT, (iv) particle tracking. The robotic platform for imaging and radiation research developed for the project is an entirely new product on the market. Preclinical research systems with precision robotic irradiation with photon/ion beams combined with multimodality high-resolution imaging do not exist currently. The researched technology can potentially cause a significant leap forward compared to the current, first-generation primary devices.

Keywords: arbitrary path CT, robotic CT, modular, multi-robot, small animal imaging

Procedia PDF Downloads 71