Search results for: fully homomorphic encryption (FHE)
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 1967

Search results for: fully homomorphic encryption (FHE)

1727 Fully Autonomous Vertical Farm to Increase Crop Production

Authors: Simone Cinquemani, Lorenzo Mantovani, Aleksander Dabek

Abstract:

New technologies in agriculture are opening new challenges and new opportunities. Among these, certainly, robotics, vision, and artificial intelligence are the ones that will make a significant leap, compared to traditional agricultural techniques, possible. In particular, the indoor farming sector will be the one that will benefit the most from these solutions. Vertical farming is a new field of research where mechanical engineering can bring knowledge and know-how to transform a highly labor-based business into a fully autonomous system. The aim of the research is to develop a multi-purpose, modular, and perfectly integrated platform for crop production in indoor vertical farming. Activities will be based both on hardware development such as automatic tools to perform different activities on soil and plants, as well as research to introduce an extensive use of monitoring techniques based on machine learning algorithms. This paper presents the preliminary results of a research project of a vertical farm living lab designed to (i) develop and test vertical farming cultivation practices, (ii) introduce a very high degree of mechanization and automation that makes all processes replicable, fully measurable, standardized and automated, (iii) develop a coordinated control and management environment for autonomous multiplatform or tele-operated robots in environments with the aim of carrying out complex tasks in the presence of environmental and cultivation constraints, (iv) integrate AI-based algorithms as decision support system to improve quality production. The coordinated management of multiplatform systems still presents innumerable challenges that require a strongly multidisciplinary approach right from the design, development, and implementation phases. The methodology is based on (i) the development of models capable of describing the dynamics of the various platforms and their interactions, (ii) the integrated design of mechatronic systems able to respond to the needs of the context and to exploit the strength characteristics highlighted by the models, (iii) implementation and experimental tests performed to test the real effectiveness of the systems created, evaluate any weaknesses so as to proceed with a targeted development. To these aims, a fully automated laboratory for growing plants in vertical farming has been developed and tested. The living lab makes extensive use of sensors to determine the overall state of the structure, crops, and systems used. The possibility of having specific measurements for each element involved in the cultivation process makes it possible to evaluate the effects of each variable of interest and allows for the creation of a robust model of the system as a whole. The automation of the laboratory is completed with the use of robots to carry out all the necessary operations, from sowing to handling to harvesting. These systems work synergistically thanks to the knowledge of detailed models developed based on the information collected, which allows for deepening the knowledge of these types of crops and guarantees the possibility of tracing every action performed on each single plant. To this end, artificial intelligence algorithms have been developed to allow synergistic operation of all systems.

Keywords: automation, vertical farming, robot, artificial intelligence, vision, control

Procedia PDF Downloads 38
1726 Integrating the Modbus SCADA Communication Protocol with Elliptic Curve Cryptography

Authors: Despoina Chochtoula, Aristidis Ilias, Yannis Stamatiou

Abstract:

Modbus is a protocol that enables the communication among devices which are connected to the same network. This protocol is, often, deployed in connecting sensor and monitoring units to central supervisory servers in Supervisory Control and Data Acquisition, or SCADA, systems. These systems monitor critical infrastructures, such as factories, power generation stations, nuclear power reactors etc. in order to detect malfunctions and ignite alerts and corrective actions. However, due to their criticality, SCADA systems are vulnerable to attacks that range from simple eavesdropping on operation parameters, exchanged messages, and valuable infrastructure information to malicious modification of vital infrastructure data towards infliction of damage. Thus, the SCADA research community has been active over strengthening SCADA systems with suitable data protection mechanisms based, to a large extend, on cryptographic methods for data encryption, device authentication, and message integrity protection. However, due to the limited computation power of many SCADA sensor and embedded devices, the usual public key cryptographic methods are not appropriate due to their high computational requirements. As an alternative, Elliptic Curve Cryptography has been proposed, which requires smaller key sizes and, thus, less demanding cryptographic operations. Until now, however, no such implementation has been proposed in the SCADA literature, to the best of our knowledge. In order to fill this gap, our methodology was focused on integrating Modbus, a frequently used SCADA communication protocol, with Elliptic Curve based cryptography and develop a server/client application to demonstrate the proof of concept. For the implementation we deployed two C language libraries, which were suitably modify in order to be successfully integrated: libmodbus (https://github.com/stephane/libmodbus) and ecc-lib https://www.ceid.upatras.gr/webpages/faculty/zaro/software/ecc-lib/). The first library provides a C implementation of the Modbus/TCP protocol while the second one offers the functionality to develop cryptographic protocols based on Elliptic Curve Cryptography. These two libraries were combined, after suitable modifications and enhancements, in order to give a modified version of the Modbus/TCP protocol focusing on the security of the data exchanged among the devices and the supervisory servers. The mechanisms we implemented include key generation, key exchange/sharing, message authentication, data integrity check, and encryption/decryption of data. The key generation and key exchange protocols were implemented with the use of Elliptic Curve Cryptography primitives. The keys established by each device are saved in their local memory and are retained during the whole communication session and are used in encrypting and decrypting exchanged messages as well as certifying entities and the integrity of the messages. Finally, the modified library was compiled for the Android environment in order to run the server application as an Android app. The client program runs on a regular computer. The communication between these two entities is an example of the successful establishment of an Elliptic Curve Cryptography based, secure Modbus wireless communication session between a portable device acting as a supervisor station and a monitoring computer. Our first performance measurements are, also, very promising and demonstrate the feasibility of embedding Elliptic Curve Cryptography into SCADA systems, filling in a gap in the relevant scientific literature.

Keywords: elliptic curve cryptography, ICT security, modbus protocol, SCADA, TCP/IP protocol

Procedia PDF Downloads 270
1725 The Scanning Vibrating Electrode Technique (SVET) as a Tool for Optimising a Printed Ni(OH)2 Electrode under Charge Conditions

Authors: C. F. Glover, J. Marinaccio, A. Barnes, I. Mabbett, G. Williams

Abstract:

The aim of the current study is to optimise formulations, in terms of charging efficiency, of a printed Ni(OH)2 precursor coating of a battery anode. Through the assessment of the current densities during charging, the efficiency of a range of formulations are compared. The Scanning vibrating electrode technique (SVET) is used extensively in the field of corrosion to measure area-averaged current densities of freely-corroding metal surfaces when fully immersed in electrolyte. Here, a Ni(OH)2 electrode is immersed in potassium hydroxide (30% w/v solution) electrolyte and charged using a range of applied currents. Samples are prepared whereby multiple coatings are applied to one substrate, separated by a non-conducting barrier, and charged using a constant current. With a known applied external current, electrode efficiencies can be calculated based on the current density outputs measured using SVET. When fully charged, a green Ni(OH)2 is oxidised to a black NiOOH surface. Distinct regions displaying high current density, and hence a faster oxidising reaction rate, are located using the SVET. This is confirmed by a darkening of the region upon transition to NiOOH. SVET is a highly effective tool for assessing homogeneity of electrodes during charge/discharge. This could prove particularly useful for electrodes where there are no visible surface appearance changes. Furthermore, a scanning Kelvin probe technique, traditionally used to assess underfilm delamination of organic coatings for the protection of metallic surfaces, is employed to study the change in phase of oxides, pre and post charging.

Keywords: battery, electrode, nickel hydroxide, SVET, printed

Procedia PDF Downloads 235
1724 Damping Optimal Design of Sandwich Beams Partially Covered with Damping Patches

Authors: Guerich Mohamed, Assaf Samir

Abstract:

The application of viscoelastic materials in the form of constrained layers in mechanical structures is an efficient and cost-effective technique for solving noise and vibration problems. This technique requires a design tool to select the best location, type, and thickness of the damping treatment. This paper presents a finite element model for the vibration of beams partially or fully covered with a constrained viscoelastic damping material. The model is based on Bernoulli-Euler theory for the faces and Timoshenko beam theory for the core. It uses four variables: the through-thickness constant deflection, the axial displacements of the faces, and the bending rotation of the beam. The sandwich beam finite element is compatible with the conventional C1 finite element for homogenous beams. To validate the proposed model, several free vibration analyses of fully or partially covered beams, with different locations of the damping patches and different percent coverage, are studied. The results show that the proposed approach can be used as an effective tool to study the influence of the location and treatment size on the natural frequencies and the associated modal loss factors. Then, a parametric study regarding the variation in the damping characteristics of partially covered beams has been conducted. In these studies, the effect of core shear modulus value, the effect of patch size variation, the thickness of constraining layer, and the core and the locations of the patches are considered. In partial coverage, the spatial distribution of additive damping by using viscoelastic material is as important as the thickness and material properties of the viscoelastic layer and the constraining layer. Indeed, to limit added mass and to attain maximum damping, the damping patches should be placed at optimum locations. These locations are often selected using the modal strain energy indicator. Following this approach, the damping patches are applied over regions of the base structure with the highest modal strain energy to target specific modes of vibration. In the present study, a more efficient indicator is proposed, which consists of placing the damping patches over regions of high energy dissipation through the viscoelastic layer of the fully covered sandwich beam. The presented approach is used in an optimization method to select the best location for the damping patches as well as the material thicknesses and material properties of the layers that will yield optimal damping with the minimum area of coverage.

Keywords: finite element model, damping treatment, viscoelastic materials, sandwich beam

Procedia PDF Downloads 146
1723 Enhancing VR Exposure Therapy for the Treatment of Phobias with the Use of Photorealistic VR Environments and Stimuli, and the Use of Tactile Feedback Suits and Responsive Systems

Authors: Vardan Melkonyan, Arman Azizyan, Astghik Boyajyan

Abstract:

Virtual reality (VR) exposure therapy is a form of cognitive-behavioral therapy that uses immersive virtual environments to expose individuals to the feared stimuli or situations that trigger their phobia. VR exposure therapy has become an increasingly popular treatment for phobias, including fear of heights, public speaking, and flying, due to its ability to provide a controlled and safe environment for individuals to confront their fears while also allowing therapists to tailor the virtual exposure to the specific needs and goals of each individual. It is also a cost-effective and accessible treatment option, as it can be delivered remotely and does not require the use of drugs. Overall, VR exposure therapy has the potential to be a valuable tool for therapists in the treatment of phobias. But current methods may be improved by incorporating advanced technology such as photorealistic VR environments, tactile feedback suits, and responsive systems. The aim of this study was to identify the most effective approach for enhancing VR exposure therapy for the treatment of phobias. Photorealistic VR environments and stimuli can greatly enhance the effectiveness of VR exposure therapy for the treatment of phobias. By creating immersive, realistic virtual environments that closely mimic the real-life situations that trigger phobia responses, patients are able to more fully engage in the therapeutic process and confront their fears in a controlled and safe manner. This can help to reduce the severity of phobia symptoms and increase treatment outcomes. The use of tactile feedback suits and responsive systems can further enhance the VR exposure therapy experience by adding a physical element to the virtual environment. These suits, which can mimic the sensations of touch, pressure, and movement, allow patients to fully immerse themselves in the virtual world and feel as if they are physically present in the situation. This can help to increase the realism of the virtual environment and make it more effective in reducing phobia symptoms. Additionally, responsive systems can be used to trigger specific events or responses within the virtual environment based on the patient's actions, providing a more interactive and personalized treatment experience. A comprehensive literature review was conducted, including studies on VR exposure therapy for phobias and the use of advanced technology to enhance the therapy. Results indicate that incorporating these enhancements may significantly increase the effectiveness of VR exposure therapy for phobias. Further research is needed to fully understand the potential of these enhancements and to determine the optimal combination and implementation.

Keywords: virtual reality, mental health, phobias, fears, treatment, photorealistic, immersive, phobia

Procedia PDF Downloads 88
1722 Teaching–Learning-Based Optimization: An Efficient Method for Chinese as a Second Language

Authors: Qi Wang

Abstract:

In the classroom, teachers have been trained to complete the target task within the limited lecture time, meanwhile learners need to receive a lot of new knowledge, however, most of the time the learners come without the proper pre-class preparation to efficiently take in the contents taught in class. Under this circumstance, teachers do have no time to check whether the learners fully understand the content or not, how the learners communicate in the different contexts, until teachers see the results when the learners are tested. In the past decade, the teaching of Chinese has taken a trend. Teaching focuses less on the use of proper grammatical terms/punctuation and is now placing a heavier focus on the materials from real life contexts. As a result, it has become a greater challenge to teachers, as this requires teachers to fully understand/prepare what they teach and explain the content with simple and understandable words to learners. On the other hand, the same challenge also applies to the learners, who come from different countries. As they have to use what they learnt, based on their personal understanding of the material to effectively communicate with others in the classroom, even in the contexts of a day to day communication. To reach this win-win stage, Feynman’s Technique plays a very important role. This practical report presents you how the Feynman’s Technique is applied into Chinese courses, both writing & oral, to motivate the learners to practice more on writing, reading and speaking in the past few years. Part 1, analysis of different teaching styles and different types of learners, to find the most efficient way to both teachers and learners. Part 2, based on the theory of Feynman’s Technique, how to let learners build the knowledge from knowing the name of something to knowing something, via different designed target tasks. Part 3. The outcomes show that Feynman’s Technique is the interaction of learning style and teaching style, the double-edged sword of Teaching & Learning Chinese as a Second Language.

Keywords: Chinese, Feynman’s technique, learners, teachers

Procedia PDF Downloads 153
1721 2D Convolutional Networks for Automatic Segmentation of Knee Cartilage in 3D MRI

Authors: Ananya Ananya, Karthik Rao

Abstract:

Accurate segmentation of knee cartilage in 3-D magnetic resonance (MR) images for quantitative assessment of volume is crucial for studying and diagnosing osteoarthritis (OA) of the knee, one of the major causes of disability in elderly people. Radiologists generally perform this task in slice-by-slice manner taking 15-20 minutes per 3D image, and lead to high inter and intra observer variability. Hence automatic methods for knee cartilage segmentation are desirable and are an active field of research. This paper presents design and experimental evaluation of 2D convolutional neural networks based fully automated methods for knee cartilage segmentation in 3D MRI. The architectures are validated based on 40 test images and 60 training images from SKI10 dataset. The proposed methods segment 2D slices one by one, which are then combined to give segmentation for whole 3D images. Proposed methods are modified versions of U-net and dilated convolutions, consisting of a single step that segments the given image to 5 labels: background, femoral cartilage, tibia cartilage, femoral bone and tibia bone; cartilages being the primary components of interest. U-net consists of a contracting path and an expanding path, to capture context and localization respectively. Dilated convolutions lead to an exponential expansion of receptive field with only a linear increase in a number of parameters. A combination of modified U-net and dilated convolutions has also been explored. These architectures segment one 3D image in 8 – 10 seconds giving average volumetric Dice Score Coefficients (DSC) of 0.950 - 0.962 for femoral cartilage and 0.951 - 0.966 for tibia cartilage, reference being the manual segmentation.

Keywords: convolutional neural networks, dilated convolutions, 3 dimensional, fully automated, knee cartilage, MRI, segmentation, U-net

Procedia PDF Downloads 259
1720 Self-Supervised Attributed Graph Clustering with Dual Contrastive Loss Constraints

Authors: Lijuan Zhou, Mengqi Wu, Changyong Niu

Abstract:

Attributed graph clustering can utilize the graph topology and node attributes to uncover hidden community structures and patterns in complex networks, aiding in the understanding and analysis of complex systems. Utilizing contrastive learning for attributed graph clustering can effectively exploit meaningful implicit relationships between data. However, existing attributed graph clustering methods based on contrastive learning suffer from the following drawbacks: 1) Complex data augmentation increases computational cost, and inappropriate data augmentation may lead to semantic drift. 2) The selection of positive and negative samples neglects the intrinsic cluster structure learned from graph topology and node attributes. Therefore, this paper proposes a method called self-supervised Attributed Graph Clustering with Dual Contrastive Loss constraints (AGC-DCL). Firstly, Siamese Multilayer Perceptron (MLP) encoders are employed to generate two views separately to avoid complex data augmentation. Secondly, the neighborhood contrastive loss is introduced to constrain node representation using local topological structure while effectively embedding attribute information through attribute reconstruction. Additionally, clustering-oriented contrastive loss is applied to fully utilize clustering information in global semantics for discriminative node representations, regarding the cluster centers from two views as negative samples to fully leverage effective clustering information from different views. Comparative clustering results with existing attributed graph clustering algorithms on six datasets demonstrate the superiority of the proposed method.

Keywords: attributed graph clustering, contrastive learning, clustering-oriented, self-supervised learning

Procedia PDF Downloads 51
1719 A Coupled Stiffened Skin-Rib Fully Gradient Based Optimization Approach for a Wing Box Made of Blended Composite Materials

Authors: F. Farzan Nasab, H. J. M. Geijselaers, I. Baran, A. De Boer

Abstract:

A method is introduced for the coupled skin-rib optimization of a wing box where mass minimization is the objective and local buckling is the constraint. The structure is made of composite materials where continuity of plies in multiple adjacent panels (blending) has to be satisfied. Blending guarantees the manufacturability of the structure; however, it is a highly challenging constraint to treat and has been under debate in recent research in the same area. To fulfill design guidelines with respect to symmetry, balance, contiguity, disorientation and percentage rule of the layup, a reference for the stacking sequences (stacking sequence table or SST) is generated first. Then, an innovative fully gradient-based optimization approach in relation to a specific SST is introduced to obtain the optimum thickness distribution all over the structure while blending is fulfilled. The proposed optimization approach aims to turn the discrete optimization problem associated with the integer number of plies into a continuous one. As a result of a wing box deflection, a rib is subjected to load values which vary nonlinearly with the amount of deflection. The bending stiffness of a skin affects the wing box deflection and thus affects the load applied to a rib. This indicates the necessity of a coupled skin-rib optimization approach for a more realistic optimized design. The proposed method is examined with the optimization of the layup of a composite stiffened skin and rib of a wing torsion box subjected to in-plane normal and shear loads. Results show that the method can successfully prescribe a valid design with a significantly cheap computation cost.

Keywords: blending, buckling optimization, composite panels, wing torsion box

Procedia PDF Downloads 407
1718 Innovative Technologies for Aeration and Feeding of Fish in Aquaculture with Minimal Impact on the Environment

Authors: Vasile Caunii, Andreea D. Serban, Mihaela Ivancia

Abstract:

The paper presents a new approach in terms of the circular economy of technologies for feeding and aeration of accumulations and water basins for fish farming and aquaculture. Because fish is and will be one of the main foods on the planet, the use of bio-eco-technologies is a priority for all producers. The technologies proposed in the paper want to reduce by a substantial percentage the costs of operation of ponds and water accumulation, using non-polluting technologies with minimal impact on the environment. The paper proposes two innovative, intelligent systems, fully automated that use a common platform, completely eco-friendly. One system is intended to aerate the water of the fish pond, and the second is intended to feed the fish by dispersing an optimal amount of fodder, depending on population size, age and habits. Both systems use a floating platform, regenerative energy sources, are equipped with intelligent and innovative systems, and in addition to fully automated operation, significantly reduce the costs of aerating water accumulations (natural or artificial) and feeding fish. The intelligent system used for feeding, in addition, to reduce operating costs, optimizes the amount of food, thus preventing water pollution and the development of bacteria, microorganisms. The advantages of the systems are: increasing the yield of fish production, these are green installations, with zero pollutant emissions, can be arranged anywhere on the water surface, depending on the user's needs, can operate autonomously or remotely controlled, if there is a component failure, the system provides the operator with accurate data on the issue, significantly reducing maintenance costs, transmit data about the water physical and chemical parameters.

Keywords: bio-eco-technologies, economy, environment, fish

Procedia PDF Downloads 148
1717 Finite Element Model to Investigate the Dynamic Behavior of Ring-Stiffened Conical Shell Fully and Partially Filled with Fluid

Authors: Mohammadamin Esmaeilzadehazimi, Morteza Shayan Arani, Mohammad Toorani, Aouni Lakis

Abstract:

This study uses a hybrid finite element method to predict the dynamic behavior of both fully and partially-filled truncated conical shells stiffened with ring stiffeners. The method combines classical shell theory and the finite element method, and employs displacement functions derived from exact solutions of Sanders' shell equilibrium equations for conical shells. The shell-fluid interface is analyzed by utilizing the velocity potential, Bernoulli's equation, and impermeability conditions to determine an explicit expression for fluid pressure. The equations of motion presented in this study apply to both conical and cylindrical shells. This study presents the first comparison of the method applied to ring-stiffened shells with other numerical and experimental findings. Vibration frequencies for conical shells with various boundary conditions and geometries in a vacuum and filled with water are compared with experimental and numerical investigations, achieving good agreement. The study thoroughly investigates the influence of geometric parameters, stiffener quantity, semi-vertex cone angle, level of water filled in the cone, and applied boundary conditions on the natural frequency of fluid-loaded ring-stiffened conical shells, and draws some useful conclusions. The primary advantage of the current method is its use of a minimal number of finite elements while achieving highly accurate results.

Keywords: finite element method, fluid–structure interaction, conical shell, natural frequency, ring-stiffener

Procedia PDF Downloads 77
1716 Existential and Possessive Constructions in Modern Standard Arabic Two Strategies Reflecting the Ontological (Non-)Autonomy of Located or Possessed Entities

Authors: Fayssal Tayalati

Abstract:

Although languages use very divergent constructional strategies, all existential constructions appear to invariably involve an implicit or explicit locative constituent. This locative constituent either surface as a true locative phrase or are realized as a possessor noun phrase. However, while much research focuses on the supposed underlying syntactic relation of locative and possessive existential constructions, not much is known about possible semantic factors that could govern the choice between these constructions. The main question that we address in this talk concerns the choice between the two related constructions in Modern Standard Arabic (MAS). Although both are used to express the existence of something somewhere, we can distinguish three contexts: First, for some types of entities, only the EL construction is possible (e.g. (1a) ṯammata raǧulun fī l-ḥadīqati vs. (1b) *(kāna) ladā l-ḥadīqati raǧulun). Second, for other types of entities, only the possessive construction is possible (e.g. (2a) ladā ṭ-ṭawilati aklun dāʾiriyyun vs. (2b) *ṯammata šaklun dāʾiriyyun ladā/fī ṭ-ṭawilati). Finally, for still other entities, both constructions can be found (e.g. (3a) ṯammata ḥubbun lā yūṣafu ladā ǧārī li-zawǧati-hi and (3b) ladā ǧārī ḥubbun lā yūṣafu li-zawǧati-hi). The data covering a range of ontologically different entities (concrete objects, events, body parts, dimensions, essential qualities, feelings, etc.) shows that the choice between the existential locative and the possessive constructions is closely linked to the conceptual autonomy of the existential theme with respect to its location or to the whole that it is a part of. The construction with ṯammata is the only possible one to express the existence of a fully autonomous (i.e. nondependent) entity (concrete objects (e.g.1) and abstract objects such as events, especially the ones that Grimshaw called ‘simple events’). The possessive construction with (kāna) ladā is the only one used to express the existence of fully non-autonomous (i.e. fully dependent on a whole) entities (body parts, dimensions (e.g. 2), essential qualities). The two constructions alternate when the existential theme is conceptually dependent but separable of the whole, either because it has an autonomous (independent) existence of the given whole (spare parts of an object), or because it receives a relative autonomy in the speech through a modifier (accidental qualities, feelings (e.g. 3a, 3b), psychological states, among some other kinds of themes). In this case, the modifier expresses an approximate boundary on a scale, and provides relative autonomy to the entity. Finally, we will show that kinship terms (e.g. son), which at first sight may seem to constitute counterexamples to our hypothesis, are nonetheless supported by it. The ontological (non-)autonomy of located or possessed entities is also reflected by morpho-syntactic properties, among them the use and the choice of determiners, pluralisation and the behavior of entities in the context of associative anaphora.

Keywords: existence, possession, autonomous entities, non-autonomous entities

Procedia PDF Downloads 348
1715 A Two-Phase Flow Interface Tracking Algorithm Using a Fully Coupled Pressure-Based Finite Volume Method

Authors: Shidvash Vakilipour, Scott Ormiston, Masoud Mohammadi, Rouzbeh Riazi, Kimia Amiri, Sahar Barati

Abstract:

Two-phase and multi-phase flows are common flow types in fluid mechanics engineering. Among the basic and applied problems of these flow types, two-phase parallel flow is the one that two immiscible fluids flow in the vicinity of each other. In this type of flow, fluid properties (e.g. density, viscosity, and temperature) are different at the two sides of the interface of the two fluids. The most challenging part of the numerical simulation of two-phase flow is to determine the location of interface accurately. In the present work, a coupled interface tracking algorithm is developed based on Arbitrary Lagrangian-Eulerian (ALE) approach using a cell-centered, pressure-based, coupled solver. To validate this algorithm, an analytical solution for fully developed two-phase flow in presence of gravity is derived, and then, the results of the numerical simulation of this flow are compared with analytical solution at various flow conditions. The results of the simulations show good accuracy of the algorithm despite using a nearly coarse and uniform grid. Temporal variations of interface profile toward the steady-state solution show that a greater difference between fluids properties (especially dynamic viscosity) will result in larger traveling waves. Gravity effect studies also show that favorable gravity will result in a reduction of heavier fluid thickness and adverse gravity leads to increasing it with respect to the zero gravity condition. However, the magnitude of variation in favorable gravity is much more than adverse gravity.

Keywords: coupled solver, gravitational force, interface tracking, Reynolds number to Froude number, two-phase flow

Procedia PDF Downloads 314
1714 Producing of Amorphous-Nanocrystalline Composite Powders

Authors: K. Tomolya, D. Janovszky, A. Sycheva, M. Sveda, A. Roosz

Abstract:

CuZrAl amorphous alloys have attracted high interest due to unique physical and mechanical properties, which can be enhanced by adding of Ni and Ti elements. It is known that this properties can be enhanced by crystallization of amorphous alloys creating nanocrystallines in the matrix. The present work intends to produce nanosized crystalline parti-cle reinforced amorphous matrix composite powders by crystallization of amorphous powders. As the first step the amorphous powders were synthe-tized by ball-milling of crystalline powders. (Cu49Zr45Al6) 80Ni10Ti10 and (Cu49Zr44Al7) 80Ni10Ti10 (at%) alloys were ball-milled for 12 hours in order to reach the fully amorphous structure. The impact en-ergy of the balls during milling causes the change of the structure in the powders. Scanning electron microscopical (SEM) images shows that the phases mixed first and then changed into a fully amorphous matrix. Furthermore, nanosized particles in the amorphous matrix were crystallized by heat treatment of the amorphous powders that was confirmed by TEM measurement. It was of importance to define the tem-perature when the amorphous phase starts to crystal-lize. Amorphous alloys have a special heating curve and characteristic temperatures, which can be meas-ured by differential scanning calorimetry (DSC). A typical DSC curve of an amorphous alloy exhibits an endothermic event characteristic of the equilibrium glass transition (Tg) and a distinct undercooled liquid region, followed by one or two exothermic events corresponding to crystallization processes (Tp). After measuring the DSC traces of the amorphous powders, the annealing temperatures should be determined between Tx and Tp. In our experiments several temperatures from the annealing temperature range were selected and de-pendency of crystallized nanoparticles fraction on their hardness was investigated.

Keywords: amorphous structure, composite, mechanical milling, powder, scanning electron microscopy (SEM), differential scanning calorimetry (DSC), transmission electronmocroscopy (TEM)

Procedia PDF Downloads 448
1713 A Biometric Template Security Approach to Fingerprints Based on Polynomial Transformations

Authors: Ramon Santana

Abstract:

The use of biometric identifiers in the field of information security, access control to resources, authentication in ATMs and banking among others, are of great concern because of the safety of biometric data. In the general architecture of a biometric system have been detected eight vulnerabilities, six of them allow obtaining minutiae template in plain text. The main consequence of obtaining minutia templates is the loss of biometric identifier for life. To mitigate these vulnerabilities several models to protect minutiae templates have been proposed. Several vulnerabilities in the cryptographic security of these models allow to obtain biometric data in plain text. In order to increase the cryptographic security and ease of reversibility, a minutiae templates protection model is proposed. The model aims to make the cryptographic protection and facilitate the reversibility of data using two levels of security. The first level of security is the data transformation level. In this level generates invariant data to rotation and translation, further transformation is irreversible. The second level of security is the evaluation level, where the encryption key is generated and data is evaluated using a defined evaluation function. The model is aimed at mitigating known vulnerabilities of the proposed models, basing its security on the impossibility of the polynomial reconstruction.

Keywords: fingerprint, template protection, bio-cryptography, minutiae protection

Procedia PDF Downloads 168
1712 Search for APN Permutations in Rings ℤ_2×ℤ_2^k

Authors: Daniel Panario, Daniel Santana de Freitas, Brett Stevens

Abstract:

Almost Perfect Nonlinear (APN) permutations with optimal resistance against differential cryptanalysis can be found in several domains. The permutation used in the standard for symmetric cryptography (the AES), for example, is based on a special kind of inversion in GF(28). Although very close to APN (2-uniform), this permutation still contains one number 4 in its differential spectrum, which means that, rigorously, it must be classified as 4-uniform. This fact motivates the search for fully APN permutations in other domains of definition. The extremely high complexity associated to this kind of problem precludes an exhaustive search for an APN permutation with 256 elements to be performed without the support of a suitable mathematical structure. On the other hand, in principle, there is nothing to indicate which mathematically structured domains can effectively help the search, and it is necessary to test several domains. In this work, the search for APN permutations in rings ℤ2×ℤ2k is investigated. After a full, exhaustive search with k=2 and k=3, all possible APN permutations in those rings were recorded, together with their differential profiles. Some very promising heuristics in these cases were collected so that, when used as a basis to prune backtracking for the same search in ℤ2×ℤ8 (search space with size 16! ≅244), just a few tenths of a second were enough to produce an APN permutation in a single CPU. Those heuristics were empirically extrapolated so that they could be applied to a backtracking search for APNs over ℤ2×ℤ16 (search space with size 32! ≅2117). The best permutations found in this search were further refined through Simulated Annealing, with a definition of neighbors suitable to this domain. The best result produced with this scheme was a 3-uniform permutation over ℤ2×ℤ16 with only 24 values equal to 3 in the differential spectrum (all the other 968 values were less than or equal 2, as it should be the case for an APN permutation). Although far from being fully APN, this result is technically better than a 4-uniform permutation and demanded only a few seconds in a single CPU. This is a strong indication that the use of mathematically structured domains, like the rings described in this work, together with heuristics based on smaller cases, can lead to dramatic cuts in the computational resources involved in the complexity of the search for APN permutations in extremely large domains.

Keywords: APN permutations, heuristic searches, symmetric cryptography, S-box design

Procedia PDF Downloads 156
1711 Fused Deposition Modelling as the Manufacturing Method of Fully Bio-Based Water Purification Filters

Authors: Natalia Fijol, Aji P. Mathew

Abstract:

We present the processing and characterisation of three-dimensional (3D) monolith filters based on polylactic acid (PLA) reinforced with various nature-derived nanospecies such as hydroxyapatite, modified cellulose fibers and chitin fibers. The nanospecies of choice were dispersed in PLA through Thermally Induced Phase Separation (TIPS) method. The biocomposites were developed via solvent-assisted blending and the obtained pellets were further single-screw extruded into 3D-printing filaments and processed into various geometries using Fused Deposition Modelling (FDM) technique. The printed prototypes included cubic, cylindrical and hour-glass shapes with diverse patterns of printing infill as well as varying pore structure including uniform and multiple level gradual pore structure. The pores and channel structure as well as overall shape of the prototypes were designed in attempt to optimize the flux and maximize the adsorption-active time. FDM is a cost and energy-efficient method, which does not require expensive tools and elaborated post-processing maintenance. Therefore, FDM offers the possibility to produce customized, highly functional water purification filters with tuned porous structures suitable for removal of wide range of common water pollutants. Moreover, as 3D printing becomes more and more available worldwide, it allows producing portable filters at the place and time where they are most needed. The study demonstrates preparation route for the PLA-based, fully biobased composite and their processing via FDM technique into water purification filters, addressing water treatment challenges on an industrial scale.

Keywords: fused deposition modelling, water treatment, biomaterials, 3D printing, nanocellulose, nanochitin, polylactic acid

Procedia PDF Downloads 114
1710 HIV/AIDS Family Dysfunction Trajectories, Child Abuse and Psychosocial Problems among Adolescents

Authors: Paul Narh Doku

Abstract:

The relationship between parental HIV/AIDS status or death and child mental health is well known, although the role of child maltreatment as a confounder or mediator in this relationship remains uncertain. This study examined the potential path mechanism through child maltreatment mediating the link between HIV/AIDS family dysfunction trajectories and psychosocial problems. A cross-sectional survey was conducted in the Lower Manya Municipal Assembly of Ghana. A questionnaire which consisted of the Strengths and Difficulties Questionnaire (SDQ), Social and Health Assessment (SAHA), Rosenberg Self-Esteem Scale (RSES), and the Conflict Tactics Scale (CTS) was completed by 291 adolescents. Controlling for relevant sociodemographic confounders, mediation analyses using linear regression were fitted to examine whether the association between family dysfunction and psychosocial problems is mediated by child maltreatment. The results indicate that, among adolescents, child maltreatment fully mediated the association between being orphaned by AIDS and self-esteem, delinquency and risky behaviours, and peer problems. Similarly, child maltreatment fully mediated the association between living with an HIV/AIDS-infected parent and self-esteem, delinquency and risky behaviours, depression/emotional problems, and peer problems. Partial mediation was found for hyperactivity. Child maltreatment mediates the association between the family dysfunction trajectories of parental HIV/AIDS or death and psychosocial problems among adolescents. This implies that efforts to address child maltreatment among families affected by HIV/AIDS may be helpful in the prevention of psychosocial problems among these children, thus enhancing their well-being. The findings, therefore, underscore the need for comprehensive psychosocial interventions that address both the unique negative exposures of HIV/AIDS and maltreatment for children affected by HIV.

Keywords: child maltreatment, child abuse, mental health, psychosocial problems, domestic violence, HIV/AIDS, adolescents

Procedia PDF Downloads 80
1709 Extracting an Experimental Relation between SMD, Mass Flow Rate, Velocity and Pressure in Swirl Fuel Atomizers

Authors: Mohammad Hassan Ziraksaz

Abstract:

Fuel atomizers are used in a wide range of IC engines, turbojets and a variety of liquid propellant rocket engines. As the fuel spray fully develops its characters approach their ultimate amounts. Fuel spray characters such as SMD, injection pressure, mass flow rate, droplet velocity and spray cone angle play important roles to atomize the liquid fuel to finely atomized fuel droplets and finally form the fine fuel spray. Well performed, fully developed, fine spray without any defections, brings the idea of finding an experimental relation between the main effective spray characters. Extracting an experimental relation between SMD and other fuel spray physical characters in swirl fuel atomizers is the main scope of this experimental work. Droplet velocity, fuel mass flow rate, SMD and spray cone angle are the parameters which are measured. A set of twelve reverse engineering atomizers without any spray defections and a set of eight original atomizers as referenced well-performed spray are contributed in this work. More than 350 tests, mostly repeated, were performed. This work shows that although spray cone angle plays a very effective role in spray formation, after formation, it smoothly approaches to an almost constant amount while the other characters are changed to create fine droplets. Therefore, the work to find the relation between the characters is focused on SMD, droplet velocity, fuel mass flow rate, and injection pressure. The process of fuel spray formation begins in 5 Psig injection pressures, where a tiny fuel onion attaches to the injector tip and ended in 250 Psig injection pressure, were fully developed fine fuel spray forms. Injection pressure is gradually increased to observe how the spray forms. In each step, all parameters are measured and recorded carefully to provide a data bank. Various diagrams have been drawn to study the behavior of the parameters in more detail. Experiments and graphs show that the power equation can best show changes in parameters. The SMD experimental relation with pressure P, fuel mass flow rate Q ̇ and droplet velocity V extracted individually in pairs. Therefore, the proportional relation of SMD with other parameters is founded. Now it is time to find an experimental relation including all the parameters. Using obtained proportional relation, replacing the parameters with experimentally measured ones and drawing the graphs of experimental SMD versus proportion SMD (〖SMD〗_P), a correctional equation and consequently the final experimental equation is obtained. This experimental equation is specified to use for swirl fuel atomizers and the use of this experimental equation in different conditions shows about 3% error, which is expected to achieve lower error and consequently higher accuracy by increasing the number of experiments and increasing the accuracy of data collection.

Keywords: droplet velocity, experimental relation, mass flow rate, SMD, swirl fuel atomizer

Procedia PDF Downloads 160
1708 Disaster Capitalism, Charter Schools, and the Reproduction of Inequality in Poor, Disabled Students: An Ethnographic Case Study

Authors: Sylvia Mac

Abstract:

This ethnographic case study examines disaster capitalism, neoliberal market-based school reforms, and disability through the lens of Disability Studies in Education. More specifically, it explores neoliberalism and special education at a small, urban charter school in a large city in California and the (re)production of social inequality. The study uses Sociology of Special Education to examine the ways in which special education is used to sort and stratify disabled students. At a time when rhetoric surrounding public schools is framed in catastrophic and dismal language in order to justify the privatization of public education, small urban charter schools must be examined to learn if they are living up to their promise or acting as another way to maintain economic and racial segregation. The study concludes that neoliberal contexts threaten successful inclusive education and normalize poor, disabled students’ continued low achievement and poor post-secondary outcomes. This ethnographic case study took place at a small urban charter school in a large city in California. Participants included three special education students, the special education teacher, the special education assistant, a regular education teacher, and the two founders and charter writers. The school claimed to have a push-in model of special education where all special education students were fully included in the general education classroom. Although presented as fully inclusive, some special education students also attended a pull-out class called Study Skills. The study found that inclusion and neoliberalism are differing ideologies that cannot co-exist. Successful inclusive environments cannot thrive while under the influences of neoliberal education policies such as efficiency and cost-cutting. Additionally, the push for students to join the global knowledge economy means that more and more low attainers are further marginalized and kept in poverty. At this school, neoliberal ideology eclipsed the promise of inclusive education for special education students. This case study has shown the need for inclusive education to be interrogated through lenses that consider macro factors, such as neoliberal ideology in public education, as well as the emerging global knowledge economy and increasing income inequality. Barriers to inclusion inside the school, such as teachers’ attitudes, teacher preparedness, and school infrastructure paint only part of the picture. Inclusive education is also threatened by neoliberal ideology that shifts the responsibility from the state to the individual. This ideology is dangerous because it reifies the stereotypes of disabled students as lazy, needs drains on already dwindling budgets. If these stereotypes persist, inclusive education will have a difficult time succeeding. In order to more fully examine the ways in which inclusive education can become truly emancipatory, we need more analysis on the relationship between neoliberalism, disability, and special education.

Keywords: case study, disaster capitalism, inclusive education, neoliberalism

Procedia PDF Downloads 219
1707 Evaluating the Capability of the Flux-Limiter Schemes in Capturing the Turbulence Structures in a Fully Developed Channel Flow

Authors: Mohamed Elghorab, Vendra C. Madhav Rao, Jennifer X. Wen

Abstract:

Turbulence modelling is still evolving, and efforts are on to improve and develop numerical methods to simulate the real turbulence structures by using the empirical and experimental information. The monotonically integrated large eddy simulation (MILES) is an attractive approach for modelling turbulence in high Re flows, which is based on the solving of the unfiltered flow equations with no explicit sub-grid scale (SGS) model. In the current work, this approach has been used, and the action of the SGS model has been included implicitly by intrinsic nonlinear high-frequency filters built into the convection discretization schemes. The MILES solver is developed using the opensource CFD OpenFOAM libraries. The role of flux limiters schemes namely, Gamma, superBee, van-Albada and van-Leer, is studied in predicting turbulent statistical quantities for a fully developed channel flow with a friction Reynolds number, ReT = 180, and compared the numerical predictions with the well-established Direct Numerical Simulation (DNS) results for studying the wall generated turbulence. It is inferred from the numerical predictions that Gamma, van-Leer and van-Albada limiters produced more diffusion and overpredicted the velocity profiles, while superBee scheme reproduced velocity profiles and turbulence statistical quantities in good agreement with the reference DNS data in the streamwise direction although it deviated slightly in the spanwise and normal to the wall directions. The simulation results are further discussed in terms of the turbulence intensities and Reynolds stresses averaged in time and space to draw conclusion on the flux limiter schemes performance in OpenFOAM context.

Keywords: flux limiters, implicit SGS, MILES, OpenFOAM, turbulence statistics

Procedia PDF Downloads 189
1706 Quantum Cryptography: Classical Cryptography Algorithms’ Vulnerability State as Quantum Computing Advances

Authors: Tydra Preyear, Victor Clincy

Abstract:

Quantum computing presents many computational advantages over classical computing methods due to the utilization of quantum mechanics. The capability of this computing infrastructure poses threats to standard cryptographic systems such as RSA and AES, which are designed for classical computing environments. This paper discusses the impact that quantum computing has on cryptography, while focusing on the evolution from classical cryptographic concepts to quantum and post-quantum cryptographic concepts. Standard Cryptography is essential for securing data by utilizing encryption and decryption methods, and these methods face vulnerability problems due to the advancement of quantum computing. In order to counter these vulnerabilities, the methods that are proposed are quantum cryptography and post-quantum cryptography. Quantum cryptography uses principles such as the uncertainty principle and photon polarization in order to provide secure data transmission. In addition, the concept of Quantum key distribution is introduced to ensure more secure communication channels by distributing cryptographic keys. There is the emergence of post-quantum cryptography which is used for improving cryptographic algorithms in order to be more secure from attacks by classical and quantum computers. Throughout this exploration, the paper mentions the critical role of the advancement of cryptographic methods to keep data integrity and privacy safe from quantum computing concepts. Future research directions that would be discussed would be more effective cryptographic methods through the advancement of technology.

Keywords: quantum computing, quantum cryptography, cryptography, data integrity and privacy

Procedia PDF Downloads 20
1705 Fully Eulerian Finite Element Methodology for the Numerical Modeling of the Dynamics of Heart Valves

Authors: Aymen Laadhari

Abstract:

During the last decade, an increasing number of contributions have been made in the fields of scientific computing and numerical methodologies applied to the study of the hemodynamics in the heart. In contrast, the numerical aspects concerning the interaction of pulsatile blood flow with highly deformable thin leaflets have been much less explored. This coupled problem remains extremely challenging and numerical difficulties include e.g. the resolution of full Fluid-Structure Interaction problem with large deformations of extremely thin leaflets, substantial mesh deformations, high transvalvular pressure discontinuities, contact between leaflets. Although the Lagrangian description of the structural motion and strain measures is naturally used, many numerical complexities can arise when studying large deformations of thin structures. Eulerian approaches represent a promising alternative to readily model large deformations and handle contact issues. We present a fully Eulerian finite element methodology tailored for the simulation of pulsatile blood flow in the aorta and sinus of Valsalva interacting with highly deformable thin leaflets. Our method enables to use a fluid solver on a fixed mesh, whilst being able to easily model the mechanical properties of the valve. We introduce a semi-implicit time integration scheme based on a consistent NewtonRaphson linearization. A variant of the classical Newton method is introduced and guarantees a third-order convergence. High-fidelity computational geometries are built and simulations are performed under physiological conditions. We address in detail the main features of the proposed method, and we report several experiments with the aim of illustrating its accuracy and efficiency.

Keywords: eulerian, level set, newton, valve

Procedia PDF Downloads 277
1704 Hybrid versus Cemented Fixation in Total Knee Arthroplasty: Mid-Term Follow-Up

Authors: Pedro Gomes, Luís Sá Castelo, António Lopes, Marta Maio, Pedro Mota, Adélia Avelar, António Marques Dias

Abstract:

Introduction: Total Knee Arthroplasty (TKA) has contributed to improvement of patient`s quality of life, although it has been associated with some complications including component loosening and polyethylene wear. To prevent these complications various fixation techniques have been employed. Hybrid TKA with cemented tibial and cementless femoral components have shown favourable outcomes, although it still lack of consensus in the literature. Objectives: To evaluate the clinical and radiographic results of hybrid versus cemented TKA with an average 5 years follow-up and analyse the survival rates. Methods: A retrospective study of 125 TKAs performed in 92 patients at our institution, between 2006 to 2008, with a minimum follow-up of 2 years. The same prosthesis was used in all knees. Hybrid TKA fixation was performed in 96 knees, with a mean follow-up of 4,8±1,7 years (range, 2–8,3 years) and 29 TKAs received fully cemented fixation with a mean follow-up of 4,9±1,9 years (range, 2-8,3 years). Selection for hybrid fixation was nonrandomized and based on femoral component fit. The Oxford Knee Score (OKS 0-48) was evaluated for clinical assessment and Knee Society Roentgenographic Evaluation Scoring System was used for radiographic outcome. The survival rate was calculated using the Kaplan-Meier method, with failures defined as revision of either the tibial or femoral component for aseptic failures and all-causes (aseptic and infection). Analysis of survivorship data was performed using the log-rank test. SPSS (v22) was the computer program used for statistical analysis. Results: The hybrid group consisted of 72 females (75%) and 24 males (25%), with mean age 64±7 years (range, 50-78 years). The preoperative diagnosis was osteoarthritis (OA) in 94 knees (98%), rheumatoid arthritis (RA) in 1 knee (1%) and Posttraumatic arthritis (PTA) in 1 Knee (1%). The fully cemented group consisted of 23 females (79%) and 6 males (21%), with mean age 65±7 years (range, 47-78 years). The preoperative diagnosis was OA in 27 knees (93%), PTA in 2 knees (7%). The Oxford Knee Scores were similar between the 2 groups (hybrid 40,3±2,8 versus cemented 40,2±3). The percentage of radiolucencies seen on the femoral side was slightly higher in the cemented group 20,7% than the hybrid group 11,5% p0.223. In the cemented group there were significantly more Zone 4 radiolucencies compared to the hybrid group (13,8% versus 2,1% p0,026). Revisions for all causes were performed in 4 of the 96 hybrid TKAs (4,2%) and 1 of the 29 cemented TKAs (3,5%). The reason for revision was aseptic loosening in 3 hybrid TKAs and 1 of the cemented TKAs. Revision was performed for infection in 1 hybrid TKA. The hybrid group demonstrated a 7 years survival rate of 93% for all-cause failures and 94% for aseptic loosening. No significant difference in survivorship was seen between the groups for all-cause failures or aseptic failures. Conclusions: Hybrid TKA yields similar intermediate-term results and survival rates as fully cemented total knee arthroplasty and remains a viable option in knee joint replacement surgery.

Keywords: hybrid, survival rate, total knee arthroplasty, orthopaedic surgery

Procedia PDF Downloads 592
1703 Versatile Variation of Fischer Indolization in Green Synthesis of Nitro Substituted Tetrahydrocarbazole

Authors: Zubi Sadiq, Ghazala Yaqub, Almas Hamid

Abstract:

We wish to report herein a very stimulating variation in Fischer indole reaction by adopting solvent and catalyst free conditions. A concise synthesis of dinitro tetrahydrocarbazole derivative 3 was introduced without the use of any accelerating agent at ambient temperature with fairly good yield. Product was fully corroborated by UV, FTIR, 1H NMR, 13C NMR, MS, and CHN analysis.

Keywords: fischer indole reaction, neutral medium, solvent free synthesis, tetrahydrocarbazole

Procedia PDF Downloads 356
1702 The Effect of the Precursor Powder Size on the Electrical and Sensor Characteristics of Fully Stabilized Zirconia-Based Solid Electrolytes

Authors: Olga Yu Kurapova, Alexander V. Shorokhov, Vladimir G. Konakov

Abstract:

Nowadays, due to their exceptional anion conductivity at high temperatures cubic zirconia solid solutions, stabilized by rare-earth and alkaline-earth metal oxides, are widely used as a solid electrolyte (SE) materials in different electrochemical devices such as gas sensors, oxygen pumps, solid oxide fuel cells (SOFC), etc. Nowadays the intensive studies are carried out in a field of novel fully stabilized zirconia based SE development. The use of precursor powders for SE manufacturing allows predetermining the microstructure, electrical and sensor characteristics of zirconia based ceramics used as SE. Thus the goal of the present work was the investigation of the effect of precursor powder size on the electrical and sensor characteristics of fully stabilized zirconia-based solid electrolytes with compositions of 0,08Y2O3∙0,92ZrO2 (YSZ), 0,06Ce2O3∙ 0,06Y2O3∙0,88ZrO2 and 0,09Ce2O3∙0,06Y2O3-0,85ZrO2. The synthesis of precursors powders with different mean particle size was performed by sol-gel synthesis in the form of reversed co-precipitation from aqueous solutions. The cakes were washed until the neutral pH and pan-dried at 110 °С. Also, YSZ ceramics was obtained by conventional solid state synthesis including milling into a planetary mill. Then the powder was cold pressed into the pellets with a diameter of 7.2 and ~4 mm thickness at P ~16 kg/cm2 and then hydrostatically pressed. The pellets were annealed at 1600 °С for 2 hours. The phase composition of as-synthesized SE was investigated by X-Ray photoelectron spectroscopy ESCA (spectrometer ESCA-5400, PHI) X-ray diffraction analysis - XRD (Shimadzu XRD-6000). Following galvanic cell О2 (РО2(1)), Pt | SE | Pt, (РО2(2) = 0.21 atm) was used for SE sensor properties investigation. The value of РО2(1) was set by mixing of O2 and N2 in the defined proportions with the accuracy of  5%. The temperature was measured by Pt/Pt-10% Rh thermocouple, The cell electromotive force (EMF) measurement was carried out with ± 0.1 mV accuracy. During the operation at the constant temperature, reproducibility was better than 5 mV. Asymmetric potential measured for all SE appeared to be negligible. It was shown that the resistivity of YSZ ceramics decreases in about two times upon the mean agglomerates decrease from 200-250 to 40 nm. It is likely due to the both surface and bulk resistivity decrease in grains. So the overall decrease of grain size in ceramic SE results in the significant decrease of the total ceramics resistivity allowing sensor operation at lower temperatures. For the SE manufactured the estimation of oxygen ion transfer number tion was carried out in the range 600-800 °С. YSZ ceramics manufactured from powders with the mean particle size 40-140 nm, shows the highest values i.e. 0.97-0.98. SE manufactured from precursors with the mean particle size 40-140 nm shows higher sensor characteristic i.e. temperature and oxygen concentration EMF dependencies, EMF (ENernst - Ereal), tion, response time, then ceramics, manufactured by conventional solid state synthesis.

Keywords: oxygen sensors, precursor powders, sol-gel synthesis, stabilized zirconia ceramics

Procedia PDF Downloads 280
1701 Comparison between Separable and Irreducible Goppa Code in McEliece Cryptosystem

Authors: Newroz Nooralddin Abdulrazaq, Thuraya Mahmood Qaradaghi

Abstract:

The McEliece cryptosystem is an asymmetric type of cryptography based on error correction code. The classical McEliece used irreducible binary Goppa code which considered unbreakable until now especially with parameter [1024, 524, and 101], but it is suffering from large public key matrix which leads to be difficult to be used practically. In this work Irreducible and Separable Goppa codes have been introduced. The Irreducible and Separable Goppa codes used are with flexible parameters and dynamic error vectors. A Comparison between Separable and Irreducible Goppa code in McEliece Cryptosystem has been done. For encryption stage, to get better result for comparison, two types of testing have been chosen; in the first one the random message is constant while the parameters of Goppa code have been changed. But for the second test, the parameters of Goppa code are constant (m=8 and t=10) while the random message have been changed. The results show that the time needed to calculate parity check matrix in separable are higher than the one for irreducible McEliece cryptosystem, which is considered expected results due to calculate extra parity check matrix in decryption process for g2(z) in separable type, and the time needed to execute error locator in decryption stage in separable type is better than the time needed to calculate it in irreducible type. The proposed implementation has been done by Visual studio C#.

Keywords: McEliece cryptosystem, Goppa code, separable, irreducible

Procedia PDF Downloads 264
1700 Improved Signal-To-Noise Ratio by the 3D-Functionalization of Fully Zwitterionic Surface Coatings

Authors: Esther Van Andel, Stefanie C. Lange, Maarten M. J. Smulders, Han Zuilhof

Abstract:

False outcomes of diagnostic tests are a major concern in medical health care. To improve the reliability of surface-based diagnostic tests, it is of crucial importance to diminish background signals that arise from the non-specific binding of biomolecules, a process called fouling. The aim is to create surfaces that repel all biomolecules except the molecule of interest. This can be achieved by incorporating antifouling protein repellent coatings in between the sensor surface and it’s recognition elements (e.g. antibodies, sugars, aptamers). Zwitterionic polymer brushes are considered excellent antifouling materials, however, to be able to bind the molecule of interest, the polymer brushes have to be functionalized and so far this was only achieved at the expense of either antifouling or binding capacity. To overcome this limitation, we combined both features into one single monomer: a zwitterionic sulfobetaine, ensuring antifouling capabilities, equipped with a clickable azide moiety which allows for further functionalization. By copolymerizing this monomer together with a standard sulfobetaine, the number of azides (and with that the number of recognition elements) can be tuned depending on the application. First, the clickable azido-monomer was synthesized and characterized, followed by copolymerizing this monomer to yield functionalizable antifouling brushes. The brushes were fully characterized using surface characterization techniques like XPS, contact angle measurements, G-ATR-FTIR and XRR. As a proof of principle, the brushes were subsequently functionalized with biotin via strain-promoted alkyne azide click reactions, which yielded a fully zwitterionic biotin-containing 3D-functionalized coating. The sensing capacity was evaluated by reflectometry using avidin and fibrinogen containing protein solutions. The surfaces showed excellent antifouling properties as illustrated by the complete absence of non-specific fibrinogen binding, while at the same time clear responses were seen for the specific binding of avidin. A great increase in signal-to-noise ratio was observed, even when the amount of functional groups was lowered to 1%, compared to traditional modification of sulfobetaine brushes that rely on a 2D-approach in which only the top-layer can be functionalized. This study was performed on stoichiometric silicon nitride surfaces for future microring resonator based assays, however, this methodology can be transferred to other biosensor platforms which are currently being investigated. The approach presented herein enables a highly efficient strategy for selective binding with retained antifouling properties for improved signal-to-noise ratios in binding assays. The number of recognition units can be adjusted to a specific need, e.g. depending on the size of the analyte to be bound, widening the scope of these functionalizable surface coatings.

Keywords: antifouling, signal-to-noise ratio, surface functionalization, zwitterionic polymer brushes

Procedia PDF Downloads 305
1699 Exploring the Connectedness of Ad Hoc Mesh Networks in Rural Areas

Authors: Ibrahim Obeidat

Abstract:

Reaching a fully-connected network of mobile nodes in rural areas got a great attention between network researchers. This attention rose due to the complexity and high costs while setting up the needed infrastructures for these networks, in addition to the low transmission range these nodes has. Terranet technology, as an example, employs ad-hoc mesh network where each node has a transmission range not exceed one kilometer, this means that every two nodes are able to communicate with each other if they are just one kilometer far from each other, otherwise a third-party will play the role of the “relay”. In Terranet, and as an idea to reduce network setup cost, every node in the network will be considered as a router that is responsible of forwarding data between other nodes which result in a decentralized collaborative environment. Most researches on Terranet presents the idea of how to encourage mobile nodes to become more cooperative by letting their devices in “ON” state as long as possible while accepting to play the role of relay (router). This research presents the issue of finding the percentage of nodes in ad-hoc mesh network within rural areas that should play the role of relay at every time slot, relating to what is the actual area coverage of nodes in order to have the network reach the fully-connectivity. Far from our knowledge, till now there is no current researches discussed this issue. The research is done by making an implementation that depends on building adjacency matrix as an indicator to the connectivity between network members. This matrix is continually updated until each value in it refers to the number of hubs that should be followed to reach from one node to another. After repeating the algorithm on different area sizes, different coverage percentages for each size, and different relay percentages for several times, results extracted shows that for area coverage less than 5% we need to have 40% of the nodes to be relays, where 10% percentage is enough for areas with node coverage greater than 5%.

Keywords: ad-hoc mesh networks, network connectivity, mobile ad-hoc networks, Terranet, adjacency matrix, simulator, wireless sensor networks, peer to peer networks, vehicular Ad hoc networks, relay

Procedia PDF Downloads 282
1698 Role of Organizational Culture in Building Sustainable Employee’s Performance in Organizations: A Case Study of Zenith Bank PLC Jalingo Taraba State Nigeria

Authors: Jerome Nyameh

Abstract:

The most valuable asset in the existence of organization is the employees and their ability in maintain appreciable level of performance which support the goal of the organization and the ability to do that depend largely on the organizational culture and culture has been considered most currently as the factor that relate positively to organizational excellence and sustainable employee’s performance over the period of time An employee engagement program will not go far without first establishing the organizational culture that is required to support sustainability. This means integrating sustainability into the overall employee’s performance, with clear vision, goals and metrics. It means having strong culture and a collaborative governance structure that has been develop as a ways of doing things in the organization for decision making and resource allocation. It requires a rewards and recognition program to support and reinforce sustainability behaviors. With such a culture in place, organization will be able to develop a strategy that fully engages employees, while fully realizing the benefits of their contributions. The study investigated empirically the role of organizational culture building sustainable employee’s performance using Zenith bank PLC a model where organizational culture will build sustainable employees performance strategy for a lasting actualization of organizational was developed. In order to achieve the research objectives of (i) to assess how organizational culture can build sustainable employee’s performance (ii) to analyze the gap that exists between organizational culture and sustainable employee’s performance in the organization, a survey questionnaires of 20 items was administered to sixty respondents. The findings of this study have practical implications for organizational leaders, managers and employees, and their organizations, particularly commercial banks in Nigeria, besides offering scope for further research in the area of organizational culture and sustainable employee’s performance. It will also show a significance and positive relationship that exist between organizational culture and sustainable employee’s performance, as means of building viable organization with cultural uniqueness and excellence performance in the world of competition.

Keywords: organizational culture, sustainable employee’s performance, organizations, Zenith Bank PLC Nigeria

Procedia PDF Downloads 510