Search results for: array signal processing
3657 Solution-Processed Threshold Switching Selectors Based on Highly Flexible, Transparent and Scratchable Silver Nanowires Conductive Films
Authors: Peiyuan Guan, Tao Wan, Dewei Chu
Abstract:
With the flash memory approaching its physical limit, the emerging resistive random-access memory (RRAM) has been considered as one of the most promising candidates for the next-generation non-volatile memory. One selector-one resistor configuration has shown the most promising way to resolve the crosstalk issue without affecting the scalability and high-density integration of the RRAM array. By comparison with other candidates of selectors (such as diodes and nonlinear devices), threshold switching selectors dominated by formation/spontaneous rupture of fragile conductive filaments have been proved to possess low voltages, high selectivity, and ultra-low current leakage. However, the flexibility and transparency of selectors are barely mentioned. Therefore, it is a matter of urgency to develop a selector with highly flexible and transparent properties to assist the application of RRAM for a diversity of memory devices. In this work, threshold switching selectors were designed using a facilely solution-processed fabrication on AgNWs@PDMS composite films, which show high flexibility, transparency and scratch resistance. As-fabricated threshold switching selectors also have revealed relatively high selectivity (~107), low operating voltages (Vth < 1 V) and good switching performance.Keywords: flexible and transparent, resistive random-access memory, silver nanowires, threshold switching selector
Procedia PDF Downloads 1283656 Generalized Up-downlink Transmission using Black-White Hole Entanglement Generated by Two-level System Circuit
Authors: Muhammad Arif Jalil, Xaythavay Luangvilay, Montree Bunruangses, Somchat Sonasang, Preecha Yupapin
Abstract:
Black and white holes form the entangled pair⟨BH│WH⟩, where a white hole occurs when the particle moves at the same speed as light. The entangled black-white hole pair is at the center with the radian between the gap. When the speed of particle motion is slower than light, the black hole is gravitational (positive gravity), where the white hole is smaller than the black hole. On the downstream side, the entangled pair appears to have a black hole outside the gap increases until the white holes disappear, which is the emptiness paradox. On the upstream side, when moving faster than light, white holes form times tunnels, with black holes becoming smaller. It will continue to move faster and further when the black hole disappears and becomes a wormhole (Singularity) that is only a white hole in emptiness (Emptiness). This research studies use of black and white holes generated by a two-level circuit for communication transmission carriers, in which high ability and capacity of data transmission can be obtained. The black and white hole pair can be generated by the two-level system circuit when the speech of a particle on the circuit is equal to the speed of light. The black hole forms when the particle speed has increased from slower to equal to the light speed, while the white hole is established when the particle comes down faster than light. They are bound by the entangled pair, signal and idler, ⟨Signal│Idler⟩, and the virtual ones for the white hole, which has an angular displacement of half of π radian. A two-level system is made from an electronic circuit to create black and white holes bound by the entangled bits that are immune or cloning-free from thieves. Start by creating a wave-particle behavior when its speed is equal to light black hole is in the middle of the entangled pair, which is the two bit gate. The required information can be input into the system and wrapped by the black hole carrier. A timeline (Tunnel) occurs when the wave-particle speed is faster than light, from which the entangle pair is collapsed. The transmitted information is safely in the time tunnel. The required time and space can be modulated via the input for the downlink operation. The downlink is established when the particle speed is given by a frequency(energy) form is down and entered into the entangled gap, where this time the white hole is established. The information with the required destination is wrapped by the white hole and retrieved by the clients at the destination. The black and white holes are disappeared, and the information can be recovered and used.Keywords: cloning free, time machine, teleportation, two-level system
Procedia PDF Downloads 743655 Pharmaceutical Science and Development in Drug Research
Authors: Adegoke Yinka Adebayo
Abstract:
An understanding of the critical product attributes that impact on in vivo performance is key to the production of safe and effective medicines. Thus, a key driver for our research is the development of new basic science and technology underpinning the development of new pharmaceutical products. Research includes the structure and properties of drugs and excipients, biopharmaceutical characterisation, pharmaceutical processing and technology and formulation and analysis.Keywords: drug discovery, drug development, drug delivery
Procedia PDF Downloads 4943654 Energy Production with Closed Methods
Authors: Bujar Ismaili, Bahti Ismajli, Venhar Ismaili, Skender Ramadani
Abstract:
In Kosovo, the problem with the electricity supply is huge and does not meet the demands of consumers. Older thermal power plants, which are regarded as big environmental polluters, produce most of the energy. Our experiment is based on the production of electricity using the closed method that does not affect environmental pollution by using waste as fuel that is considered to pollute the environment. The experiment was carried out in the village of Godanc, municipality of Shtime - Kosovo. In the experiment, a production line based on the production of electricity and central heating was designed at the same time. The results are the benefits of electricity as well as the release of temperature for heating with minimal expenses and with the release of 0% gases into the atmosphere. During this experiment, coal, plastic, waste from wood processing, and agricultural wastes were used as raw materials. The method utilized in the experiment allows for the release of gas through pipes and filters during the top-to-bottom combustion of the raw material in the boiler, followed by the method of gas filtration from waste wood processing (sawdust). During this process, the final product is obtained - gas, which passes through the carburetor, which enables the gas combustion process and puts into operation the internal combustion machine and the generator and produces electricity that does not release gases into the atmosphere. The obtained results show that the system provides energy stability without environmental pollution from toxic substances and waste, as well as with low production costs. From the final results, it follows that: in the case of using coal fuel, we have benefited from more electricity and higher temperature release, followed by plastic waste, which also gave good results. The results obtained during these experiments prove that the current problems of lack of electricity and heating can be met at a lower cost and have a clean environment and waste management.Keywords: energy, heating, atmosphere, waste, gasification
Procedia PDF Downloads 2353653 Deep Learning-Based Approach to Automatic Abstractive Summarization of Patent Documents
Authors: Sakshi V. Tantak, Vishap K. Malik, Neelanjney Pilarisetty
Abstract:
A patent is an exclusive right granted for an invention. It can be a product or a process that provides an innovative method of doing something, or offers a new technical perspective or solution to a problem. A patent can be obtained by making the technical information and details about the invention publicly available. The patent owner has exclusive rights to prevent or stop anyone from using the patented invention for commercial uses. Any commercial usage, distribution, import or export of a patented invention or product requires the patent owner’s consent. It has been observed that the central and important parts of patents are scripted in idiosyncratic and complex linguistic structures that can be difficult to read, comprehend or interpret for the masses. The abstracts of these patents tend to obfuscate the precise nature of the patent instead of clarifying it via direct and simple linguistic constructs. This makes it necessary to have an efficient access to this knowledge via concise and transparent summaries. However, as mentioned above, due to complex and repetitive linguistic constructs and extremely long sentences, common extraction-oriented automatic text summarization methods should not be expected to show a remarkable performance when applied to patent documents. Other, more content-oriented or abstractive summarization techniques are able to perform much better and generate more concise summaries. This paper proposes an efficient summarization system for patents using artificial intelligence, natural language processing and deep learning techniques to condense the knowledge and essential information from a patent document into a single summary that is easier to understand without any redundant formatting and difficult jargon.Keywords: abstractive summarization, deep learning, natural language Processing, patent document
Procedia PDF Downloads 1233652 Health Care Waste Management Practices in Liberia: An Investigative Case Study
Authors: V. Emery David Jr., J. Wenchao, D. Mmereki, Y. John, F. Heriniaina
Abstract:
Healthcare waste management continues to present an array of challenges for developing countries, and Liberia is of no exception. There is insufficient information available regarding the generation, handling, and disposal of health care waste. This face serves as an impediment to healthcare management schemes. The specific objective of this study is to present an evaluation of the current health care management practices in Liberia. It also presented procedures, techniques used, methods of handling, transportation, and disposal methods of wastes as well as the quantity and composition of health care waste. This study was conducted as an investigative case study, covering three different health care facilities; a hospital, a health center, and a clinic in Monrovia, Montserrado County. The average waste generation was found to be 0-7kg per day at the clinic and health center and 8-15kg per/day at the hospital. The composition of the waste includes hazardous and non-hazardous waste i.e. plastic, papers, sharps, and pathological elements etc. Nevertheless, the investigation showed that the healthcare waste generated by the surveyed healthcare facilities were not properly handled because of insufficient guidelines for separate collection, and classification, and adequate methods for storage and proper disposal of generated wastes. This therefore indicates that there is a need for improvement within the healthcare waste management system to improve the existing situation.Keywords: disposal, healthcare waste, management, Montserrado County, Monrovia
Procedia PDF Downloads 3453651 Output Voltage Analysis of CMOS Colpitts Oscillator with Short Channel Device
Authors: Maryam Ebrahimpour, Amir Ebrahimi
Abstract:
This paper presents the steady-state amplitude analysis of MOS Colpitts oscillator with short channel device. The proposed method is based on a large signal analysis and the nonlinear differential equations that govern the oscillator circuit behaviour. Also, the short channel effects are considered in the proposed analysis and analytical equations for finding the steady-state oscillation amplitude are derived. The output voltage calculated from this analysis is in excellent agreement with simulations for a wide range of circuit parameters.Keywords: colpitts oscillator, CMOS, electronics, circuits
Procedia PDF Downloads 3513650 Natural Gas Flow Optimization Using Pressure Profiling and Isolation Techniques
Authors: Syed Tahir Shah, Fazal Muhammad, Syed Kashif Shah, Maleeha Gul
Abstract:
In recent days, natural gas has become a relatively clean and quality source of energy, which is recovered from deep wells by expensive drilling activities. The recovered substance is purified by processing in multiple stages to remove the unwanted/containments like dust, dirt, crude oil and other particles. Mostly, gas utilities are concerned with essential objectives of quantity/quality of natural gas delivery, financial outcome and safe natural gas volumetric inventory in the transmission gas pipeline. Gas quantity and quality are primarily related to standards / advanced metering procedures in processing units/transmission systems, and the financial outcome is defined by purchasing and selling gas also the operational cost of the transmission pipeline. SNGPL (Sui Northern Gas Pipelines Limited) Pakistan has a wide range of diameters of natural gas transmission pipelines network of over 9125 km. This research results in answer a few of the issues in accuracy/metering procedures via multiple advanced gadgets for gas flow attributes after being utilized in the transmission system and research. The effects of good pressure management in transmission gas pipeline network in contemplation to boost the gas volume deposited in the existing network and finally curbing gas losses UFG (Unaccounted for gas) for financial benefits. Furthermore, depending on the results and their observation, it is directed to enhance the maximum allowable working/operating pressure (MAOP) of the system to 1235 PSIG from the current round about 900 PSIG, such that the capacity of the network could be entirely utilized. In gross, the results depict that the current model is very efficient and provides excellent results in the minimum possible time.Keywords: natural gas, pipeline network, UFG, transmission pack, AGA
Procedia PDF Downloads 953649 The Influence of Structural Disorder and Phonon on Metal-To-Insulator Transition of VO₂
Authors: Sang-Wook Han, In-Hui Hwang, Zhenlan Jin, Chang-In Park
Abstract:
We used temperature-dependent X-Ray absorption fine structure (XAFS) measurements to examine the local structural properties around vanadium atoms at the V K edge from VO₂ films. A direct comparison of simultaneously-measured resistance and XAFS from the VO₂ films showed that the thermally-driven structural phase transition (SPT) occurred prior to the metal-insulator transition (MIT) during heating, whereas these changed simultaneously during cooling. XAFS revealed a significant increase in the Debye-Waller factors of the V-O and V-V pairs in the {111} direction of the R-phase VO₂ due to the phonons of the V-V arrays along the direction in a metallic phase. A substantial amount of structural disorder existing on the V-V pairs along the c-axis in both M₁ and R phases indicates the structural instability of V-V arrays in the axis. The anomalous structural disorder observed on all atomic sites at the SPT prevents the migration of the V 3d¹ electrons, resulting in a Mott insulator in the M₂-phase VO₂. The anomalous structural disorder, particularly, at vanadium sites, effectively affects the migration of metallic electrons, resulting in the Mott insulating properties in M₂ phase and a non-congruence of the SPT, MIT, and local density of state. The thermally-induced phonons in the {111} direction assist the delocalization of the V 3d¹ electrons in the R phase VO₂ and the electrons likely migrate via the V-V array in the {111} direction as well as the V-V dimerization along the c-axis. This study clarifies that the tetragonal symmetry is essentially important for the metallic phase in VO₂.Keywords: metal-insulator transition, XAFS, VO₂, structural-phase transition
Procedia PDF Downloads 2713648 Optimizing CNC Production Line Efficiency Using NSGA-II: Adaptive Layout and Operational Sequence for Enhanced Manufacturing Flexibility
Authors: Yi-Ling Chen, Dung-Ying Lin
Abstract:
In the manufacturing process, computer numerical control (CNC) machining plays a crucial role. CNC enables precise machinery control through computer programs, achieving automation in the production process and significantly enhancing production efficiency. However, traditional CNC production lines often require manual intervention for loading and unloading operations, which limits the production line's operational efficiency and production capacity. Additionally, existing CNC automation systems frequently lack sufficient intelligence and fail to achieve optimal configuration efficiency, resulting in the need for substantial time to reconfigure production lines when producing different products, thereby impacting overall production efficiency. Using the NSGA-II algorithm, we generate production line layout configurations that consider field constraints and select robotic arm specifications from an arm list. This allows us to calculate loading and unloading times for each job order, perform demand allocation, and assign processing sequences. The NSGA-II algorithm is further employed to determine the optimal processing sequence, with the aim of minimizing demand completion time and maximizing average machine utilization. These objectives are used to evaluate the performance of each layout, ultimately determining the optimal layout configuration. By employing this method, it enhance the configuration efficiency of CNC production lines and establish an adaptive capability that allows the production line to respond promptly to changes in demand. This will minimize production losses caused by the need to reconfigure the layout, ensuring that the CNC production line can maintain optimal efficiency even when adjustments are required due to fluctuating demands.Keywords: evolutionary algorithms, multi-objective optimization, pareto optimality, layout optimization, operations sequence
Procedia PDF Downloads 213647 Copywriting and the Creative Edge
Authors: Dandeswar Bisoyi, Preeti Yadav, Utpal Barua
Abstract:
This study address particular way that verbal information can affect the processing of positive and interesting qualities which help in making the brand attractive to the consumer. Also, it address the development of a communication strategy which is a very important part of the marketing plan we have to take into account many factors. Out of all the product strengths, the strategy has to outline one marked differential which will drive our brand. This is the fundamental base on which the entire creative strategy will be big idea-based.Keywords: copy writing, advertisement, marketing, branding, recall
Procedia PDF Downloads 5823646 Markov-Chain-Based Optimal Filtering and Smoothing
Authors: Garry A. Einicke, Langford B. White
Abstract:
This paper describes an optimum filter and smoother for recovering a Markov process message from noisy measurements. The developments follow from an equivalence between a state space model and a hidden Markov chain. The ensuing filter and smoother employ transition probability matrices and approximate probability distribution vectors. The properties of the optimum solutions are retained, namely, the estimates are unbiased and minimize the variance of the output estimation error, provided that the assumed parameter set are correct. Methods for estimating unknown parameters from noisy measurements are discussed. Signal recovery examples are described in which performance benefits are demonstrated at an increased calculation cost.Keywords: optimal filtering, smoothing, Markov chains
Procedia PDF Downloads 3173645 The Foundation Binary-Signals Mechanics and Actual-Information Model of Universe
Authors: Elsadig Naseraddeen Ahmed Mohamed
Abstract:
In contrast to the uncertainty and complementary principle, it will be shown in the present paper that the probability of the simultaneous occupation event of any definite values of coordinates by any definite values of momentum and energy at any definite instance of time can be described by a binary definite function equivalent to the difference between their numbers of occupation and evacuation epochs up to that time and also equivalent to the number of exchanges between those occupation and evacuation epochs up to that times modulus two, these binary definite quantities can be defined at all point in the time’s real-line so it form a binary signal represent a complete mechanical description of physical reality, the time of these exchanges represent the boundary of occupation and evacuation epochs from which we can calculate these binary signals using the fact that the time of universe events actually extends in the positive and negative of time’s real-line in one direction of extension when these number of exchanges increase, so there exists noninvertible transformation matrix can be defined as the matrix multiplication of invertible rotation matrix and noninvertible scaling matrix change the direction and magnitude of exchange event vector respectively, these noninvertible transformation will be called actual transformation in contrast to information transformations by which we can navigate the universe’s events transformed by actual transformations backward and forward in time’s real-line, so these information transformations will be derived as an elements of a group can be associated to their corresponded actual transformations. The actual and information model of the universe will be derived by assuming the existence of time instance zero before and at which there is no coordinate occupied by any definite values of momentum and energy, and then after that time, the universe begin its expanding in spacetime, this assumption makes the need for the existence of Laplace’s demon who at one moment can measure the positions and momentums of all constituent particle of the universe and then use the law of classical mechanics to predict all future and past of universe’s events, superfluous, we only need for the establishment of our analog to digital converters to sense the binary signals that determine the boundaries of occupation and evacuation epochs of the definite values of coordinates relative to its origin by the definite values of momentum and energy as present events of the universe from them we can predict approximately in high precision it's past and future events.Keywords: binary-signal mechanics, actual-information model of the universe, actual-transformation, information-transformation, uncertainty principle, Laplace's demon
Procedia PDF Downloads 1753644 Breath Ethanol Imaging System Using Real Time Biochemical Luminescence for Evaluation of Alcohol Metabolic Capacity
Authors: Xin Wang, Munkbayar Munkhjargal, Kumiko Miyajima, Takahiro Arakawa, Kohji Mitsubayashi
Abstract:
The measurement of gaseous ethanol plays an important role of evaluation of alcohol metabolic capacity in clinical and forensic analysis. A 2-dimensional visualization system for gaseous ethanol was constructed and tested in visualization of breath and transdermal alcohol. We demonstrated breath ethanol measurement using developed high-sensitive visualization system. The concentration of breath ethanol calculated with the imaging signal was significantly different between the volunteer subjects of ALDH2 (+) and (-).Keywords: breath ethanol, ethnaol imaging, biochemical luminescence, alcohol metabolism
Procedia PDF Downloads 3513643 Hot Deformability of Si-Steel Strips Containing Al
Authors: Mohamed Yousef, Magdy Samuel, Maha El-Meligy, Taher El-Bitar
Abstract:
The present work is dealing with 2% Si-steel alloy. The alloy contains 0.05% C as well as 0.85% Al. The alloy under investigation would be used for electrical transformation purposes. A heating (expansion) - cooling (contraction) dilation investigation was executed to detect the a, a+g, and g transformation temperatures at the inflection points of the dilation curve. On heating, primary a was detected at a temperature range between room temperature and 687 oC. The domain of a+g was detected in the range between 687 oC and 746 oC. g phase exists in the closed g region at the range between 746 oC and 1043 oC. The domain of a phase appears again at a temperature range between 1043 and 1105 oC, and followed by secondary a at temperature higher than 1105 oC. A physical simulation of thermo-mechanical processing on the as-cast alloy was carried out. The simulation process took into consideration the hot flat rolling pilot plant parameters. The process was executed on the thermo-mechanical simulator (Gleeble 3500). The process was designed to include seven consecutive passes. The 1st pass represents the roughing stage, while the remaining six passes represent finish rolling stage. The whole process was executed at the temperature range from 1100 oC to 900 oC. The amount of strain starts with 23.5% at the roughing pass and decreases continuously to reach 7.5 % at the last finishing pass. The flow curve of the alloy can be abstracted from the stress-strain curves representing simulated passes. It shows alloy hardening from a pass to the other up to pass no. 6, as a result of decreasing the deformation temperature and increasing of cumulative strain. After pass no. 6, the deformation process enhances the dynamic recrystallization phenomena to appear, where the z-parameter would be high.Keywords: si- steel, hot deformability, critical transformation temperature, physical simulation, thermo-mechanical processing, flow curve, dynamic softening.
Procedia PDF Downloads 2453642 Comprehensive Evaluation of COVID-19 Through Chest Images
Authors: Parisa Mansour
Abstract:
The coronavirus disease 2019 (COVID-19) was discovered and rapidly spread to various countries around the world since the end of 2019. Computed tomography (CT) images have been used as an important alternative to the time-consuming RT. PCR test. However, manual segmentation of CT images alone is a major challenge as the number of suspected cases increases. Thus, accurate and automatic segmentation of COVID-19 infections is urgently needed. Because the imaging features of the COVID-19 infection are different and similar to the background, existing medical image segmentation methods cannot achieve satisfactory performance. In this work, we try to build a deep convolutional neural network adapted for the segmentation of chest CT images with COVID-19 infections. First, we maintain a large and novel chest CT image database containing 165,667 annotated chest CT images from 861 patients with confirmed COVID-19. Inspired by the observation that the boundary of an infected lung can be improved by global intensity adjustment, we introduce a feature variable block into the proposed deep CNN, which adjusts the global features of features to segment the COVID-19 infection. The proposed PV array can effectively and adaptively improve the performance of functions in different cases. We combine features of different scales by proposing a progressive atrocious space pyramid fusion scheme to deal with advanced infection regions with various aspects and shapes. We conducted experiments on data collected in China and Germany and showed that the proposed deep CNN can effectively produce impressive performance.Keywords: chest, COVID-19, chest Image, coronavirus, CT image, chest CT
Procedia PDF Downloads 573641 Heavy Metal Contents in Vegetable Oils of Kazakhstan Origin and Life Risk Assessment
Authors: A. E. Mukhametov, M. T. Yerbulekova, D. R. Dautkanova, G. A. Tuyakova, G. Aitkhozhayeva
Abstract:
The accumulation of heavy metals in food is a constant problem in many parts of the world. Vegetable oils are widely used, both for cooking and for processing in the food industry, meeting the main dietary requirements. One of the main chemical pollutants, heavy metals, is usually found in vegetable oils. These chemical pollutants are carcinogenic, teratogenic and immunotoxic, harmful to consumption and have a negative effect on human health even in trace amounts. Residues of these substances can easily accumulate in vegetable oil during cultivation, processing and storage. In this article, the content of the concentration of heavy metal ions in vegetable oils of Kazakhstan production is studied: sunflower, rapeseed, safflower and linseed oil. Heavy metals: arsenic, cadmium, lead and nickel, were determined in three repetitions by the method of flame atomic absorption. Analysis of vegetable oil samples revealed that the largest lead contamination (Pb) was determined to be 0.065 mg/kg in linseed oil. The content of cadmium (Cd) in the largest amount of 0.009 mg/kg was found in safflower oil. Arsenic (As) content was determined in rapeseed and safflower oils at 0.003 mg/kg, and arsenic (As) was not detected in linseed and sunflower oil. The nickel (Ni) content in the largest amount of 0.433 mg/kg was in linseed oil. The heavy metal contents in the test samples complied with the requirements of regulatory documents for vegetable oils. An assessment of the health risk of vegetable oils with a daily consumption of 36 g per day shows that all samples of vegetable oils produced in Kazakhstan are safe for consumption. But further monitoring is needed, since all these metals are toxic and their harmful effects become apparent only after several years of exposure.Keywords: vegetable oil, sunflower oil, linseed oil, safflower oil, toxic metals, food safety, rape oil
Procedia PDF Downloads 1333640 The Use of Political Savviness in Dealing with Workplace Ostracism: A Social Information Processing Perspective
Authors: Amy Y. Wang, Eko L. Yi
Abstract:
Can vicarious experiences of workplace ostracism affect employees’ willingness to voice? Given the increasingly interdependent nature of the modern workplace in which employees rely on social interactions to fulfill organizational goals, workplace ostracism –the extent to which an individual perceives that he or she is ignored or excluded by others in the workplace– has garnered significant interest from scholars and practitioners alike. Extending beyond conventional studies that largely focus on the perspectives and outcomes of ostracized targets, we address the indirect effects of workplace ostracism on third-party employees embedded in the same social context. Using a social information processing approach, we propose that the ostracism of coworkers acts as political information that influences third-party employees in their decisions to engage in risky and discretionary behaviors such as employee voice. To make sense of and to navigate through experiences of workplace ostracism, we posit that both political understanding and political skill allow third party employees to minimize the risks and uncertainty of voicing. This conceptual model was tested by a study involving 154 supervisor-subordinate dyads of a publicly listed bio-technology firm located in Mainland China. Each supervisor and their direct subordinates composed of a work team; each team had a minimum of two subordinates and a maximum of four subordinates. Human resources used the master list to distribute the ID coded questionnaires to the matching names. All studied constructs were measured using existing scales proved effective in previous literature. Hypotheses were tested using Confirmatory Factor Analysis and Hierarchal Multiple Regression. All three hypotheses were supported which showed that employees were less likely to engage in voice behaviors when their coworkers reported having experienced ostracism in the workplace. Results also showed a significant three-way interaction between political understanding and political skill on the relationship between coworkers’ ostracism and employee voice, indicating that political savviness is a valuable resource in mitigating ostracism’s negative and indirect effects. Our results illustrated that an employee’s coworkers being ostracized indeed adversely impacted his or her own voice behavior. However, not all individuals reacted passively to the social context; rather, we found that politically savvy individuals – possessing both political understanding and political skill – and their voice behaviors were less impacted by ostracism in their work environment. At the same time, we found that having only political understanding or only political skill was significantly less effective in mitigating ostracism’s negative effects, suggesting a necessary duality of political knowledge and political skill in combatting ostracism. Organizational implications, recommendations, and future research ideas are also discussed.Keywords: employee voice, organizational politics, social information processing, workplace ostracism
Procedia PDF Downloads 1403639 Effect of Constant and Variable Temperature on the Morphology of TiO₂ Nanotubes Prepared by Two-Step Anodization Method
Authors: Tayyaba Ghani, Mazhar Mehmood, Mohammad Mujahid
Abstract:
TiO₂ nanotubes are receiving immense attraction in the field of dye-sensitized solar cells due to their well-defined nanostructures, efficient electron transport and large surface area as compared to other one dimensional structures. In the present work, we have investigated the influence of temperature on the morphology of anodically produced self-organized Titanium oxide nanotubes (TiNTs). TiNTs are synthesized by two-step anodization method in an ethylene glycol based electrolytes containing ammonium fluoride. Experiments are performed at constant anodization voltage for two hours. An investigation by the SEM images reveals that if the temperature is kept constant during the anodizing experiment, variation in the average tube diameter is significantly reduced. However, if the temperature is not controlled then due to the exothermic nature of reactions for the formation of TiNTs, the temperature of electrolyte keep on increasing. This variation in electrolyte bath temperature introduced strong variations in tube diameter (20 nm to 160 nm) along the length of tubes. Current profiles, recorded during the anodization experiment, predict the effect of constant and varying experimental temperatures as well. In both cases, XRD results show the complete anatase crystal structure of nanotube upon annealing at 450 °C. Present work highlights the importance of constant temperature during the anodization experiments in order to develop an ordered array of nanotubes with a uniform tube diameter.Keywords: anodization, ordering, temperature, TiO₂ nanotubes
Procedia PDF Downloads 1713638 Computational Fluid Dynamics Analysis of Cyclone Separator Performance Using Discrete Phase Model
Authors: Sandeep Mohan Ahuja, Gulshan Kumar Jawa
Abstract:
Cyclone separators are crucial components in various industries tasked with efficiently separating particulate matter from gas streams. Achieving optimal performance hinges on a deep understanding of flow dynamics and particle behaviour within these separators. In this investigation, Computational Fluid Dynamics (CFD) simulations are conducted utilizing the Discrete Phase Model (DPM) to dissect the intricate flow patterns, particle trajectories, and separation efficiency within cyclone separators. The study delves into the influence of pivotal parameters like inlet velocity, particle size distribution, and cyclone geometry on separation efficiency. Through numerical simulations, a comprehensive comprehension of fluid-particle interaction phenomena within cyclone separators is attained, allowing for the assessment of solid collection efficiency across diverse operational conditions and geometrical setups. The insights gleaned from this study promise to advance our understanding of the complex interplay between fluid and particle within cyclone separators, thereby enabling optimization across a wide array of industrial applications. By harnessing the power of CFD simulations and the DPM, this research endeavours to furnish valuable insights for designing, operating, and evaluating the performance of cyclone separators, ultimately fostering greater efficiency and environmental sustainability within industrial processes.Keywords: cyclone separator, computational fluid dynamics, enhancing efficiency, discrete phase model
Procedia PDF Downloads 523637 Controllable Modification of Glass-Crystal Composites with Ion-Exchange Technique
Authors: Andrey A. Lipovskii, Alexey V. Redkov, Vyacheslav V. Rusan, Dmitry K. Tagantsev, Valentina V. Zhurikhina
Abstract:
The presented research is related to the development of recently proposed technique of the formation of composite materials, like optical glass-ceramics, with predetermined structure and properties of the crystalline component. The technique is based on the control of the size and concentration of the crystalline grains using the phenomenon of glass-ceramics decrystallization (vitrification) induced by ion-exchange. This phenomenon was discovered and explained in the beginning of the 2000s, while related theoretical description was given in 2016 only. In general, the developed theory enables one to model the process and optimize the conditions of ion-exchange processing of glass-ceramics, which provide given properties of crystalline component, in particular, profile of the average size of the crystalline grains. The optimization is possible if one knows two dimensionless parameters of the theoretical model. One of them (β) is the value which is directly related to the solubility of crystalline component of the glass-ceramics in the glass matrix, and another (γ) is equal to the ratio of characteristic times of ion-exchange diffusion and crystalline grain dissolution. The presented study is dedicated to the development of experimental technique and simulation which allow determining these parameters. It is shown that these parameters can be deduced from the data on the space distributions of diffusant concentrations and average size of crystalline grains in the glass-ceramics samples subjected to ion-exchange treatment. Measurements at least at two temperatures and two processing times at each temperature are necessary. The composite material used was a silica-based glass-ceramics with crystalline grains of Li2OSiO2. Cubical samples of the glass-ceramics (6x6x6 mm3) underwent the ion exchange process in NaNO3 salt melt at 520 oC (for 16 and 48 h), 540 oC (for 8 and 24 h), 560 oC (for 4 and 12 h), and 580 oC (for 2 and 8 h). The ion exchange processing resulted in the glass-ceramics vitrification in the subsurface layers where ion-exchange diffusion took place. Slabs about 1 mm thick were cut from the central part of the samples and their big facets were polished. These slabs were used to find profiles of diffusant concentrations and average size of the crystalline grains. The concentration profiles were determined from refractive index profiles measured with Max-Zender interferometer, and profiles of the average size of the crystalline grains were determined with micro-Raman spectroscopy. Numerical simulation were based on the developed theoretical model of the glass-ceramics decrystallization induced by ion exchange. The simulation of the processes was carried out for different values of β and γ parameters under all above-mentioned ion exchange conditions. As a result, the temperature dependences of the parameters, which provided a reliable coincidence of the simulation and experimental data, were found. This ensured the adequate modeling of the process of the glass-ceramics decrystallization in 520-580 oC temperature interval. Developed approach provides a powerful tool for fine tuning of the glass-ceramics structure, namely, concentration and average size of crystalline grains.Keywords: diffusion, glass-ceramics, ion exchange, vitrification
Procedia PDF Downloads 2693636 Tribological Properties of Non-Stick Coatings Used in Bread Baking Process
Authors: Maurice Brogly, Edwige Privas, Rajesh K. Gajendran, Sophie Bistac
Abstract:
Anti-sticky coatings based on perfluoroalkoxy (PFA) coatings are widely used in food processing industry especially for bread making. Their tribological performance, such as low friction coefficient, low surface energy and high heat resistance, make them an appropriate choice for anti-sticky coating application in moulds for food processing industry. This study is dedicated to evidence the transfer of contaminants from the coating due to wear and thermal ageing of the mould. The risk of contamination is induced by the damage of the coating by bread crust during the demoulding stage. The study focuses on the wear resistance and potential transfer of perfluorinated polymer from the anti-sticky coating. Friction between perfluorinated coating and bread crust is modeled by a tribological pin-on-disc test. The cellular nature of the bread crust is modeled by a polymer foam. FTIR analysis of the polymer foam after friction allow the evaluation of the transfer from the perfluorinated coating to polymer foam. Influence of thermal ageing on the physical, chemical and wear properties of the coating are also investigated. FTIR spectroscopic results show that the increase of PFA transfer onto the foam counterface is associated to the decrease of the friction coefficient. Increasing lubrication by film transfer results in the decrease of the friction coefficient. Moreover increasing the friction test parameters conditions (load, speed and sliding distance) also increase the film transfer onto the counterface. Thermal ageing increases the hydrophobic character of the PFA coating and thus also decreases the friction coefficient.Keywords: fluorobased polymer coatings, FTIR spectroscopy, non-stick food moulds, wear and friction
Procedia PDF Downloads 3303635 Biosensor Design through Molecular Dynamics Simulation
Authors: Wenjun Zhang, Yunqing Du, Steven W. Cranford, Ming L. Wang
Abstract:
The beginning of 21st century has witnessed new advancements in the design and use of new materials for biosensing applications, from nano to macro, protein to tissue. Traditional analytical methods lack a complete toolset to describe the complexities introduced by living systems, pathological relations, discrete hierarchical materials, cross-phase interactions, and structure-property dependencies. Materiomics – via systematic molecular dynamics (MD) simulation – can provide structure-process-property relations by using a materials science approach linking mechanisms across scales and enables oriented biosensor design. With this approach, DNA biosensors can be utilized to detect disease biomarkers present in individuals’ breath such as acetone for diabetes. Our wireless sensor array based on single-stranded DNA (ssDNA)-decorated single-walled carbon nanotubes (SWNT) has successfully detected trace amount of various chemicals in vapor differentiated by pattern recognition. Here, we present how MD simulation can revolutionize the way of design and screening of DNA aptamers for targeting biomarkers related to oral diseases and oral health monitoring. It demonstrates great potential to be utilized to build a library of DNDA sequences for reliable detection of several biomarkers of one specific disease, and as well provides a new methodology of creating, designing, and applying of biosensors.Keywords: biosensor, DNA, biomarker, molecular dynamics simulation
Procedia PDF Downloads 4633634 Digimesh Wireless Sensor Network-Based Real-Time Monitoring of ECG Signal
Authors: Sahraoui Halima, Dahani Ameur, Tigrine Abedelkader
Abstract:
DigiMesh technology represents a pioneering advancement in wireless networking, offering cost-effective and energy-efficient capabilities. Its inherent simplicity and adaptability facilitate the seamless transfer of data between network nodes, extending the range and ensuring robust connectivity through autonomous self-healing mechanisms. In light of these advantages, this study introduces a medical platform harnessed with DigiMesh wireless network technology characterized by low power consumption, immunity to interference, and user-friendly operation. The primary application of this platform is the real-time, long-distance monitoring of Electrocardiogram (ECG) signals, with the added capacity for simultaneous monitoring of ECG signals from multiple patients. The experimental setup comprises key components such as Raspberry Pi, E-Health Sensor Shield, and Xbee DigiMesh modules. The platform is composed of multiple ECG acquisition devices labeled as Sensor Node 1 and Sensor Node 2, with a Raspberry Pi serving as the central hub (Sink Node). Two communication approaches are proposed: Single-hop and multi-hop. In the Single-hop approach, ECG signals are directly transmitted from a sensor node to the sink node through the XBee3 DigiMesh RF Module, establishing peer-to-peer connections. This approach was tested in the first experiment to assess the feasibility of deploying wireless sensor networks (WSN). In the multi-hop approach, two sensor nodes communicate with the server (Sink Node) in a star configuration. This setup was tested in the second experiment. The primary objective of this research is to evaluate the performance of both Single-hop and multi-hop approaches in diverse scenarios, including open areas and obstructed environments. Experimental results indicate the DigiMesh network's effectiveness in Single-hop mode, with reliable communication over distances of approximately 300 meters in open areas. In the multi-hop configuration, the network demonstrated robust performance across approximately three floors, even in the presence of obstacles, without the need for additional router devices. This study offers valuable insights into the capabilities of DigiMesh wireless technology for real-time ECG monitoring in healthcare applications, demonstrating its potential for use in diverse medical scenarios.Keywords: DigiMesh protocol, ECG signal, real-time monitoring, medical platform
Procedia PDF Downloads 793633 Evaluation of PV Orientation Effect on Mismatch between Consumption Load and PV Power Profiles
Authors: Iyad M. Muslih, Yehya Abdellatif, Sara Qutishat
Abstract:
Renewable energy and in particular solar photovoltaic energy is emerging as a reasonable power generation source. The intermittent and unpredictable nature of solar energy can represent a serious challenge to the utility grids, specifically at relatively high penetration. To minimize the impact of PV power systems on the grid, self-consumption is encouraged. Self-consumption can be improved by matching the PV power generation with the electrical load consumption profile. This study will focus in studying different load profiles and comparing them to typical solar PV power generation at the selected sites with the purpose of analyzing the mismatch in consumption load profile for different users; residential, commercial, and industrial versus the solar photovoltaic output generation. The PV array orientation can be adjusted to reduce the mismatch effects. The orientation shift produces a corresponding shift in the energy production curve. This shift has a little effect on the mismatch for residential loads due to the fact the peak load occurs at night due to lighting loads. This minor gain does not justify the power production loss associated with the orientation shift. The orientation shift for both commercial and industrial cases lead to valuable decrease in the mismatch effects. Such a design is worth considering for reducing grid penetration. Furthermore, the proposed orientation shift yielded better results during the summer time due to the extended daylight hours.Keywords: grid impact, HOMER, power mismatch, solar PV energy
Procedia PDF Downloads 6043632 A Quality Index Optimization Method for Non-Invasive Fetal ECG Extraction
Authors: Lucia Billeci, Gennaro Tartarisco, Maurizio Varanini
Abstract:
Fetal cardiac monitoring by fetal electrocardiogram (fECG) can provide significant clinical information about the healthy condition of the fetus. Despite this potentiality till now the use of fECG in clinical practice has been quite limited due to the difficulties in its measuring. The recovery of fECG from the signals acquired non-invasively by using electrodes placed on the maternal abdomen is a challenging task because abdominal signals are a mixture of several components and the fetal one is very weak. This paper presents an approach for fECG extraction from abdominal maternal recordings, which exploits the characteristics of pseudo-periodicity of fetal ECG. It consists of devising a quality index (fQI) for fECG and of finding the linear combinations of preprocessed abdominal signals, which maximize these fQI (quality index optimization - QIO). It aims at improving the performances of the most commonly adopted methods for fECG extraction, usually based on maternal ECG (mECG) estimating and canceling. The procedure for the fECG extraction and fetal QRS (fQRS) detection is completely unsupervised and based on the following steps: signal pre-processing; maternal ECG (mECG) extraction and maternal QRS detection; mECG component approximation and canceling by weighted principal component analysis; fECG extraction by fQI maximization and fetal QRS detection. The proposed method was compared with our previously developed procedure, which obtained the highest at the Physionet/Computing in Cardiology Challenge 2013. That procedure was based on removing the mECG from abdominal signals estimated by a principal component analysis (PCA) and applying the Independent component Analysis (ICA) on the residual signals. Both methods were developed and tuned using 69, 1 min long, abdominal measurements with fetal QRS annotation of the dataset A provided by PhysioNet/Computing in Cardiology Challenge 2013. The QIO-based and the ICA-based methods were compared in analyzing two databases of abdominal maternal ECG available on the Physionet site. The first is the Abdominal and Direct Fetal Electrocardiogram Database (ADdb) which contains the fetal QRS annotations thus allowing a quantitative performance comparison, the second is the Non-Invasive Fetal Electrocardiogram Database (NIdb), which does not contain the fetal QRS annotations so that the comparison between the two methods can be only qualitative. In particular, the comparison on NIdb was performed defining an index of quality for the fetal RR series. On the annotated database ADdb the QIO method, provided the performance indexes Sens=0.9988, PPA=0.9991, F1=0.9989 overcoming the ICA-based one, which provided Sens=0.9966, PPA=0.9972, F1=0.9969. The comparison on NIdb was performed defining an index of quality for the fetal RR series. The index of quality resulted higher for the QIO-based method compared to the ICA-based one in 35 records out 55 cases of the NIdb. The QIO-based method gave very high performances with both the databases. The results of this study foresees the application of the algorithm in a fully unsupervised way for the implementation in wearable devices for self-monitoring of fetal health.Keywords: fetal electrocardiography, fetal QRS detection, independent component analysis (ICA), optimization, wearable
Procedia PDF Downloads 2803631 Impact of Traditional Male Circumcision Mishaps Towards Newly Initiated Men's Advancement in Education in South Africa
Authors: Thanduxolo Nomngcoyiya, Simon M. Kang’ethe
Abstract:
The aim of this article is to explore whether a relationship exists between traditional male circumcision mishaps and level of education in the Eastern Cape, South Africa, exemplified by an empirical case study. The study used qualitative paradigm; was exploratory in nature and used case study design that was descriptive and exploratory; and entailed interviewing twenty-eight (28) research participants comprising of eleven (11) newly initiated men and their families on one-on-one in-depth interviews, twelve (12) traditional nurses and community members in focus group discussions; and five (5) society key informants on key informant method. An interview guide served as a data collection instrument for focus group discussions, key informant method and in-depth interviews with unstructured open-ended questions. Findings indicated an array of traditional male circumcision (TMC) gaps, some of which were indicative of a relationship between the mishaps and level of education: the phenomenon of schooling became secondary in newly initiated men’s lives; TMC mishaps became a drawback towards the newly initiated men’s education progression; the newly initiated men are sacrificed at the altar of culture, and TMC mishaps ushered in socioeconomic setback to the newly initiated men. The study suggested that: TMC be developmental; TMC as a cultural endeavor be educational and human rights friendly; and the need to identify and integrate all other players with diverse specialties.Keywords: culture, education for all, EFA, millennium development goals, traditional male circumcision
Procedia PDF Downloads 1993630 Parallelization of Random Accessible Progressive Streaming of Compressed 3D Models over Web
Authors: Aayushi Somani, Siba P. Samal
Abstract:
Three-dimensional (3D) meshes are data structures, which store geometric information of an object or scene, generally in the form of vertices and edges. Current technology in laser scanning and other geometric data acquisition technologies acquire high resolution sampling which leads to high resolution meshes. While high resolution meshes give better quality rendering and hence is used often, the processing, as well as storage of 3D meshes, is currently resource-intensive. At the same time, web applications for data processing have become ubiquitous owing to their accessibility. For 3D meshes, the advancement of 3D web technologies, such as WebGL, WebVR, has enabled high fidelity rendering of huge meshes. However, there exists a gap in ability to stream huge meshes to a native client and browser application due to high network latency. Also, there is an inherent delay of loading WebGL pages due to large and complex models. The focus of our work is to identify the challenges faced when such meshes are streamed into and processed on hand-held devices, owing to its limited resources. One of the solutions that are conventionally used in the graphics community to alleviate resource limitations is mesh compression. Our approach deals with a two-step approach for random accessible progressive compression and its parallel implementation. The first step includes partition of the original mesh to multiple sub-meshes, and then we invoke data parallelism on these sub-meshes for its compression. Subsequent threaded decompression logic is implemented inside the Web Browser Engine with modification of WebGL implementation in Chromium open source engine. This concept can be used to completely revolutionize the way e-commerce and Virtual Reality technology works for consumer electronic devices. These objects can be compressed in the server and can be transmitted over the network. The progressive decompression can be performed on the client device and rendered. Multiple views currently used in e-commerce sites for viewing the same product from different angles can be replaced by a single progressive model for better UX and smoother user experience. Can also be used in WebVR for commonly and most widely used activities like virtual reality shopping, watching movies and playing games. Our experiments and comparison with existing techniques show encouraging results in terms of latency (compressed size is ~10-15% of the original mesh), processing time (20-22% increase over serial implementation) and quality of user experience in web browser.Keywords: 3D compression, 3D mesh, 3D web, chromium, client-server architecture, e-commerce, level of details, parallelization, progressive compression, WebGL, WebVR
Procedia PDF Downloads 1703629 Investigating the English Speech Processing System of EFL Japanese Older Children
Authors: Hiromi Kawai
Abstract:
This study investigates the nature of EFL older children’s L2 perceptive and productive abilities using classroom data, in order to find a pedagogical solution to the teaching of L2 sounds at an early stage of learning in a formal school setting. It is still inconclusive whether older children with only EFL formal school instruction at the initial stage of L2 learning are able to attain native-like perception and production in English within the very limited amount of exposure to the target language available. Based on the notion of the lack of study of EFL Japanese children’s acquisition of English segments, the researcher uses a model of L1 speech processing which was developed for investigating L1 English children’s speech and literacy difficulties using a psycholinguistic framework. The model is composed of input channel, output channel, and lexical representation, and examines how a child receives information from spoken or written language, remembers and stores it within the lexical representations and how the child selects and produces spoken or written words. Concerning language universality and language specificity in the language acquisitional process, the aim of finding any sound errors in L1 English children seemed to conform to the author’s intention to find abilities of English sounds in older Japanese children at the novice level of English in an EFL setting. 104 students in Grade 5 (between the ages of 10 and 11 years old) of an elementary school in Tokyo participated in this study. Four tests to measure their perceptive ability and three oral repetition tests to measure their productive ability were conducted with/without reference to lexical representation. All the test items were analyzed to calculate item facility (IF) indices, and correlational analyses and Structural Equation Modeling (SEM) were conducted to examine the relationship between the receptive ability and the productive ability. IF analysis showed that (1) the participants were better at perceiving a segment than producing a segment, (2) they had difficulty in auditory discrimination of paired consonants when one of them does not exist in the Japanese inventory, (3) they had difficulty in both perceiving and producing English vowels, and (4) their L1 loan word knowledge had an influence on their ability to perceive and produce L2 sounds. The result of the Multiple Regression Modeling showed that the two production tests could predict the participants’ auditory ability of real words in English. The result of SEM showed that the hypothesis that perceptive ability affects productive ability was supported. Based on these findings, the author discusses the possible explicit method of teaching English segments to EFL older children in a formal school setting.Keywords: EFL older children, english segments, perception, production, speech processing system
Procedia PDF Downloads 2433628 Roof Integrated Photo Voltaic with Air Collection on Glasgow School of Art Campus Building: A Feasibility Study
Authors: Rosalie Menon, Angela Reid
Abstract:
Building integrated photovoltaic systems with air collectors (hybrid PV-T) have proved successful however there are few examples of their application in the UK. The opportunity to pull heat from behind the PV system to contribute to a building’s heating system is an efficient use of waste energy and its potential to improve the performance of the PV array is well documented. As part of Glasgow School of Art’s estate expansion, the purchase and redevelopment of an existing 1950’s college building was used as a testing vehicle for the hybrid PV-T system as an integrated element of the upper floor and roof. The primary objective of the feasibility study was to determine if hybrid PV-T was technically and financially suitable for the refurbished building. The key consideration was whether the heat recovered from the PV panels (to increase the electrical efficiency) can be usefully deployed as a heat source within the building. Dynamic thermal modelling (IES) and RetScreen Software were used to carry out the feasibility study not only to simulate overshadowing and optimise the PV-T locations but also to predict the atrium temperature profile; predict the air load for the proposed new 4 No. roof mounted air handling units and to predict the dynamic electrical efficiency of the PV element. The feasibility study demonstrates that there is an energy reduction and carbon saving to be achieved with each hybrid PV-T option however the systems are subject to lengthy payback periods and highlights the need for enhanced government subsidy schemes to reward innovation with this technology in the UK.Keywords: building integrated, photovoltatic thermal, pre-heat air, ventilation
Procedia PDF Downloads 171