Search results for: matrix geometric method
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 20463

Search results for: matrix geometric method

19533 Developing a Sustainable Transit Planning Index Using Analytical Hierarchy Process Method for ZEB Implementation in Canada

Authors: Mona Ghafouri-Azar, Sara Diamond, Jeremy Bowes, Grace Yuan, Aimee Burnett, Michelle Wyndham-West, Sara Wagner, Anand Pariyarath

Abstract:

Transportation is the fastest growing source of greenhouse gas emissions worldwide. In Canada, it is responsible for 23% of total CO2emissions from fuel combustion, and emissions from the transportation sector are the second largest source of emissions after the oil and gas sector. Currently, most Canadian public transportation systems rely on buses that operateon fossil fuels.Canada is currently investing billions of dollars to replacediesel buses with electric busesas this isperceived to have a significant impact on climate mitigation. This paper focuses on the possible impacts of zero emission buses (ZEB) on sustainable development, considering three dimensions of sustainability; environmental quality, economic growth, and social development.A sustainable transportation system is one that is safe, affordable, accessible, efficient, and resilient and that contributes minimal emissions of carbon and other pollutants.To enable implementation of these goals, relevant indicators were selected and defined that measure progress towards a sustainable transportation system. These were drawn from Canadian and international examples. Studies compare different European cities in terms of development, sustainability, and infrastructures, by using transport performance indicators. A Normalized Transport Sustainability index measures and compares policies in different urban areas and allows fine-tuning of policies. Analysts use a number ofmethods for sustainable analysis, like cost-benefit analysis (CBA) toassess economic benefit, life-cycle assessment (LCA) to assess social, economic, and environment factors and goals, and multi-criteria decision making (MCDM) analysis which can comparediffering stakeholder preferences.A multi criteria decision making approach is an appropriate methodology to plan and evaluate sustainable transit development and to provide insights and meaningful information for decision makers and transit agencies. It is essential to develop a system thataggregates specific discrete indices to assess the sustainability of transportation systems.Theseprioritize indicators appropriate for the differentCanadian transit system agencies and theirpreferences and requirements. This studywill develop an integrating index that alliesexistingdiscrete indexes to supporta reliable comparison between the current transportation system (diesel buses) and the new ZEB system emerging in Canada. As a first step, theindexes for each category are selected, and the index matrix constructed. Second, the selected indicators arenormalized to remove anyinconsistency between them. Next, the normalized matrix isweighted based on the relative importance of each index to the main domains of sustainability using the analytical hierarchy process (AHP) method. This is accomplished through expert judgement around the relative importance of different attributes with respect to the goals through apairwise comparison matrix. The considerationof multiple environmental, economic, and social factors (including equity and health) is integrated intoa sustainable transit planning index (STPI) which supportsrealistic ZEB implementation in Canada and beyond and is useful to different stakeholders, agencies, and ministries.

Keywords: zero emission buses, sustainability, sustainable transit, transportation, analytical hierarchy process, environment, economy, social

Procedia PDF Downloads 116
19532 Building a Parametric Link between Mapping and Planning: A Sunlight-Adaptive Urban Green System Plan Formation Process

Authors: Chenhao Zhu

Abstract:

Quantitative mapping is playing a growing role in guiding urban planning, such as using a heat map created by CFX, CFD2000, or Envi-met, to adjust the master plan. However, there is no effective quantitative link between the mappings and planning formation. So, in many cases, the decision-making is still based on the planner's subjective interpretation and understanding of these mappings, which limits the improvement of scientific and accuracy brought by the quantitative mapping. Therefore, in this paper, an effort has been made to give a methodology of building a parametric link between the mapping and planning formation. A parametric planning process based on radiant mapping has been proposed for creating an urban green system. In the first step, a script is written in Grasshopper to build a road network and form the block, while the Ladybug Plug-in is used to conduct a radiant analysis in the form of mapping. Then, the research creatively transforms the radiant mapping from a polygon into a data point matrix, because polygon is hard to engage in the design formation. Next, another script is created to select the main green spaces from the road network based on the criteria of radiant intensity and connect the green spaces' central points to generate a green corridor. After that, a control parameter is introduced to adjust the corridor's form based on the radiant intensity. Finally, a green system containing greenspace and green corridor is generated under the quantitative control of the data matrix. The designer only needs to modify the control parameter according to the relevant research results and actual conditions to realize the optimization of the green system. This method can also be applied to much other mapping-based analysis, such as wind environment analysis, thermal environment analysis, and even environmental sensitivity analysis. The parameterized link between the mapping and planning will bring about a more accurate, objective, and scientific planning.

Keywords: parametric link, mapping, urban green system, radiant intensity, planning strategy, grasshopper

Procedia PDF Downloads 129
19531 Genetic Association and Functional Significance of Matrix Metalloproteinase-14 Promoter Variants rs1004030 and rs1003349 in Gallbladder Cancer Pathogenesis

Authors: J. Vinay , Kusumbati Besra, Niharika Pattnaik, Shivaram Prasad Singh, Manjusha Dixit

Abstract:

Gallbladder cancer (GBC) is rare but highly malignant cancer; its prevalence is more in certain geographical regions and ethnic groups, which include the Northern and Eastern states of India. Previous studies in India have reported genetic predisposition as one of the risk factors in GBC pathogenesis. Although the matrix metalloproteinase-14 (MMP14) is a well-known modulator of the tumor microenvironment and tumorigenesis and TCGA data also suggests its upregulation yet, its role in the genetic predisposition for GBC is completely unknown. We elucidated the role of MMP14 promoter variants as genetic risk factors and their implications in expression modulation. We screened MMP14 promoter variants association with GBC using Sanger’s sequencing in approximately 300 GBC and 300 control subjects and 26 GBC tissue samples of Indian ethnicity. The immunohistochemistry was used to check the MMP14 protein expression in GBC tissue samples. The role of promoter variants on expression levels was elucidated using a luciferase reporter assay. The variants rs1004030 (p-value = 0.0001) and rs1003349 (p-value = 0.0008) were significantly associated with gallbladder cancer. The luciferase assay in two different cell lines, HEK-293 (p = 0.0006) and TGBC1TKB (p = 0.0036) showed a significant increase in relative luciferase activity in the presence of risk alleles for both the single nucleotide polymorphisms (SNPs). Similarly, genotype-phenotype correlation in patients samples confirmed that the presence of risk alleles at rs1004030 and rs1003349 increased MMP14 expression. Overall, this study unravels the genetic association of MMP14 promoter variants with gallbladder cancer, which may contribute to pathogenesis by increasing its expression.

Keywords: gallbladder cancer, matrix metalloproteinase-14, single nucleotide polymorphism, case control study, genetic association study

Procedia PDF Downloads 168
19530 Design and Development of an Optimal Fault Tolerant 3 Degree of Freedom Robotic Manipulator

Authors: Ramish, Farhan Khalique Awan

Abstract:

Kinematic redundancy within the manipulators presents extended dexterity and manipulability to the manipulators. Redundant serial robotic manipulators are very popular in industries due to its competencies to keep away from singularities during normal operation and fault tolerance because of failure of one or more joints. Such fault tolerant manipulators are extraordinarily beneficial in applications where human interference for repair and overhaul is both impossible or tough; like in case of robotic arms for space programs, nuclear applications and so on. The design of this sort of fault tolerant serial 3 DoF manipulator is presented in this paper. This work was the extension of the author’s previous work of designing the simple 3R serial manipulator. This work is the realization of the previous design with optimizing the link lengths for incorporating the feature of fault tolerance. Various measures have been followed by the researchers to quantify the fault tolerance of such redundant manipulators. The fault tolerance in this work has been described in terms of the worst-case measure of relative manipulability that is, in fact, a local measure of optimization that works properly for certain configuration of the manipulators. An optimum fault tolerant Jacobian matrix has been determined first based on prescribed null space properties after which the link parameters have been described to meet the given Jacobian matrix. A solid model of the manipulator was then developed to realize the mathematically rigorous design. Further work was executed on determining the dynamic properties of the fault tolerant design and simulations of the movement for various trajectories have been carried out to evaluate the joint torques. The mathematical model of the system was derived via the Euler-Lagrange approach after which the same has been tested using the RoboAnalyzer© software. The results have been quite in agreement. From the CAD model and dynamic simulation data, the manipulator was fabricated in the workshop and Advanced Machining lab of NED University of Engineering and Technology.

Keywords: fault tolerant, Graham matrix, Jacobian, kinematics, Lagrange-Euler

Procedia PDF Downloads 210
19529 Durability of Wood Shavel Composites with Environmental Friendly Based Binder

Authors: Jul Endawati

Abstract:

The composite element of 20 mm in thickness were manufactured using high volume fly ash, silica fume as alternative hydraulic binders and Portland cement Type II. Pine wood shavel as by product of local small wood working industries were used as the composite filler. The elements were given in situ wet and dry treatment for 9 months. Visually there is no fiber degradation as a result of the interaction of the environment. The assessment were done to the elements bending strength and dimensional properties. Increase in MoR after 180 days of exposure shown that mechanically this degradation is not seen yet. The increment of MoR (213%) compare to that of 28 days might be affected by the formation of calcium hydroxide (CH) or ettringite in the transition zone. The use of pozzolan showed also a delay or minimize degradation of composites while improving the pore structure, and minimize the mineralization of the fiber bond with the cement matrix. The water absorption is 4,22% at 180 days, 7,94% at 120 days and 12,38% at 28 days, in line with the 68% decrease in Thickness Swelling (TS). This unoccured degradation could also be affected by the presence of silica fume in the binder matrix. After 270 days of exposure under tropical condition, the flexural strength started to decrease.

Keywords: durability, fly ash, natural fibre, silica fume

Procedia PDF Downloads 254
19528 Heroin and Opiates Metabolites Tracing by Gas-Chromatography Isotope Ratio Mass Spectrometry

Authors: Yao-Te Yen, Chao-Hsin Cheng, Meng-Shun Huang, Shan-Zong Cyue

Abstract:

'Poppy-seed defense' has been a serious problem all over the world, that is because the opiates metabolites in urine are difficult to distinguish where they come from precisely. In this research, a powerful analytic method has been developed to trace the opiates metabolites in urine by Gas-Chromatography Isotope Ratio Mass Spectrometry (GC-IRMS). In order to eliminate the interference of synthesis to heroin or metabolism through human body, opiates metabolites in urine and sized heroin were hydrolyzed to morphine. Morphine is the key compound for tracing between opiates metabolites and seized heroin in this research. By matching δ13C and δ15N values through morphine, it is successful to distinguish the opiates metabolites coming from heroin or medicine. We tested seven heroin abuser’s metabolites and seized heroin in crime sites, the result showed that opiates metabolites coming from seized heroin, the variation of δ13C and δ15N for morphine are within 0.2 and 2.5‰, respectively. The variation of δ13C and δ15N for morphine are reasonable with the result of matrix match experiments. Above all, the uncertainty of 'Poppy-seed defense' can be solved easily by this analytic method, it provides the direct evidence for judge to make accurate conviction without hesitation.

Keywords: poppy-seed defense, heroin, opiates metabolites, isotope ratio mass spectrometry

Procedia PDF Downloads 229
19527 Exploring Disruptive Innovation Capacity Effects on Firm Performance: An Investigation in Industries 4.0

Authors: Selma R. Oliveira, E. W. Cazarini

Abstract:

Recently, studies have referenced innovation as a key factor affecting the performance of firms. Companies make use of its innovative capacities to achieve sustainable competitive advantage. In this perspective, the objective of this paper is to contribute to innovation planning policies in industry 4.0. Thus, this paper examines the disruptive innovation capacity on firm performance in Europe. This procedure was prepared according to the following phases: Phase 1: Determination of the conceptual model; and Phase 2: Verification of the conceptual model. The research was initially conducted based on the specialized literature, which extracted the data regarding the constructs/structure and content in order to build the model. The research involved the intervention of experts knowledgeable on the object studied, selected by technical-scientific criteria. The data were extracted using an assessment matrix. To reduce subjectivity in the results achieved the following methods were used complementarily and in combination: multicriteria analysis, multivariate analysis, psychometric scaling and neurofuzzy technology. The data were extracted using an assessment matrix and the results were satisfactory, validating the modeling approach.

Keywords: disruptive innovation, capacity, performance, Industry 4.0

Procedia PDF Downloads 152
19526 Problems and Solutions in the Application of ICP-MS for Analysis of Trace Elements in Various Samples

Authors: Béla Kovács, Éva Bódi, Farzaneh Garousi, Szilvia Várallyay, Áron Soós, Xénia Vágó, Dávid Andrási

Abstract:

In agriculture for analysis of elements in different food and food raw materials, moreover environmental samples generally flame atomic absorption spectrometers (FAAS), graphite furnace atomic absorption spectrometers (GF-AAS), inductively coupled plasma optical emission spectrometers (ICP-OES) and inductively coupled plasma mass spectrometers (ICP-MS) are routinely applied. An inductively coupled plasma mass spectrometer (ICP-MS) is capable for analysis of 70-80 elements in multielemental mode, from 1-5 cm3 volume of a sample, moreover the detection limits of elements are in µg/kg-ng/kg (ppb-ppt) concentration range. All the analytical instruments have different physical and chemical interfering effects analysing the above types of samples. The smaller the concentration of an analyte and the larger the concentration of the matrix the larger the interfering effects. Nowadays there is very important to analyse growingly smaller concentrations of elements. From the above analytical instruments generally the inductively coupled plasma mass spectrometer is capable of analysing the smallest concentration of elements. The applied ICP-MS instrument has Collision Cell Technology (CCT) also. Using CCT mode certain elements have better (smaller) detection limits with 1-3 magnitudes comparing to a normal ICP-MS analytical method. The CCT mode has better detection limits mainly for analysis of selenium, arsenic, germanium, vanadium and chromium. To elaborate an analytical method for trace elements with an inductively coupled plasma mass spectrometer the most important interfering effects (problems) were evaluated: 1) Physical interferences; 2) Spectral interferences (elemental and molecular isobaric); 3) Effect of easily ionisable elements; 4) Memory interferences. Analysing food and food raw materials, moreover environmental samples an other (new) interfering effect emerged in ICP-MS, namely the effect of various matrixes having different evaporation and nebulization effectiveness, moreover having different quantity of carbon content of food and food raw materials, moreover environmental samples. In our research work the effect of different water-soluble compounds furthermore the effect of various quantity of carbon content (as sample matrix) were examined on changes of intensity of the applied elements. So finally we could find “opportunities” to decrease or eliminate the error of the analyses of applied elements (Cr, Co, Ni, Cu, Zn, Ge, As, Se, Mo, Cd, Sn, Sb, Te, Hg, Pb, Bi). To analyse these elements in the above samples, the most appropriate inductively coupled plasma mass spectrometer is a quadrupole instrument applying a collision cell technique (CCT). The extent of interfering effect of carbon content depends on the type of compounds. The carbon content significantly affects the measured concentration (intensities) of the above elements, which can be corrected using different internal standards.

Keywords: elements, environmental and food samples, ICP-MS, interference effects

Procedia PDF Downloads 490
19525 Poly(L-Lactic Acid) Scaffolds for Bone Tissue Engineering

Authors: Aleksandra BužArovska, Gordana Bogoeva Gaceva

Abstract:

Biodegradable polymers have received significant scientific attention in tissue engineering (TE) application, in particular their composites consisting of inorganic nanoparticles. In the last 15 years, they are subject of intensive research by many groups, aiming to develop polymer scaffolds with defined biodegradability, porosity and adequate mechanical stability. The most important characteristic making these materials attractive for TE is their biodegradability, a process that could be time controlled and long enough to enable generation of a new tissue as a replacement for the degraded polymer scaffold. In this work poly(L-lactic acid) scaffolds, filled with TiO2 nanoparticles functionalized with oleic acid, have been prepared by thermally induced phase separation method (TIPS). The functionalization of TiO2 nanoparticles with oleic acid was performed in order to improve the nanoparticles dispersibility within the polymer matrix and at the same time to inhibit the cytotoxicity of the nanofiller. The oleic acid was chosen as amphiphilic molecule belonging to the fatty acid family because of its non-toxicity and possibility for mediation between the hydrophilic TiO2 nanoparticles and hydrophobic PLA matrix. The produced scaffolds were characterized with thermogravimetric analysis (TGA), differential scanning calorimetry (DSC), scanning electron microscopy (SEM) and mechanical compression measurements. The bioactivity for bone tissue engineering application was tested in supersaturated simulated body fluid. The degradation process was followed by Fourier transform infrared spectroscopy (FTIR). The results showed anisotropic morphology with elongated open pores (100 µm), high porosity (around 92%) and perfectly dispersed nanofiller. The compression moduli up to 10 MPa were identified independent on the nanofiller content. Functionalized TiO2 nanoparticles induced formation of hydroxyapatite clusters as much as unfunctionalized TiO2. The prepared scaffolds showed properties ideal for scaffold vascularization, cell attachment, growth and proliferation.

Keywords: biodegradation, bone tissue engineering, mineralization, PLA scaffolds

Procedia PDF Downloads 256
19524 Computational Fluid Dynamics Based Analysis of Heat Exchanging Performance of Rotary Thermal Wheels

Authors: H. M. D. Prabhashana Herath, M. D. Anuradha Wickramasinghe, A. M. C. Kalpani Polgolla, R. A. C. Prasad Ranasinghe, M. Anusha Wijewardane

Abstract:

The demand for thermal comfort in buildings in hot and humid climates increases progressively. In general, buildings in hot and humid climates spend more than 60% of the total energy cost for the functionality of the air conditioning (AC) system. Hence, it is required to install energy efficient AC systems or integrate energy recovery systems for both new and/or existing AC systems whenever possible, to reduce the energy consumption by the AC system. Integrate a Rotary Thermal Wheel as the energy recovery device of an existing AC system has shown very promising with attractive payback periods of less than 5 years. A rotary thermal wheel can be located in the Air Handling Unit (AHU) of a central AC system to recover the energy available in the return air stream. During this study, a sensitivity analysis was performed using a CFD (Computational Fluid Dynamics) software to determine the optimum design parameters (i.e., rotary speed and parameters of the matrix profile) of a rotary thermal wheel for hot and humid climates. The simulations were performed for a sinusoidal matrix geometry. Variation of sinusoidal matrix parameters, i.e., span length and height, were also analyzed to understand the heat exchanging performance and the induced pressure drop due to the air flow. The results show that the heat exchanging performance increases when increasing the wheel rpm. However, the performance increment rate decreases when increasing the rpm. As a result, it is more advisable to operate the wheel at 10-20 rpm. For the geometry, it was found that the sinusoidal geometries with lesser spans and higher heights have higher heat exchanging capabilities. Considering the sinusoidal profiles analyzed during the study, the geometry with 4mm height and 3mm width shows better performance than the other combinations.

Keywords: air conditioning, computational fluid dynamics, CFD, energy recovery, heat exchangers

Procedia PDF Downloads 117
19523 Fuzzy Multi-Criteria Decision-Making Framework for Risk Management in Construction Supply Chain

Authors: Abdullah Ali Salamai

Abstract:

Risk management in the construction supply chain (CSC) is vital in construction project risks. CSC has various risks affecting product quality and project timeline, such as operational, social, financial, technical, design, and safety risks. These risks should be mitigated in project construction. So, this paper proposed a set of technologies to overcome risks in CSC, like artificial intelligence (AI), blockchain, data analytics, and IoT, to select the best one. So, the multi-criteria decision-making (MCDM) methodology is used to deal with various risks. The Multi-Attribute Utility Theory (MAUT) method is used to rank technologies. The weights of risks are obtained by the average method by using the decision matrix. The MCDM methodology is integrated with a fuzzy set to overcome uncertainty data. Experts used triangular fuzzy numbers to express their opinions instead of exact numbers. These allow the model to overcome inconsistent and vague data. The MCDM methodology was applied to 18 risks and 5 technologies. The results show that social risks have the highest weight. AI is the best technology for overcoming risks in CSC. AI can integrate with CSC from raw data to final products to deliver to the user.

Keywords: risk management, construction supply chain, fuzzy sets, multi-criteria decision making, supply chain management, artificial intelligence, blockchain

Procedia PDF Downloads 12
19522 A Structuring and Classification Method for Assigning Application Areas to Suitable Digital Factory Models

Authors: R. Hellmuth

Abstract:

The method of factory planning has changed a lot, especially when it is about planning the factory building itself. Factory planning has the task of designing products, plants, processes, organization, areas, and the building of a factory. Regular restructuring is becoming more important in order to maintain the competitiveness of a factory. Restrictions in new areas, shorter life cycles of product and production technology as well as a VUCA world (Volatility, Uncertainty, Complexity and Ambiguity) lead to more frequent restructuring measures within a factory. A digital factory model is the planning basis for rebuilding measures and becomes an indispensable tool. Furthermore, digital building models are increasingly being used in factories to support facility management and manufacturing processes. The main research question of this paper is, therefore: What kind of digital factory model is suitable for the different areas of application during the operation of a factory? First, different types of digital factory models are investigated, and their properties and usabilities for use cases are analysed. Within the scope of investigation are point cloud models, building information models, photogrammetry models, and these enriched with sensor data are examined. It is investigated which digital models allow a simple integration of sensor data and where the differences are. Subsequently, possible application areas of digital factory models are determined by means of a survey and the respective digital factory models are assigned to the application areas. Finally, an application case from maintenance is selected and implemented with the help of the appropriate digital factory model. It is shown how a completely digitalized maintenance process can be supported by a digital factory model by providing information. Among other purposes, the digital factory model is used for indoor navigation, information provision, and display of sensor data. In summary, the paper shows a structuring of digital factory models that concentrates on the geometric representation of a factory building and its technical facilities. A practical application case is shown and implemented. Thus, the systematic selection of digital factory models with the corresponding application cases is evaluated.

Keywords: building information modeling, digital factory model, factory planning, maintenance

Procedia PDF Downloads 100
19521 Fe Modified Tin Oxide Thin Film Based Matrix for Reagentless Uric Acid Biosensing

Authors: Kashima Arora, Monika Tomar, Vinay Gupta

Abstract:

Biosensors have found potential applications ranging from environmental testing and biowarfare agent detection to clinical testing, health care, and cell analysis. This is driven in part by the desire to decrease the cost of health care and to obtain precise information more quickly about the health status of patient by the development of various biosensors, which has become increasingly prevalent in clinical testing and point of care testing for a wide range of biological elements. Uric acid is an important byproduct in human body and a number of pathological disorders are related to its high concentration in human body. In past few years, rapid growth in the development of new materials and improvements in sensing techniques have led to the evolution of advanced biosensors. In this context, metal oxide thin film based matrices due to their bio compatible nature, strong adsorption ability, high isoelectric point (IEP) and abundance in nature have become the materials of choice for recent technological advances in biotechnology. In the past few years, wide band-gap metal oxide semiconductors including ZnO, SnO₂ and CeO₂ have gained much attention as a matrix for immobilization of various biomolecules. Tin oxide (SnO₂), wide band gap semiconductor (Eg =3.87 eV), despite having multifunctional properties for broad range of applications including transparent electronics, gas sensors, acoustic devices, UV photodetectors, etc., it has not been explored much for biosensing purpose. To realize a high performance miniaturized biomolecular electronic device, rf sputtering technique is considered to be the most promising for the reproducible growth of good quality thin films, controlled surface morphology and desired film crystallization with improved electron transfer property. Recently, iron oxide and its composites have been widely used as matrix for biosensing application which exploits the electron communication feature of Fe, for the detection of various analytes using urea, hemoglobin, glucose, phenol, L-lactate, H₂O₂, etc. However, to the authors’ knowledge, no work is being reported on modifying the electronic properties of SnO₂ by implanting with suitable metal (Fe) to induce the redox couple in it and utilizing it for reagentless detection of uric acid. In present study, Fe implanted SnO₂ based matrix has been utilized for reagentless uric acid biosensor. Implantation of Fe into SnO₂ matrix is confirmed by energy-dispersive X-Ray spectroscopy (EDX) analysis. Electrochemical techniques have been used to study the response characteristics of Fe modified SnO₂ matrix before and after uricase immobilization. The developed uric acid biosensor exhibits a high sensitivity to about 0.21 mA/mM and a linear variation in current response over concentration range from 0.05 to 1.0 mM of uric acid besides high shelf life (~20 weeks). The Michaelis-Menten kinetic parameter (Km) is found to be relatively very low (0.23 mM), which indicates high affinity of the fabricated bioelectrode towards uric acid (analyte). Also, the presence of other interferents present in human serum has negligible effect on the performance of biosensor. Hence, obtained results highlight the importance of implanted Fe:SnO₂ thin film as an attractive matrix for realization of reagentless biosensors towards uric acid.

Keywords: Fe implanted tin oxide, reagentless uric acid biosensor, rf sputtering, thin film

Procedia PDF Downloads 171
19520 Performance of Buildings with Base-Isolation System under Geometric Irregularities

Authors: Firoz Alam Faroque, Ankur Neog

Abstract:

Earthquake causes significant loss of lives and severe damage to infrastructure. Base isolator is one of the most suitable solutions to make a building earthquake resistant. Base isolation consists of installing an isolator along with the steel plates covered with pads of strong material like steel, rubber, etc. In our study, we have used lead rubber bearing (LRB). The basic idea of seismic isolation is based on the reduction of the earthquake-induced inertia forces by shifting the fundamental period of the structure out of dangerous resonance range, and concentration of the deformation and energy dissipation demands at the isolation and energy dissipation systems, which are designed for this purpose. In this paper, RC frame buildings have been modeled and analyzed by response spectrum method using ETABS software. The LRB used in the model is designed as per uniform building code (UBC) 97. It is found that time period for the base isolated structures are higher than that of the fixed base structure and the value of base shear significantly reduces in the case of base-isolated buildings. It has also been found that buildings with vertical irregularities give better performance as compared to building with plan irregularities using base isolators.

Keywords: base isolation, base shear, irregularities in buildings, lead rubber bearing (LRB)

Procedia PDF Downloads 316
19519 The Foundation Binary-Signals Mechanics and Actual-Information Model of Universe

Authors: Elsadig Naseraddeen Ahmed Mohamed

Abstract:

In contrast to the uncertainty and complementary principle, it will be shown in the present paper that the probability of the simultaneous occupation event of any definite values of coordinates by any definite values of momentum and energy at any definite instance of time can be described by a binary definite function equivalent to the difference between their numbers of occupation and evacuation epochs up to that time and also equivalent to the number of exchanges between those occupation and evacuation epochs up to that times modulus two, these binary definite quantities can be defined at all point in the time’s real-line so it form a binary signal represent a complete mechanical description of physical reality, the time of these exchanges represent the boundary of occupation and evacuation epochs from which we can calculate these binary signals using the fact that the time of universe events actually extends in the positive and negative of time’s real-line in one direction of extension when these number of exchanges increase, so there exists noninvertible transformation matrix can be defined as the matrix multiplication of invertible rotation matrix and noninvertible scaling matrix change the direction and magnitude of exchange event vector respectively, these noninvertible transformation will be called actual transformation in contrast to information transformations by which we can navigate the universe’s events transformed by actual transformations backward and forward in time’s real-line, so these information transformations will be derived as an elements of a group can be associated to their corresponded actual transformations. The actual and information model of the universe will be derived by assuming the existence of time instance zero before and at which there is no coordinate occupied by any definite values of momentum and energy, and then after that time, the universe begin its expanding in spacetime, this assumption makes the need for the existence of Laplace’s demon who at one moment can measure the positions and momentums of all constituent particle of the universe and then use the law of classical mechanics to predict all future and past of universe’s events, superfluous, we only need for the establishment of our analog to digital converters to sense the binary signals that determine the boundaries of occupation and evacuation epochs of the definite values of coordinates relative to its origin by the definite values of momentum and energy as present events of the universe from them we can predict approximately in high precision it's past and future events.

Keywords: binary-signal mechanics, actual-information model of the universe, actual-transformation, information-transformation, uncertainty principle, Laplace's demon

Procedia PDF Downloads 168
19518 Calycosin Ameliorates Osteoarthritis by Regulating the Imbalance Between Chondrocyte Synthesis and Catabolism

Authors: Hong Su, Qiuju Yan, Wei Du, En Hu, Zhaoyu Yang, Wei Zhang, Yusheng Li, Tao Tang, Wang yang, Shushan Zhao

Abstract:

Osteoarthritis (OA) is a severe chronic inflammatory disease. As the main active component of Astragalus mongholicus Bunge, a classic traditional ethnic herb, calycosin exhibits anti-inflammatory action and its mechanism of exact targets for OA have yet to be determined. In this study, we established an anterior cruciate ligament transection (ACLT) mouse model. Mice were randomized to sham, OA, and calycosin groups. Cartilage synthesis markers type II collagen (Col-2) and SRY-Box Transcription Factor 9 (Sox-9) increased significantly after calycosin gavage. While cartilage matrix degradation index cyclooxygenase-2 (COX-2), phosphor-epidermal growth factor receptor (p-EGFR), and matrix metalloproteinase-9 (MMP9) expression were decreased. With the help of network pharmacology and molecular docking, these results were confirmed in chondrocyte ATDC5 cells. Our results indicated that the calycosin treatment significantly improved cartilage damage, this was probably attributed to reversing the imbalance between chondrocyte synthesis and catabolism.

Keywords: calycosin, osteoarthritis, network pharmacology, molecular docking, inflammatory, cyclooxygenase 2

Procedia PDF Downloads 89
19517 A Hybrid Method for Determination of Effective Poles Using Clustering Dominant Pole Algorithm

Authors: Anuj Abraham, N. Pappa, Daniel Honc, Rahul Sharma

Abstract:

In this paper, an analysis of some model order reduction techniques is presented. A new hybrid algorithm for model order reduction of linear time invariant systems is compared with the conventional techniques namely Balanced Truncation, Hankel Norm reduction and Dominant Pole Algorithm (DPA). The proposed hybrid algorithm is known as Clustering Dominant Pole Algorithm (CDPA) is able to compute the full set of dominant poles and its cluster center efficiently. The dominant poles of a transfer function are specific eigenvalues of the state space matrix of the corresponding dynamical system. The effectiveness of this novel technique is shown through the simulation results.

Keywords: balanced truncation, clustering, dominant pole, Hankel norm, model reduction

Procedia PDF Downloads 589
19516 Using Google Distance Matrix Application Programming Interface to Reveal and Handle Urban Road Congestion Hot Spots: A Case Study from Budapest

Authors: Peter Baji

Abstract:

In recent years, a growing body of literature emphasizes the increasingly negative impacts of urban road congestion in the everyday life of citizens. Although there are different responses from the public sector to decrease traffic congestion in urban regions, the most effective public intervention is using congestion charges. Because travel is an economic asset, its consumption can be controlled by extra taxes or prices effectively, but this demand-side intervention is often unpopular. Measuring traffic flows with the help of different methods has a long history in transport sciences, but until recently, there was not enough sufficient data for evaluating road traffic flow patterns on the scale of an entire road system of a larger urban area. European cities (e.g., London, Stockholm, Milan), in which congestion charges have already been introduced, designated a particular zone in their downtown for paying, but it protects only the users and inhabitants of the CBD (Central Business District) area. Through the use of Google Maps data as a resource for revealing urban road traffic flow patterns, this paper aims to provide a solution for a fairer and smarter congestion pricing method in cities. The case study area of the research contains three bordering districts of Budapest which are linked by one main road. The first district (5th) is the original downtown that is affected by the congestion charge plans of the city. The second district (13th) lies in the transition zone, and it has recently been transformed into a new CBD containing the biggest office zone in Budapest. The third district (4th) is a mainly residential type of area on the outskirts of the city. The raw data of the research was collected with the help of Google’s Distance Matrix API (Application Programming Interface) which provides future estimated traffic data via travel times between freely fixed coordinate pairs. From the difference of free flow and congested travel time data, the daily congestion patterns and hot spots are detectable in all measured roads within the area. The results suggest that the distribution of congestion peak times and hot spots are uneven in the examined area; however, there are frequently congested areas which lie outside the downtown and their inhabitants also need some protection. The conclusion of this case study is that cities can develop a real-time and place-based congestion charge system that forces car users to avoid frequently congested roads by changing their routes or travel modes. This would be a fairer solution for decreasing the negative environmental effects of the urban road transportation instead of protecting a very limited downtown area.

Keywords: Budapest, congestion charge, distance matrix API, application programming interface, pilot study

Procedia PDF Downloads 184
19515 Mode II Fracture Toughness of Hybrid Fiber Reinforced Concrete

Authors: H. S. S Abou El-Mal, A. S. Sherbini, H. E. M. Sallam

Abstract:

Mode II fracture toughness (KIIc) of fiber reinforced concrete has been widely investigated under various patterns of testing geometries. The effect of fiber type, concrete matrix properties, and testing mechanisms were extensively studied. The area of hybrid fiber addition shows a lake of reported research data. In this paper an experimental investigation of hybrid fiber embedded in high strength concrete matrix is reported. Three different types of fibers; namely steel (S), glass (G), and polypropylene (PP) fibers were mixed together in four hybridization patterns, (S/G), (S/PP), (G/PP), (S/G/PP) with constant cumulative volume fraction (Vf) of 1.5%. The concrete matrix properties were kept the same for all hybrid fiber reinforced concrete patterns. In an attempt to estimate a fairly accepted value of fracture toughness KIIc, four testing geometries and loading types are employed in this investigation. Four point shear, Brazilian notched disc, double notched cube, and double edge notched specimens are investigated in a trial to avoid the limitations and sensitivity of each test regarding geometry, size effect, constraint condition, and the crack length to specimen width ratio a/w. The addition of all hybridization patterns of fiber reduced the compressive strength and increased mode II fracture toughness in pure mode II tests. Mode II fracture toughness of concrete KIIc decreased with the increment of a/w ratio for all concretes and test geometries. Mode II fracture toughness KIIc is found to be sensitive to the hybridization patterns of fiber. The (S/PP) hybridization pattern showed higher values than all other patterns, while the (S/G/PP) showed insignificant enhancement on mode II fracture toughness (KIIc). Four point shear (4PS) test set up reflects the most reliable values of mode II fracture toughness KIIc of concrete. Mode II fracture toughness KIIc of concrete couldn’t be assumed as a real material property.

Keywords: fiber reinforced concrete, Hybrid fiber, Mode II fracture toughness, testing geometry

Procedia PDF Downloads 319
19514 Constant Order Predictor Corrector Method for the Solution of Modeled Problems of First Order IVPs of ODEs

Authors: A. A. James, A. O. Adesanya, M. R. Odekunle, D. G. Yakubu

Abstract:

This paper examines the development of one step, five hybrid point method for the solution of first order initial value problems. We adopted the method of collocation and interpolation of power series approximate solution to generate a continuous linear multistep method. The continuous linear multistep method was evaluated at selected grid points to give the discrete linear multistep method. The method was implemented using a constant order predictor of order seven over an overlapping interval. The basic properties of the derived corrector was investigated and found to be zero stable, consistent and convergent. The region of absolute stability was also investigated. The method was tested on some numerical experiments and found to compete favorably with the existing methods.

Keywords: interpolation, approximate solution, collocation, differential system, half step, converges, block method, efficiency

Procedia PDF Downloads 322
19513 Exploring the Connectedness of Ad Hoc Mesh Networks in Rural Areas

Authors: Ibrahim Obeidat

Abstract:

Reaching a fully-connected network of mobile nodes in rural areas got a great attention between network researchers. This attention rose due to the complexity and high costs while setting up the needed infrastructures for these networks, in addition to the low transmission range these nodes has. Terranet technology, as an example, employs ad-hoc mesh network where each node has a transmission range not exceed one kilometer, this means that every two nodes are able to communicate with each other if they are just one kilometer far from each other, otherwise a third-party will play the role of the “relay”. In Terranet, and as an idea to reduce network setup cost, every node in the network will be considered as a router that is responsible of forwarding data between other nodes which result in a decentralized collaborative environment. Most researches on Terranet presents the idea of how to encourage mobile nodes to become more cooperative by letting their devices in “ON” state as long as possible while accepting to play the role of relay (router). This research presents the issue of finding the percentage of nodes in ad-hoc mesh network within rural areas that should play the role of relay at every time slot, relating to what is the actual area coverage of nodes in order to have the network reach the fully-connectivity. Far from our knowledge, till now there is no current researches discussed this issue. The research is done by making an implementation that depends on building adjacency matrix as an indicator to the connectivity between network members. This matrix is continually updated until each value in it refers to the number of hubs that should be followed to reach from one node to another. After repeating the algorithm on different area sizes, different coverage percentages for each size, and different relay percentages for several times, results extracted shows that for area coverage less than 5% we need to have 40% of the nodes to be relays, where 10% percentage is enough for areas with node coverage greater than 5%.

Keywords: ad-hoc mesh networks, network connectivity, mobile ad-hoc networks, Terranet, adjacency matrix, simulator, wireless sensor networks, peer to peer networks, vehicular Ad hoc networks, relay

Procedia PDF Downloads 266
19512 Low Frequency Ultrasonic Degassing to Reduce Void Formation in Epoxy Resin and Its Effect on the Thermo-Mechanical Properties of the Cured Polymer

Authors: A. J. Cobley, L. Krishnan

Abstract:

The demand for multi-functional lightweight materials in sectors such as automotive, aerospace, electronics is growing, and for this reason fibre-reinforced, epoxy polymer composites are being widely utilized. The fibre reinforcing material is mainly responsible for the strength and stiffness of the composites whilst the main role of the epoxy polymer matrix is to enhance the load distribution applied on the fibres as well as to protect the fibres from the effect of harmful environmental conditions. The superior properties of the fibre-reinforced composites are achieved by the best properties of both of the constituents. Although factors such as the chemical nature of the epoxy and how it is cured will have a strong influence on the properties of the epoxy matrix, the method of mixing and degassing of the resin can also have a significant impact. The production of a fibre-reinforced epoxy polymer composite will usually begin with the mixing of the epoxy pre-polymer with a hardener and accelerator. Mechanical methods of mixing are often employed for this stage but such processes naturally introduce air into the mixture, which, if it becomes entrapped, will lead to voids in the subsequent cured polymer. Therefore, degassing is normally utilised after mixing and this is often achieved by placing the epoxy resin mixture in a vacuum chamber. Although this is reasonably effective, it is another process stage and if a method of mixing could be found that, at the same time, degassed the resin mixture this would lead to shorter production times, more effective degassing and less voids in the final polymer. In this study the effect of four different methods for mixing and degassing of the pre-polymer with hardener and accelerator were investigated. The first two methods were manual stirring and magnetic stirring which were both followed by vacuum degassing. The other two techniques were ultrasonic mixing/degassing using a 40 kHz ultrasonic bath and a 20 kHz ultrasonic probe. The cured cast resin samples were examined under scanning electron microscope (SEM), optical microscope, and Image J analysis software to study morphological changes, void content and void distribution. Three point bending test and differential scanning calorimetry (DSC) were also performed to determine the thermal and mechanical properties of the cured resin. It was found that the use of the 20 kHz ultrasonic probe for mixing/degassing gave the lowest percentage voids of all the mixing methods in the study. In addition, the percentage voids found when employing a 40 kHz ultrasonic bath to mix/degas the epoxy polymer mixture was only slightly higher than when magnetic stirrer mixing followed by vacuum degassing was utilized. The effect of ultrasonic mixing/degassing on the thermal and mechanical properties of the cured resin will also be reported. The results suggest that low frequency ultrasound is an effective means of mixing/degassing a pre-polymer mixture and could enable a significant reduction in production times.

Keywords: degassing, low frequency ultrasound, polymer composites, voids

Procedia PDF Downloads 292
19511 Geometric Properties of Some q-Bessel Functions

Authors: İbrahim Aktaş, Árpád Baricz

Abstract:

In this paper, the radii of star likeness of the Jackson and Hahn-Exton q-Bessel functions are considered, and for each of them three different normalizations is applied. By applying Euler-Rayleigh inequalities for the first positive zeros of these functions tight lower, and upper bounds for the radii of starlikeness of these functions are obtained. The Laguerre-Pólya class of real entire functions plays an important role in this study. In particular, we obtain some new bounds for the first positive zero of the derivative of the classical Bessel function of the first kind.

Keywords: bessel function, lommel function, radius of starlikeness and convexity, Struve function

Procedia PDF Downloads 268
19510 Novel Recommender Systems Using Hybrid CF and Social Network Information

Authors: Kyoung-Jae Kim

Abstract:

Collaborative Filtering (CF) is a popular technique for the personalization in the E-commerce domain to reduce information overload. In general, CF provides recommending items list based on other similar users’ preferences from the user-item matrix and predicts the focal user’s preference for particular items by using them. Many recommender systems in real-world use CF techniques because it’s excellent accuracy and robustness. However, it has some limitations including sparsity problems and complex dimensionality in a user-item matrix. In addition, traditional CF does not consider the emotional interaction between users. In this study, we propose recommender systems using social network and singular value decomposition (SVD) to alleviate some limitations. The purpose of this study is to reduce the dimensionality of data set using SVD and to improve the performance of CF by using emotional information from social network data of the focal user. In this study, we test the usability of hybrid CF, SVD and social network information model using the real-world data. The experimental results show that the proposed model outperforms conventional CF models.

Keywords: recommender systems, collaborative filtering, social network information, singular value decomposition

Procedia PDF Downloads 276
19509 Ascidian Styela rustica Proteins’ Structural Domains Predicted to Participate in the Tunic Formation

Authors: M. I. Tyletc, O. I. Podgornya, T. G. Shaposhnikova, S. V. Shabelnikov, A. G. Mittenberg, M. A. Daugavet

Abstract:

Ascidiacea is the most numerous class of the Tunicata subtype. These chordates' distinctive feature of the anatomical structure is a tunic consisting of cellulose fibrils, protein molecules, and single cells. The mechanisms of the tunic formation are not known in detail; tunic formation could be used as the model system for studying the interaction of cells with the extracellular matrix. Our model species is the ascidian Styela rustica, which is prevalent in benthic communities of the White Sea. As previously shown, the tunic formation involves morula blood cells, which contain the major 48 kDa protein p48. P48 participation in the tunic formation was proved using antibodies against the protein. The nature of the protein and its function remains unknown. The current research aims to determine the amino acid sequence of p48, as well as to clarify its role in the tunic formation. The peptides that make up the p48 amino acid sequence were determined by mass spectrometry. A search for peptides in protein sequence databases identified sequences homologous to p48 in Styela clava, Styela plicata, and Styela canopus. Based on sequence alignment, their level of similarity was determined as 81-87%. The correspondent sequence of ascidian Styela canopus was used for further analysis. The Styela rustica p48 sequence begins with a signal peptide, which could indicate that the protein is secretory. This is consistent with experimentally obtained data: the contents of morula cells secreted in the tunic matrix. The isoelectric point of p48 is 9.77, which is consistent with the experimental results of acid electrophoresis of morula cell proteins. However, the molecular weight of the amino acid sequence of ascidian Styela canopus is 103 kDa, so p48 of Styela rustica is a shorter homolog. The search for conservative functional domains revealed the presence of two Ca-binding EGF-like domains, thrombospondin (TSP1) and tyrosinase domains. The p48 peptides determined by mass spectrometry fall into the region of the sequence corresponding to the last two domains and have amino acid substitutions as compared to Styela canopus homolog. The tyrosinase domain (pfam00264) is known to be part of the phenoloxidase enzyme, which participates in melanization processes and the immune response. The thrombospondin domain (smart00209) interacts with a wide range of proteins, and is involved in several biological processes, including coagulation, cell adhesion, modulation of intercellular and cell-matrix interactions, angiogenesis, wound healing and tissue remodeling. It can be assumed that the tyrosinase domain in p48 plays the role of the phenoloxidase enzyme, and TSP1 provides a link between the extracellular matrix and cell surface receptors, and may also be responsible for the repair of the tunic. The results obtained are consistent with experimental data on p48. The domain organization of protein suggests that p48 is an enzyme involved in the tunic tunning and is an important regulator of the organization of the extracellular matrix.

Keywords: ascidian, p48, thrombospondin, tyrosinase, tunic, tunning

Procedia PDF Downloads 100
19508 Student Loan Debt among Students with Disabilities

Authors: Kaycee Bills

Abstract:

This study will determine if students with disabilities have higher student loan debt payments than other student populations. The hypothesis was that students with disabilities would have significantly higher student loan debt payments than other students due to the length of time they spend in school. Using the Bachelorette and Beyond Study Wave 2015/017 dataset, quantitative methods were employed. These data analysis methods included linear regression and a correlation matrix. Due to the exploratory nature of the study, the significance levels for the overall model and each variable were set at .05. The correlation matrix demonstrated that students with certain types of disabilities are more likely to fall under higher student loan payment brackets than students without disabilities. These results also varied among the different types of disabilities. The result of the overall linear regression model was statistically significant (p = .04). Despite the overall model being statistically significant, the majority of the significance values for the different types of disabilities were null. However, several other variables had statistically significant results, such as veterans, people of minority races, and people who attended private schools. Implications for how this impacts the economy, capitalism, and financial wellbeing of various students are discussed.

Keywords: disability, student loan debt, higher education, social work

Procedia PDF Downloads 157
19507 Mueller Matrix Polarimetry for Analysis Scattering Biological Fluid Media

Authors: S. Cherif, A. Medjahed, M. Bouafia, A. Manallah

Abstract:

A light wave is characterized by 4 characteristics: its amplitude, its frequency, its phase and the direction of polarization of its luminous vector (the electric field). It is in this last characteristic that we will be interested. The polarization of the light was introduced in order to describe the vectorial behavior of the light; it describes the way in which the electric field evolves in a point of space. Our work consists in studying diffusing mediums. Different types of biological fluids were selected to study the evolution of each with increasing scattering power of the medium, and in the same time to make a comparison between them. When crossing these mediums, the light undergoes modifications and/or deterioration of its initial state of polarization. This phenomenon is related to the properties of the medium, the idea is to compare the characteristics of the entering and outgoing light from the studied medium by a white light. The advantage of this model is that it is experimentally accessible workable intensity measurements with CCD sensors and allows operation in 2D. The latter information is used to discriminate some physical properties of the studied areas. We chose four types of milk to study the evolution of each with increasing scattering power of the medium.

Keywords: light polarization, Mueller matrix, Mueller images, diffusing medium, milk

Procedia PDF Downloads 321
19506 Motion Planning and Simulation Design of a Redundant Robot for Sheet Metal Bending Processes

Authors: Chih-Jer Lin, Jian-Hong Hou

Abstract:

Industry 4.0 is a vision of integrated industry implemented by artificial intelligent computing, software, and Internet technologies. The main goal of industry 4.0 is to deal with the difficulty owing to competitive pressures in the marketplace. For today’s manufacturing factories, the type of production is changed from mass production (high quantity production with low product variety) to medium quantity-high variety production. To offer flexibility, better quality control, and improved productivity, robot manipulators are used to combine material processing, material handling, and part positioning systems into an integrated manufacturing system. To implement the automated system for sheet metal bending operations, motion planning of a 7-degrees of freedom (DOF) robot is studied in this paper. A virtual reality (VR) environment of a bending cell, which consists of the robot and a bending machine, is established using the virtual robot experimentation platform (V-REP) simulator. For sheet metal bending operations, the robot only needs six DOFs for the pick-and-place or tracking tasks. Therefore, this 7 DOF robot has more DOFs than the required to execute a specified task; it can be called a redundant robot. Therefore, this robot has kinematic redundancies to deal with the task-priority problems. For redundant robots, Pseudo-inverse of the Jacobian is the most popular motion planning method, but the pseudo-inverse methods usually lead to a kind of chaotic motion with unpredictable arm configurations as the Jacobian matrix lose ranks. To overcome the above problem, we proposed a method to formulate the motion planning problems as optimization problem. Moreover, a genetic algorithm (GA) based method is proposed to deal with motion planning of the redundant robot. Simulation results validate the proposed method feasible for motion planning of the redundant robot in an automated sheet-metal bending operations.

Keywords: redundant robot, motion planning, genetic algorithm, obstacle avoidance

Procedia PDF Downloads 134
19505 Improvement of the Geometric of Dental Bridge Framework through Automatic Program

Authors: Rong-Yang Lai, Jia-Yu Wu, Chih-Han Chang, Yung-Chung Chen

Abstract:

The dental bridge is one of the clinical methods of the treatment for missing teeth. The dental bridge is generally designed for two layers, containing the inner layer of the framework(zirconia) and the outer layer of the porcelain-fused to framework restorations. The design of a conventional bridge is generally based on the antagonist tooth profile so that the framework evenly indented by an equal thickness from outer contour. All-ceramic dental bridge made of zirconia have well demonstrated remarkable potential to withstand a higher physiological occlusal load in posterior region, but it was found that there is still the risk of all-ceramic bridge failure in five years. Thus, how to reduce the incidence of failure is still a problem to be solved. Therefore, the objective of this study is to develop mechanical designs for all-ceramic dental bridges framework by reducing the stress and enhancing fracture resistance under given loading conditions by finite element method. In this study, dental design software is used to design dental bridge based on tooth CT images. After building model, Bi-directional Evolutionary Structural Optimization (BESO) Method algorithm implemented in finite element software was employed to analyze results of finite element software and determine the distribution of the materials in dental bridge; BESO searches the optimum distribution of two different materials, namely porcelain and zirconia. According to the previous calculation of the stress value of each element, when the element stress value is higher than the threshold value, the element would be replaced by the framework material; besides, the difference of maximum stress peak value is less than 0.1%, calculation is complete. After completing the design of dental bridge, the stress distribution of the whole structure is changed. BESO reduces the peak values of principle stress of 10% in outer-layer porcelain and avoids producing tensile stress failure.

Keywords: dental bridge, finite element analysis, framework, automatic program

Procedia PDF Downloads 272
19504 Silver-Doped Magnetite Titanium Oxide Nanoparticles for Photocatalytic Degradation of Organic Pollutants

Authors: Hanna Abbo, Siyasanga Noganta, Salam Titinchi

Abstract:

The global lack of clean water for human sanitation and other purposes has become an emerging dilemma for human beings. The presence of organic pollutants in wastewater produced by textile industries, leather manufacturing and chemical industries is an alarming matter for a safe environment and human health. For the last decades, conventional methods have been applied for the purification of water but due to industrialization these methods fall short. Advanced oxidation processes and their reliable application in degradation of many contaminants have been reported as a potential method to reduce and/or alleviate this problem. Lately it has been assumed that incorporation of some metal nanoparticles such as magnetite nanoparticles as photocatalyst for Fenton reaction which could improve the degradation efficiency of contaminants. Core/shell nanoparticles, are extensively studied because of their wide applications in the biomedical, drug delivery, electronics fields and water treatment. The current study is centred on the synthesis of silver-doped Fe3O4/SiO2/TiO2 photocatalyst. Magnetically separable Fe3O4@SiO2@TiO2 composite with core–shell structure were synthesized by the deposition of uniform anatase TiO2 NPs on Fe3O4@SiO2 by using titanium butoxide (TBOT) as titanium source. Then, the silver is doped on SiO2 layer by hydrothermal method. Integration of magnetic nanoparticles was suggested to avoid the post separation difficulties associated with the powder form of the TiO2 catalyst, increase of the surface area and adsorption properties. The morphology, structure, composition, and magnetism of the resulting composites were characterized and their photocatalytic activities were also evaluated. The results demonstrate that TiO2 NPs were uniformly deposited on the Fe3O4@SiO2 surface. The silver nanoparticles were also uniformly distributed on the surface of TiO2 nanoparticles. The aim of this work is to study the suitability of photocatalysis for the treatment of aqueous streams containing organic pollutants such as methylene blue which is selected as a model compound to represent one of the pollutants existing in wastewaters. Various factors such as initial pollutant concentration, photocatalyst dose and wastewater matrix were studied for their effect on the photocatalytic degradation of the organic model pollutants using the as synthesized catalysts and compared with the commercial titanium dioxide (Aeroxide P25). Photocatalysis was found to be a potential purification method for the studied pollutant also in an industrial wastewater matrix with the removal percentages of over 81 % within 15 minutes. Methylene blue was removed most efficiently and its removal consumed the least of energy in terms of the specific applied energy. The magnetic Ag/SiO2/TiO2 composites show high photocatalytic performance and can be recycled three times by magnetic separation without major loss of activity, which meant that they can be used as efficient and conveniently renewable photocatalyst.

Keywords: Magnetite nanoparticles, Titanium, Photocatalyst, Organic pollutant, Water treatment

Procedia PDF Downloads 253