Search results for: testing techniques
7596 Innovative Housing Construction Technologies in Slum Upgrading
Authors: Edmund M. Muthigani
Abstract:
Innovation in the construction industry has been characterized by new products and processes especially in slum upgrading. The need for low cost housing has motivated stakeholders to think outside the box in coming up with solutions. This paper explored innovative construction technologies that have been used in slum upgrading. The main objectives of the paper was to examine innovations in the construction housing sector and to show how incremental derived demand for decent housing has led to adoption of innovative technologies and materials. Systematic literature review was used to review studies on innovative construction technologies in slum upgrading. The review revealed slow process of innovations in the construction industry due to risk aversion by firms and the hesitance to adopt by firms and individuals. Low profit margins in low cost housing and lack of sufficient political support remain the major hurdles to innovative techniques adoption that can actualize right to decent housing. Conventional construction materials have remained unaffordable to many people and this has negated them decent housing. This has necessitated exploration of innovative materials to realize low cost housing. Stabilized soil blocks and sisal-cement roofing blocks are some of the innovative construction materials that have been utilized in slum upgrading. These innovative materials have not only lowered the cost of production of building elements but also eased costs of transport as the raw materials to produce them are readily available in or within the slum sites. Despite their shortcomings in durability and compressive strength, they have proved worthwhile in slum upgrading. Production of innovative construction materials and use of innovative techniques in slum upgrading also provided employment to the locals.Keywords: construction, housing, innovation, slum, technology
Procedia PDF Downloads 2077595 Making of Alloy Steel by Direct Alloying with Mineral Oxides during Electro-Slag Remelting
Authors: Vishwas Goel, Kapil Surve, Somnath Basu
Abstract:
In-situ alloying in steel during the electro-slag remelting (ESR) process has already been achieved by the addition of necessary ferroalloys into the electro-slag remelting mold. However, the use of commercially available ferroalloys during ESR processing is often found to be financially less favorable, in comparison with the conventional alloying techniques. However, a process of alloying steel with elements like chromium and manganese using the electro-slag remelting route is under development without any ferrochrome addition. The process utilizes in-situ reduction of refined mineral chromite (Cr₂O₃) and resultant enrichment of chromium in the steel ingot produced. It was established in course of this work that this process can become more advantageous over conventional alloying techniques, both economically and environmentally, for applications which inherently demand the use of the electro-slag remelting process, such as manufacturing of superalloys. A key advantage is the lower overall CO₂ footprint of this process relative to the conventional route of production, storage, and the addition of ferrochrome. In addition to experimentally validating the feasibility of the envisaged reactions, a mathematical model to simulate the reduction of chromium (III) oxide and transfer to chromium to the molten steel droplets was also developed as part of the current work. The developed model helps to correlate the amount of chromite input and the magnitude of chromium alloying that can be achieved through this process. Experiments are in progress to validate the predictions made by this model and to fine-tune its parameters.Keywords: alloying element, chromite, electro-slag remelting, ferrochrome
Procedia PDF Downloads 2237594 Influence of Boron Doping and Thermal Treatment on Internal Friction of Monocrystalline Si1-xGex(x≤0,02) Alloys
Authors: I. Kurashvili, G. Darsavelidze, G. Bokuchava, A. Sichinava, I. Tabatadze
Abstract:
The impact of boron doping on the internal friction (IF) and shear modulus temperature spectra of Si1-xGex(x≤0,02) monocrsytals has been investigated by reverse torsional pendulum oscillations characteristics testing. At room temperatures, microhardness and indentation modulus of the same specimens have been measured by dynamic ultra microhardness tester. It is shown that boron doping causes two kinds effect: At low boron concentration (~1015 cm-3) significant strengthening is revealed, while at the high boron concentration (~1019 cm-3) strengthening effect and activation characteristics of relaxation origin IF processes are reduced.Keywords: boron, doping, internal friction, si-ge alloys, thermal treatment
Procedia PDF Downloads 4587593 Optimum Design of Steel Space Frames by Hybrid Teaching-Learning Based Optimization and Harmony Search Algorithms
Authors: Alper Akin, Ibrahim Aydogdu
Abstract:
This study presents a hybrid metaheuristic algorithm to obtain optimum designs for steel space buildings. The optimum design problem of three-dimensional steel frames is mathematically formulated according to provisions of LRFD-AISC (Load and Resistance factor design of American Institute of Steel Construction). Design constraints such as the strength requirements of structural members, the displacement limitations, the inter-story drift and the other structural constraints are derived from LRFD-AISC specification. In this study, a hybrid algorithm by using teaching-learning based optimization (TLBO) and harmony search (HS) algorithms is employed to solve the stated optimum design problem. These algorithms are two of the recent additions to metaheuristic techniques of numerical optimization and have been an efficient tool for solving discrete programming problems. Using these two algorithms in collaboration creates a more powerful tool and mitigates each other’s weaknesses. To demonstrate the powerful performance of presented hybrid algorithm, the optimum design of a large scale steel building is presented and the results are compared to the previously obtained results available in the literature.Keywords: optimum structural design, hybrid techniques, teaching-learning based optimization, harmony search algorithm, minimum weight, steel space frame
Procedia PDF Downloads 5457592 Power Iteration Clustering Based on Deflation Technique on Large Scale Graphs
Authors: Taysir Soliman
Abstract:
One of the current popular clustering techniques is Spectral Clustering (SC) because of its advantages over conventional approaches such as hierarchical clustering, k-means, etc. and other techniques as well. However, one of the disadvantages of SC is the time consuming process because it requires computing the eigenvectors. In the past to overcome this disadvantage, a number of attempts have been proposed such as the Power Iteration Clustering (PIC) technique, which is one of versions from SC; some of PIC advantages are: 1) its scalability and efficiency, 2) finding one pseudo-eigenvectors instead of computing eigenvectors, and 3) linear combination of the eigenvectors in linear time. However, its worst disadvantage is an inter-class collision problem because it used only one pseudo-eigenvectors which is not enough. Previous researchers developed Deflation-based Power Iteration Clustering (DPIC) to overcome problems of PIC technique on inter-class collision with the same efficiency of PIC. In this paper, we developed Parallel DPIC (PDPIC) to improve the time and memory complexity which is run on apache spark framework using sparse matrix. To test the performance of PDPIC, we compared it to SC, ESCG, ESCALG algorithms on four small graph benchmark datasets and nine large graph benchmark datasets, where PDPIC proved higher accuracy and better time consuming than other compared algorithms.Keywords: spectral clustering, power iteration clustering, deflation-based power iteration clustering, Apache spark, large graph
Procedia PDF Downloads 1897591 Rhythm-Reading Success Using Conversational Solfege
Authors: Kelly Jo Hollingsworth
Abstract:
Conversational Solfege, a research-based, 12-step music literacy instructional method using the sound-before-sight approach, was used to teach rhythm-reading to 128-second grade students at a public school in the southeastern United States. For each step, multiple scripted techniques are supplied to teach each skill. Unit one was the focus of this study, which is quarter note and barred eighth note rhythms. During regular weekly music instruction, students completed method steps one through five, which includes aural discrimination, decoding familiar and unfamiliar rhythm patterns, and improvising rhythmic phrases using quarter notes and barred eighth notes. Intact classes were randomly assigned to two treatment groups for teaching steps six through eight, which was the visual presentation and identification of quarter notes and barred eighth notes, visually presenting and decoding familiar patterns, and visually presenting and decoding unfamiliar patterns using said notation. For three weeks, students practiced steps six through eight during regular weekly music class. One group spent five-minutes of class time on steps six through eight technique work, while the other group spends ten-minutes of class time practicing the same techniques. A pretest and posttest were administered, and ANOVA results reveal both the five-minute (p < .001) and ten-minute group (p < .001) reached statistical significance suggesting Conversational Solfege is an efficient, effective approach to teach rhythm-reading to second grade students. After two weeks of no instruction, students were retested to measure retention. Using a repeated-measures ANOVA, both groups reached statistical significance (p < .001) on the second posttest, suggesting both the five-minute and ten-minute group retained rhythm-reading skill after two weeks of no instruction. Statistical significance was not reached between groups (p=.252), suggesting five-minutes is equally as effective as ten-minutes of rhythm-reading practice using Conversational Solfege techniques. Future research includes replicating the study with other grades and units in the text.Keywords: conversational solfege, length of instructional time, rhythm-reading, rhythm instruction
Procedia PDF Downloads 1577590 Constitutional Courts as Positive Legislators: The Role of Indonesian Constitutional Court in Interpreting and Applying the Constitution
Authors: Masnur Marzuki
Abstract:
As in other democratic countries, the constitutional court of Indonesia has the role of interpreting and applying the Constitution in order to preserve its supremacy testing the constitutionality of statutes. With its strong power to enforce and guard the Constitution, the court is now challenged to provide people an opportunity to understand their constitutional rights close up. At the same time, the court has built up an enviable reputation among constitutional courts in new democracies for the technical quality of its legitimacy in the legal sense. Since its establishment in 2003, the Constitutional Court of Indonesia has decided more than 190 statutes in judicial review case. It has been remarkably successful to make a credible start on its work of guarding the Constitution. Unsurprisingly, many argue that the Court has elevated Indonesia’s democracy to a whole new level. In accomplishing its roles judicial review, the basic principle that can be identified is that the Constitutional Court must always be subordinated to the Constitution. It is not being allowed to invade the field of the legislator. In doing so, the court does not have any discretionary political basis in order to create legal norms or provisions that could not be deducted from the Constitution itself. When interpreting a statute “in accordance with the constitution”, the court recognizes and reasserts that it is strictly forbidden to extend the scope of a legal provision in such a way that would create a general norm not established by the law-maker. This paper aims to identify and assess the latest role of Indonesian Constitutional Court in interpreting and applying the Constitution. In particular, it questions 1) the role of the Constitutional Court in judicial review; and 2) the role of the court to assist the legislators in the accomplishment of their functions in order to preserve its supremacy testing the constitutionality of statutes. Concerning positive legislator, jurisprudential and judicial review theories will be approached. The empirical part will include qualitative and comparative research. Main questions to be addressed: Can the Constitutional Court be functionalized as positive legislator? What are the criteria for conducting role of Constitutional Courts as Positive Legislators and how can it be accepted? Concerning the subordination of Constitutional Courts to the Constitution and judicial review, both qualitative and quantitative methods will be used, and differences between Indonesia and German Constitutional Court will be observed. Other questions to be addressed: Can Constitutional Courts have any discretionary political basis in order to create legal norms or provisions that could not be deducted from the Constitution itself. Should the Constitutional Court always act as a negative legislator? However, the Constitutional Court in Indonesia has played role as positive legislators which create dynamic of Indonesian legal development. In performing the task of reviewing the constitutionality of statutes, the Constitutional Court has created legal norms or provisions that could be deducted from the Constitution itself.Keywords: constitution, court, law, rights
Procedia PDF Downloads 4257589 Intelligent Prediction System for Diagnosis of Heart Attack
Authors: Oluwaponmile David Alao
Abstract:
Due to an increase in the death rate as a result of heart attack. There is need to develop a system that can be useful in the diagnosis of the disease at the medical centre. This system will help in preventing misdiagnosis that may occur from the medical practitioner or the physicians. In this research work, heart disease dataset obtained from UCI repository has been used to develop an intelligent prediction diagnosis system. The system is modeled on a feedforwad neural network and trained with back propagation neural network. A recognition rate of 86% is obtained from the testing of the network.Keywords: heart disease, artificial neural network, diagnosis, prediction system
Procedia PDF Downloads 4507588 Iterative Reconstruction Techniques as a Dose Reduction Tool in Pediatric Computed Tomography Imaging: A Phantom Study
Authors: Ajit Brindhaban
Abstract:
Background and Purpose: Computed Tomography (CT) scans have become the largest source of radiation in radiological imaging. The purpose of this study was to compare the quality of pediatric Computed Tomography (CT) images reconstructed using Filtered Back Projection (FBP) with images reconstructed using different strengths of Iterative Reconstruction (IR) technique, and to perform a feasibility study to assess the use of IR techniques as a dose reduction tool. Materials and Methods: An anthropomorphic phantom representing a 5-year old child was scanned, in two stages, using a Siemens Somatom CT unit. In stage one, scans of the head, chest and abdomen were performed using standard protocols recommended by the scanner manufacturer. Images were reconstructed using FBP and 5 different strengths of IR. Contrast-to-Noise Ratios (CNR) were calculated from average CT number and its standard deviation measured in regions of interest created in the lungs, bone, and soft tissues regions of the phantom. Paired t-test and the one-way ANOVA were used to compare the CNR from FBP images with IR images, at p = 0.05 level. The lowest strength value of IR that produced the highest CNR was identified. In the second stage, scans of the head was performed with decreased mA(s) values relative to the increase in CNR compared to the standard FBP protocol. CNR values were compared in this stage using Paired t-test at p = 0.05 level. Results: Images reconstructed using IR technique had higher CNR values (p < 0.01.) in all regions compared to the FBP images, at all strengths of IR. The CNR increased with increasing IR strength of up to 3, in the head and chest images. Increases beyond this strength were insignificant. In abdomen images, CNR continued to increase up to strength 5. The results also indicated that, IR techniques improve CNR by a up to factor of 1.5. Based on the CNR values at strength 3 of IR images and CNR values of FBP images, a reduction in mA(s) of about 20% was identified. The images of the head acquired at 20% reduced mA(s) and reconstructed using IR at strength 3, had similar CNR as FBP images at standard mA(s). In the head scans of the phantom used in this study, it was demonstrated that similar CNR can be achieved even when the mA(s) is reduced by about 20% if IR technique with strength of 3 is used for reconstruction. Conclusions: The IR technique produced better image quality at all strengths of IR in comparison to FBP. IR technique can provide approximately 20% dose reduction in pediatric head CT while maintaining the same image quality as FBP technique.Keywords: filtered back projection, image quality, iterative reconstruction, pediatric computed tomography imaging
Procedia PDF Downloads 1487587 Geomatic Techniques to Filter Vegetation from Point Clouds
Authors: M. Amparo Núñez-Andrés, Felipe Buill, Albert Prades
Abstract:
More and more frequently, geomatics techniques such as terrestrial laser scanning or digital photogrammetry, either terrestrial or from drones, are being used to obtain digital terrain models (DTM) used for the monitoring of geological phenomena that cause natural disasters, such as landslides, rockfalls, debris-flow. One of the main multitemporal analyses developed from these models is the quantification of volume changes in the slopes and hillsides, either caused by erosion, fall, or land movement in the source area or sedimentation in the deposition zone. To carry out this task, it is necessary to filter the point clouds of all those elements that do not belong to the slopes. Among these elements, vegetation stands out as it is the one we find with the greatest presence and its constant change, both seasonal and daily, as it is affected by factors such as wind. One of the best-known indexes to detect vegetation on the image is the NVDI (Normalized Difference Vegetation Index), which is obtained from the combination of the infrared and red channels. Therefore it is necessary to have a multispectral camera. These cameras are generally of lower resolution than conventional RGB cameras, while their cost is much higher. Therefore we have to look for alternative indices based on RGB. In this communication, we present the results obtained in Georisk project (PID2019‐103974RB‐I00/MCIN/AEI/10.13039/501100011033) by using the GLI (Green Leaf Index) and ExG (Excessive Greenness), as well as the change to the Hue-Saturation-Value (HSV) color space being the H coordinate the one that gives us the most information for vegetation filtering. These filters are applied both to the images, creating binary masks to be used when applying the SfM algorithms, and to the point cloud obtained directly by the photogrammetric process without any previous filter or the one obtained by TLS (Terrestrial Laser Scanning). In this last case, we have also tried to work with a Riegl VZ400i sensor that allows the reception, as in the aerial LiDAR, of several returns of the signal. Information to be used for the classification on the point cloud. After applying all the techniques in different locations, the results show that the color-based filters allow correct filtering in those areas where the presence of shadows is not excessive and there is a contrast between the color of the slope lithology and the vegetation. As we have advanced in the case of using the HSV color space, it is the H coordinate that responds best for this filtering. Finally, the use of the various returns of the TLS signal allows filtering with some limitations.Keywords: RGB index, TLS, photogrammetry, multispectral camera, point cloud
Procedia PDF Downloads 1547586 Analytical Study and Conservation Processes of Scribe Box from Old Kingdom
Authors: Mohamed Moustafa, Medhat Abdallah, Ramy Magdy, Ahmed Abdrabou, Mohamed Badr
Abstract:
The scribe box under study dates back to the old kingdom. It was excavated by the Italian expedition in Qena (1935-1937). The box consists of 2pieces, the lid and the body. The inner side of the lid is decorated with ancient Egyptian inscriptions written with a black pigment. The box was made using several panels assembled together by wooden dowels and secured with plant ropes. The entire box is covered with a red pigment. This study aims to use analytical techniques in order to identify and have deep understanding for the box components. Moreover, the authors were significantly interested in using infrared reflectance transmission imaging (RTI-IR) to improve the hidden inscriptions on the lid. The identification of wood species included in this study. The visual observation and assessment were done to understand the condition of this box. 3Ddimensions and 2D programs were used to illustrate wood joints techniques. Optical microscopy (OM), X-ray diffraction (XRD), X-ray fluorescence portable (XRF) and Fourier Transform Infrared spectroscopy (FTIR) were used in this study in order to identify wood species, remains of insects bodies, red pigment, fibers plant and previous conservation adhesives, also RTI-IR technique was very effective to improve hidden inscriptions. The analysis results proved that wooden panels and dowels were identified as Acacia nilotica, wooden rail was Salix sp. the insects were identified as Lasioderma serricorne and Gibbium psylloids, the red pigment was Hematite, while the fiber plants were linen, previous adhesive was identified as cellulose nitrates. The historical study for the inscriptions proved that it’s a Hieratic writings of a funerary Text. After its transportation from the Egyptian museum storage to the wood conservation laboratory of the Grand Egyptian museum –conservation center (GEM-CC), conservation techniques were applied with high accuracy in order to restore the object including cleaning , consolidating of friable pigments and writings, removal of previous adhesive and reassembly, finally the conservation process that were applied were extremely effective for this box which became ready for display or storage in the grand Egyptian museum.Keywords: scribe box, hieratic, 3D program, Acacia nilotica, XRD, cellulose nitrate, conservation
Procedia PDF Downloads 2717585 Usability Guidelines for Arab E-Government Websites
Authors: Omyma Alosaimi, Asma Alsumait
Abstract:
The website developer and designer should follow usability guidelines to provide a user-friendly interface. Many guidelines and heuristics have been developed by previous studies to help both the developer and designer in this task, but E-government websites are special cases that require specialized guidelines. This paper introduces a set of eighteen guidelines for evaluating the usability of e-government websites in general and Arabic e-government websites specifically, along with a check list of how to apply them. The validity and effectiveness of these guidelines were evaluated against a variety of user characteristics. The results indicated that the proposed set of guidelines can be used to identify qualitative similarities and differences with user testing and that the new set is best suited for evaluating general and e-governmental usability.Keywords: e-government, human computer interaction, usability evaluation, usability guidelines
Procedia PDF Downloads 3957584 A Dynamic Solution Approach for Heart Disease Prediction
Authors: Walid Moudani
Abstract:
The healthcare environment is generally perceived as being information rich yet knowledge poor. However, there is a lack of effective analysis tools to discover hidden relationships and trends in data. In fact, valuable knowledge can be discovered from application of data mining techniques in healthcare system. In this study, a proficient methodology for the extraction of significant patterns from the coronary heart disease warehouses for heart attack prediction, which unfortunately continues to be a leading cause of mortality in the whole world, has been presented. For this purpose, we propose to enumerate dynamically the optimal subsets of the reduced features of high interest by using rough sets technique associated to dynamic programming. Therefore, we propose to validate the classification using Random Forest (RF) decision tree to identify the risky heart disease cases. This work is based on a large amount of data collected from several clinical institutions based on the medical profile of patient. Moreover, the experts’ knowledge in this field has been taken into consideration in order to define the disease, its risk factors, and to establish significant knowledge relationships among the medical factors. A computer-aided system is developed for this purpose based on a population of 525 adults. The performance of the proposed model is analyzed and evaluated based on set of benchmark techniques applied in this classification problem.Keywords: multi-classifier decisions tree, features reduction, dynamic programming, rough sets
Procedia PDF Downloads 4107583 Understanding Evidence Dispersal Caused by the Effects of Using Unmanned Aerial Vehicles in Active Indoor Crime Scenes
Authors: Elizabeth Parrott, Harry Pointon, Frederic Bezombes, Heather Panter
Abstract:
Unmanned aerial vehicles (UAV’s) are making a profound effect within policing, forensic and fire service procedures worldwide. These intelligent devices have already proven useful in photographing and recording large-scale outdoor and indoor sites using orthomosaic and three-dimensional (3D) modelling techniques, for the purpose of capturing and recording sites during and post-incident. UAV’s are becoming an established tool as they are extending the reach of the photographer and offering new perspectives without the expense and restrictions of deploying full-scale aircraft. 3D reconstruction quality is directly linked to the resolution of captured images; therefore, close proximity flights are required for more detailed models. As technology advances deployment of UAVs in confined spaces is becoming more common. With this in mind, this study investigates the effects of UAV operation within active crimes scenes with regard to the dispersal of particulate evidence. To date, there has been little consideration given to the potential effects of using UAV’s within active crime scenes aside from a legislation point of view. Although potentially the technology can reduce the likelihood of contamination by replacing some of the roles of investigating practitioners. There is the risk of evidence dispersal caused by the effect of the strong airflow beneath the UAV, from the downwash of the propellers. The initial results of this study are therefore presented to determine the height of least effect at which to fly, and the commercial propeller type to choose to generate the smallest amount of disturbance from the dataset tested. In this study, a range of commercially available 4-inch propellers were chosen as a starting point due to the common availability and their small size makes them well suited for operation within confined spaces. To perform the testing, a rig was configured to support a single motor and propeller powered with a standalone mains power supply and controlled via a microcontroller. This was to mimic a complete throttle cycle and control the device to ensure repeatability. By removing the variances of battery packs and complex UAV structures to allow for a more robust setup. Therefore, the only changing factors were the propeller and operating height. The results were calculated via computer vision analysis of the recorded dispersal of the sample particles placed below the arm-mounted propeller. The aim of this initial study is to give practitioners an insight into the technology to use when operating within confined spaces as well as recognizing some of the issues caused by UAV’s within active crime scenes.Keywords: dispersal, evidence, propeller, UAV
Procedia PDF Downloads 1637582 Development and Range Testing of a LoRaWAN System in an Urban Environment
Authors: N. R. Harris, J. Curry
Abstract:
This paper describes the construction and operation of an experimental LoRaWAN network surrounding the University of Southampton in the United Kingdom. Following successful installation, an experimental node design is built and characterised, with particular emphasis on radio range. Several configurations are investigated, including different data rates, and varying heights of node. It is concluded that although range can be great (over 8 km in this case), environmental topology is critical. However, shorter range implementations, up to about 2 km in an urban environment, are relatively insensitive although care is still needed. The example node and the relatively simple base station reported demonstrate that LoraWan can be a very low cost and practical solution to Internet of Things type applications for distributed monitoring systems with sensors spread over distances of several km.Keywords: long-range, wireless, sensor, network
Procedia PDF Downloads 1377581 Effect of Plasma Treatment on UV Protection Properties of Fabrics
Authors: Sheila Shahidi
Abstract:
UV protection by fabrics has recently become a focus of great interest, particularly in connection with environmental degradation or ozone layer depletion. Fabrics provide simple and convenient protection against UV radiation (UVR), but not all fabrics offer sufficient UV protection. To describe the degree of UVR protection offered by clothing materials, the ultraviolet protection factor (UPF) is commonly used. UV-protective fabric can be generated by application of a chemical finish using normal wet-processing methodologies. However, traditional wet-processing techniques are known to consume large quantities of water and energy and may lead to adverse alterations of the bulk properties of the substrate. Recently, usage of plasmas to generate physicochemical surface modifications of textile substrates has become an intriguing approach to replace or enhance conventional wet-processing techniques. In this research work the effect of plasma treatment on UV protection properties of fabrics was investigated. DC magnetron sputtering was used and the parameters of plasma such as gas type, electrodes, time of exposure, power and, etc. were studied. The morphological and chemical properties of samples were analyzed using Scanning Electron Microscope (SEM) and Furrier Transform Infrared Spectroscopy (FTIR), respectively. The transmittance and UPF values of the original and plasma-treated samples were measured using a Shimadzu UV3101 PC (UV–Vis–NIR scanning spectrophotometer, 190–2, 100 nm range). It was concluded that, plasma which is an echo-friendly, cost effective and dry technique is being used in different branches of the industries, and will conquer textile industry in the near future. Also it is promising method for preparation of UV protection textile.Keywords: fabric, plasma, textile, UV protection
Procedia PDF Downloads 5197580 Elimination of Mother to Child Transmission of HIV/AIDS: A Study of the Knowledge, Attitudes and Perceptions of Healthcare Workers in Abuja Nigeria
Authors: Ezinne K. Okoro, Takahiko Katoh, Yoko Kawamura, Stanley C. Meribe
Abstract:
HIV infection in children is largely as a result of vertical transmission (mother to child transmission [MTCT]). Thus, elimination of mother to child transmission of HIV/AIDS is critical in eliminating HIV infection in children. In Nigeria, drawbacks such as; limited pediatric screening, limited human capital, insufficient advocacy and poor understanding of ART guidelines, have impacted efforts at combating the disease, even as treatment services are free. Prevention of Mother to Child Transmission (PMTCT) program relies on health workers who not only counsel pregnant women on first contact but can competently provide HIV-positive pregnant women with accurate information about the PMTCT program such as feeding techniques and drug adherence. In developing regions like Nigeria where health care delivery faces a lot of drawbacks, it becomes paramount to address these issues of poor PMTCT coverage by conducting a baseline assessment of the knowledge, practices and perceptions related to HIV prevention amongst healthcare workers in Nigeria. A descriptive cross-sectional study was conducted amongst 250 health workers currently employed in health facilities in Abuja, Nigeria where PMTCT services were offered with the capacity to carry out early infant diagnosis testing (EID). Data was collected using a self-administered, pretested, structured questionnaire. This study showed that the knowledge of PMTCT of HIV was poor (30%) among healthcare workers who offer this service day-to-day to pregnant women. When PMTCT practices were analyzed in keeping with National PMTCT guidelines, over 61% of the respondents reported observing standard practices and the majority (58%) had good attitudes towards caring for patients with HIV/AIDS. Although 61% of the respondents reported being satisfied with the quality of service being rendered, 63% reported not being satisfied with their level of knowledge. Predictors of good knowledge were job designation and level of educational attainment. Health workers who were more satisfied with their working conditions and those who had worked for a longer time in the PMTCT service were more likely to observe standard PMTCT practices. With over 62% of the healthcare workers suggesting that more training would improve the quality of service being rendered, this is a strong pointer to stakeholders to consider a ‘healthcare worker-oriented approach’ when planning and conducting PMTCT training for healthcare workers. This in turn will increase pediatric ARV coverage, the knowledge and effectiveness of the healthcare workers in carrying out appropriate PMTCT interventions and culminating in the reduction/elimination of HIV transmission to newborns.Keywords: attitudes, HIV/AIDS, healthcare workers, knowledge, mother to child transmission, Nigeria, perceptions
Procedia PDF Downloads 2057579 Dynamic Modeling of the Exchange Rate in Tunisia: Theoretical and Empirical Study
Authors: Chokri Slim
Abstract:
The relative failure of simultaneous equation models in the seventies has led researchers to turn to other approaches that take into account the dynamics of economic and financial systems. In this paper, we use an approach based on vector autoregressive model that is widely used in recent years. Their popularity is due to their flexible nature and ease of use to produce models with useful descriptive characteristics. It is also easy to use them to test economic hypotheses. The standard econometric techniques assume that the series studied are stable over time (stationary hypothesis). Most economic series do not verify this hypothesis, which assumes, when one wishes to study the relationships that bind them to implement specific techniques. This is cointegration which characterizes non-stationary series (integrated) with a linear combination is stationary, will also be presented in this paper. Since the work of Johansen, this approach is generally presented as part of a multivariate analysis and to specify long-term stable relationships while at the same time analyzing the short-term dynamics of the variables considered. In the empirical part, we have applied these concepts to study the dynamics of of the exchange rate in Tunisia, which is one of the most important economic policy of a country open to the outside. According to the results of the empirical study by the cointegration method, there is a cointegration relationship between the exchange rate and its determinants. This relationship shows that the variables have a significant influence in determining the exchange rate in Tunisia.Keywords: stationarity, cointegration, dynamic models, causality, VECM models
Procedia PDF Downloads 3647578 Generation of Numerical Data for the Facilitation of the Personalized Hyperthermic Treatment of Cancer with An Interstital Antenna Array Using the Method of Symmetrical Components
Authors: Prodromos E. Atlamazoglou
Abstract:
The method of moments combined with the method of symmetrical components is used for the analysis of interstitial hyperthermia applicators. The basis and testing functions are both piecewise sinusoids, qualifying our technique as a Galerkin one. The dielectric coatings are modeled by equivalent volume polarization currents, which are simply related to the conduction current distribution, avoiding in that way the introduction of additional unknowns or numerical integrations. The results of our method for a four dipole circular array, are in agreement with those already published in literature for a same hyperthermia configuration. Apart from being accurate, our approach is more general, more computationally efficient and takes into account the coupling between the antennas.Keywords: hyperthermia, integral equations, insulated antennas, method of symmetrical components
Procedia PDF Downloads 2577577 Direct Phoenix Identification and Antimicrobial Susceptibility Testing from Positive Blood Culture Broths
Authors: Waad Al Saleemi, Badriya Al Adawi, Zaaima Al Jabri, Sahim Al Ghafri, Jalila Al Hadhramia
Abstract:
Objectives: Using standard lab methods, a positive blood culture requires a minimum of two days (two occasions of overnight incubation) to obtain a final identification (ID) and antimicrobial susceptibility results (AST) report. In this study, we aimed to evaluate the accuracy and precision of identification and antimicrobial susceptibility testing of an alternative method (direct method) that will reduce the turnaround time by 24 hours. This method involves the direct inoculation of positive blood culture broths into the Phoenix system using serum separation tubes (SST). Method: This prospective study included monomicrobial-positive blood cultures obtained from January 2022 to May 2023 in SQUH. Blood cultures containing a mixture of organisms, fungi, or anaerobic organisms were excluded from this study. The result of the new “direct method” under study was compared with the current “standard method” used in the lab. The accuracy and precision were evaluated for the ID and AST using Clinical and Laboratory Standards Institute (CLSI) recommendations. The categorical agreement, essential agreement, and the rates of very major errors (VME), major errors (ME), and minor errors (MIE) for both gram-negative and gram-positive bacteria were calculated. Passing criteria were set according to CLSI. Result: The results of ID and AST were available for a total of 158 isolates. Of 77 isolates of gram-negative bacteria, 71 (92%) were correctly identified at the species level. Of 70 isolates of gram-positive bacteria, 47(67%) isolates were correctly identified. For gram-negative bacteria, the essential agreement of the direct method was ≥92% when compared to the standard method, while the categorical agreement was ≥91% for all tested antibiotics. The precision of ID and AST were noted to be 100% for all tested isolates. For gram-positive bacteria, the essential agreement was >93%, while the categorical agreement was >92% for all tested antibiotics except moxifloxacin. Many antibiotics were noted to have an unacceptable higher rate of very major errors including penicillin, cotrimoxazole, clindamycin, ciprofloxacin, and moxifloxacin. However, no error was observed in the results of vancomycin, linezolid, and daptomycin. Conclusion: The direct method of ID and AST for positive blood cultures using SST is reliable for gram negative bacteria. It will significantly decrease the turnaround time and will facilitate antimicrobial stewardship.Keywords: bloodstream infection, oman, direct ast, blood culture, rapid identification, antimicrobial susceptibility, phoenix, direct inoculation
Procedia PDF Downloads 637576 Modern Scotland Yard: Improving Surveillance Policies Using Adversarial Agent-Based Modelling and Reinforcement Learning
Authors: Olaf Visker, Arnout De Vries, Lambert Schomaker
Abstract:
Predictive policing refers to the usage of analytical techniques to identify potential criminal activity. It has been widely implemented by various police departments. Being a relatively new area of research, there are, to the author’s knowledge, no absolute tried, and true methods and they still exhibit a variety of potential problems. One of those problems is closely related to the lack of understanding of how acting on these prediction influence crime itself. The goal of law enforcement is ultimately crime reduction. As such, a policy needs to be established that best facilitates this goal. This research aims to find such a policy by using adversarial agent-based modeling in combination with modern reinforcement learning techniques. It is presented here that a baseline model for both law enforcement and criminal agents and compare their performance to their respective reinforcement models. The experiments show that our smart law enforcement model is capable of reducing crime by making more deliberate choices regarding the locations of potential criminal activity. Furthermore, it is shown that the smart criminal model presents behavior consistent with popular crime theories and outperforms the baseline model in terms of crimes committed and time to capture. It does, however, still suffer from the difficulties of capturing long term rewards and learning how to handle multiple opposing goals.Keywords: adversarial, agent based modelling, predictive policing, reinforcement learning
Procedia PDF Downloads 1487575 Component Test of Martensitic/Ferritic Steels and Nickel-Based Alloys and Their Welded Joints under Creep and Thermo-Mechanical Fatigue Loading
Authors: Daniel Osorio, Andreas Klenk, Stefan Weihe, Andreas Kopp, Frank Rödiger
Abstract:
Future power plants currently face high design requirements due to worsening climate change and environmental restrictions, which demand high operational flexibility, superior thermal performance, minimal emissions, and higher cyclic capability. The aim of the paper is, therefore, to investigate the creep and thermo-mechanical material behavior of improved materials experimentally and welded joints at component scale under near-to-service operating conditions, which are promising for application in highly efficient and flexible future power plants. These materials promise an increase in flexibility and a reduction in manufacturing costs by providing enhanced creep strength and, therefore, the possibility for wall thickness reduction. At the temperature range between 550°C and 625°C, the investigation focuses on the in-phase thermo-mechanical fatigue behavior of dissimilar welded joints of conventional materials (ferritic and martensitic material T24 and T92) to nickel-based alloys (A617B and HR6W) by means of membrane test panels. The temperature and external load are varied in phase during the test, while the internal pressure remains constant. At the temperature range between 650°C and 750°C, it focuses on the creep behavior under multiaxial stress loading of similar and dissimilar welded joints of high temperature resistant nickel-based alloys (A740H, A617B, and HR6W) by means of a thick-walled-component test. In this case, the temperature, the external axial load, and the internal pressure remain constant during testing. Numerical simulations are used for the estimation of the axial component load in order to induce a meaningful damage evolution without causing a total component failure. Metallographic investigations after testing will provide support for understanding the damage mechanism and the influence of the thermo-mechanical load and multiaxiality on the microstructure change and on the creep and TMF- strength.Keywords: creep, creep-fatigue, component behaviour, weld joints, high temperature material behaviour, nickel-alloys, high temperature resistant steels
Procedia PDF Downloads 1197574 Life Time Improvement of Clamp Structural by Using Fatigue Analysis
Authors: Pisut Boonkaew, Jatuporn Thongsri
Abstract:
In hard disk drive manufacturing industry, the process of reducing an unnecessary part and qualifying the quality of part before assembling is important. Thus, clamp was designed and fabricated as a fixture for holding in testing process. Basically, testing by trial and error consumes a long time to improve. Consequently, the simulation was brought to improve the part and reduce the time taken. The problem is the present clamp has a low life expectancy because of the critical stress that occurred. Hence, the simulation was brought to study the behavior of stress and compressive force to improve the clamp expectancy with all probability of designs which are present up to 27 designs, which excluding the repeated designs. The probability was calculated followed by the full fractional rules of six sigma methodology which was provided correctly. The six sigma methodology is a well-structured method for improving quality level by detecting and reducing the variability of the process. Therefore, the defective will be decreased while the process capability increasing. This research focuses on the methodology of stress and fatigue reduction while compressive force still remains in the acceptable range that has been set by the company. In the simulation, ANSYS simulates the 3D CAD with the same condition during the experiment. Then the force at each distance started from 0.01 to 0.1 mm will be recorded. The setting in ANSYS was verified by mesh convergence methodology and compared the percentage error with the experimental result; the error must not exceed the acceptable range. Therefore, the improved process focuses on degree, radius, and length that will reduce stress and still remain in the acceptable force number. Therefore, the fatigue analysis will be brought as the next process in order to guarantee that the lifetime will be extended by simulating through ANSYS simulation program. Not only to simulate it, but also to confirm the setting by comparing with the actual clamp in order to observe the different of fatigue between both designs. This brings the life time improvement up to 57% compared with the actual clamp in the manufacturing. This study provides a precise and trustable setting enough to be set as a reference methodology for the future design. Because of the combination and adaptation from the six sigma method, finite element, fatigue and linear regressive analysis that lead to accurate calculation, this project will able to save up to 60 million dollars annually.Keywords: clamp, finite element analysis, structural, six sigma, linear regressive analysis, fatigue analysis, probability
Procedia PDF Downloads 2357573 Quality of Age Reporting from Tanzania 2012 Census Results: An Assessment Using Whipple’s Index, Myer’s Blended Index, and Age-Sex Accuracy Index
Authors: A. Sathiya Susuman, Hamisi F. Hamisi
Abstract:
Background: Many socio-economic and demographic data are age-sex attributed. However, a variety of irregularities and misstatement are noted with respect to age-related data and less to sex data because of its biological differences between the genders. Noting the misstatement/misreporting of age data regardless of its significance importance in demographics and epidemiological studies, this study aims at assessing the quality of 2012 Tanzania Population and Housing Census Results. Methods: Data for the analysis are downloaded from Tanzania National Bureau of Statistics. Age heaping and digit preference were measured using summary indices viz., Whipple’s index, Myers’ blended index, and Age-Sex Accuracy index. Results: The recorded Whipple’s index for both sexes was 154.43; male has the lowest index of about 152.65 while female has the highest index of about 156.07. For Myers’ blended index, the preferences were at digits ‘0’ and ‘5’ while avoidance were at digits ‘1’ and ‘3’ for both sexes. Finally, Age-sex index stood at 59.8 where sex ratio score was 5.82 and age ratio scores were 20.89 and 21.4 for males and female respectively. Conclusion: The evaluation of the 2012 PHC data using the demographic techniques has qualified the data inaccurate as the results of systematic heaping and digit preferences/avoidances. Thus, innovative methods in data collection along with measuring and minimizing errors using statistical techniques should be used to ensure accuracy of age data.Keywords: age heaping, digit preference/avoidance, summary indices, Whipple’s index, Myer’s index, age-sex accuracy index
Procedia PDF Downloads 4767572 Structural Equation Modeling Exploration for the Multiple College Admission Criteria in Taiwan
Authors: Tzu-Ling Hsieh
Abstract:
When the Taiwan Ministry of Education implemented a new university multiple entrance policy in 2002, most colleges and universities still use testing scores as mainly admission criteria. With forthcoming 12 basic-year education curriculum, the Ministry of Education provides a new college admission policy, which will be implemented in 2021. The new college admission policy will highlight the importance of holistic education by more emphases on the learning process of senior high school, except only on the outcome of academic testing. However, the development of college admission criteria doesn’t have a thoughtful process. Universities and colleges don’t have an idea about how to make suitable multi-admission criteria. Although there are lots of studies in other countries which have implemented multi-college admission criteria for years, these studies still cannot represent Taiwanese students. Also, these studies are limited without the comparison of two different academic fields. Therefore, this study investigated multiple admission criteria and its relationship with college success. This study analyzed the Taiwan Higher Education Database with 12,747 samples from 156 universities and tested a conceptual framework that examines factors by structural equation model (SEM). The conceptual framework of this study was adapted from Pascarella's general causal model and focused on how different admission criteria predict students’ college success. It discussed the relationship between admission criteria and college success, also the relationship how motivation (one of admission standard) influence college success through engagement behaviors of student effort and interactions with agents of socialization. After processing missing value, reliability and validity analysis, the study found three indicators can significantly predict students’ college success which was defined as average grade of last semester. These three indicators are the Chinese language scores at college entrance exam, high school class rank, and quality of student academic engagement. In addition, motivation can significantly predict quality of student academic engagement and interactions with agents of socialization. However, the multi-group SEM analysis showed that there is no difference to predict college success between the students from liberal arts and science. Finally, this study provided some suggestions for universities and colleges to develop multi-admission criteria through the empirical research of Taiwanese higher education students.Keywords: college admission, admission criteria, structural equation modeling, higher education, education policy
Procedia PDF Downloads 1787571 Preliminary Results on a Study of Antimicrobial Susceptibility Testing of Bacillus anthracis Strains Isolated during Anthrax Outbreaks in Italy from 2001 to 2017
Authors: Viviana Manzulli, Luigina Serrecchia, Adelia Donatiello, Valeria Rondinone, Sabine Zange, Alina Tscherne, Antonio Parisi, Antonio Fasanella
Abstract:
Anthrax is a zoonotic disease that affects a wide range of animal species (primarily ruminant herbivores), and can be transmitted to humans through consumption or handling of contaminated animal products. The etiological agent B.anthracis is able to survive in unfavorable environmental conditions by forming endospore which remain viable in the soil for many decades. Furthermore, B.anthracis is considered as one of the most feared agents to be potentially misused as a biological weapon and the importance of the disease and its treatment in humans has been underscored before the bioterrorism events in the United States in 2001. Due to the often fatal outcome of human cases, antimicrobial susceptibility testing plays especially in the management of anthrax infections an important role. In Italy, animal anthrax is endemic (predominantly found in the southern regions and on islands) and is characterized by sporadic outbreaks occurring mainly during summer. Between 2012 and 2017 single human cases of cutaneous anthrax occurred. In this study, 90 diverse strains of B.anthracis, isolated in Italy from 2001 to 2017, were screened to their susceptibility to sixteen clinically relevant antimicrobial agents by using the broth microdilution method. B.anthracis strains selected for this study belong to the strain collection stored at the Anthrax Reference Institute of Italy located inside the Istituto Zooprofilattico Sperimentale of Puglia and Basilicata. The strains were isolated at different time points and places from various matrices (human, animal and environmental). All strains are a representative of over fifty distinct MLVA 31 genotypes. The following antibiotics were used for testing: gentamicin, ceftriaxone, streptomycin, penicillin G, clindamycin, chloramphenicol, vancomycin, linezolid, cefotaxime, tetracycline, erythromycin, rifampin, amoxicillin, ciprofloxacin, doxycycline and trimethoprim. A standard concentration of each antibiotic was prepared in a specific diluent, which were then twofold serial diluted. Therefore, each wells contained: bacterial suspension of 1–5x104 CFU/mL in Mueller-Hinton Broth (MHB), the antibiotic to be tested at known concentration and resazurin, an indicator of cell growth. After incubation overnight at 37°C, the wells were screened for color changes caused by the resazurin: a change from purple to pink/colorless indicated cell growth. The lowest concentration of antibiotic that prevented growth represented the minimal inhibitory concentration (MIC). This study suggests that B.anthracis remains susceptible in vitro to many antibiotics, in addition to doxycycline (MICs ≤ 0,03 µg/ml), ciprofloxacin (MICs ≤ 0,03 µg/ml) and penicillin G (MICs ≤ 0,06 µg/ml), recommend by CDC for the treatment of human cases and for prophylactic use after exposure to the spores. In fact, the good activity of gentamicin (MICs ≤ 0,25 µg/ml), streptomycin (MICs ≤ 1 µg/ml), clindamycin (MICs ≤ 0,125 µg/ml), chloramphenicol(MICs ≤ 4 µg/ml), vancomycin (MICs ≤ 2 µg/ml), linezolid (MICs ≤ 2 µg/ml), tetracycline (MICs ≤ 0,125 µg/ml), erythromycin (MICs ≤ 0,25 µg/ml), rifampin (MICs ≤ 0,25 µg/ml), amoxicillin (MICs ≤ 0,06 µg/ml), towards all tested B.anthracis strains demonstrates an appropriate alternative choice for prophylaxis and/or treatment. All tested B.anthracis strains showed intermediate susceptibility to the cephalosporins (MICs ≥ 16 µg/ml) and resistance to trimethoprim (MICs ≥ 128 µg/ml).Keywords: Bacillus anthracis, antibiotic susceptibility, treatment, minimum inhibitory concentration
Procedia PDF Downloads 2127570 Efficient Control of Some Dynamic States of Wheeled Robots
Authors: Boguslaw Schreyer
Abstract:
In some types of wheeled robots it is important to secure starting acceleration and deceleration maxima while at the same time maintaining transversal stability. In this paper torque distribution between the front and rear wheels as well as the timing of torque application have been calculated. Both secure an optimum traction coefficient. This paper also identifies required input signals to a control unit, which controls the torque values and timing. Using a three dimensional, two mass model of a robot developed by the author a computer simulation was performed confirming the calculations presented in this paper. These calculations were also implemented and confirmed during military robot testing.Keywords: robot dynamics, torque distribution, traction coefficient, wheeled robots
Procedia PDF Downloads 3127569 Advances in Machine Learning and Deep Learning Techniques for Image Classification and Clustering
Authors: R. Nandhini, Gaurab Mudbhari
Abstract:
Ranging from the field of health care to self-driving cars, machine learning and deep learning algorithms have revolutionized the field with the proper utilization of images and visual-oriented data. Segmentation, regression, classification, clustering, dimensionality reduction, etc., are some of the Machine Learning tasks that helped Machine Learning and Deep Learning models to become state-of-the-art models for the field where images are key datasets. Among these tasks, classification and clustering are essential but difficult because of the intricate and high-dimensional characteristics of image data. This finding examines and assesses advanced techniques in supervised classification and unsupervised clustering for image datasets, emphasizing the relative efficiency of Convolutional Neural Networks (CNNs), Vision Transformers (ViTs), Deep Embedded Clustering (DEC), and self-supervised learning approaches. Due to the distinctive structural attributes present in images, conventional methods often fail to effectively capture spatial patterns, resulting in the development of models that utilize more advanced architectures and attention mechanisms. In image classification, we investigated both CNNs and ViTs. One of the most promising models, which is very much known for its ability to detect spatial hierarchies, is CNN, and it serves as a core model in our study. On the other hand, ViT is another model that also serves as a core model, reflecting a modern classification method that uses a self-attention mechanism which makes them more robust as this self-attention mechanism allows them to lean global dependencies in images without relying on convolutional layers. This paper evaluates the performance of these two architectures based on accuracy, precision, recall, and F1-score across different image datasets, analyzing their appropriateness for various categories of images. In the domain of clustering, we assess DEC, Variational Autoencoders (VAEs), and conventional clustering techniques like k-means, which are used on embeddings derived from CNN models. DEC, a prominent model in the field of clustering, has gained the attention of many ML engineers because of its ability to combine feature learning and clustering into a single framework and its main goal is to improve clustering quality through better feature representation. VAEs, on the other hand, are pretty well known for using latent embeddings for grouping similar images without requiring for prior label by utilizing the probabilistic clustering method.Keywords: machine learning, deep learning, image classification, image clustering
Procedia PDF Downloads 107568 Energy Conservation in Heat Exchangers
Authors: Nadia Allouache
Abstract:
Energy conservation is one of the major concerns in the modern high tech era due to the limited amount of energy resources and the increasing cost of energy. Predicting an efficient use of energy in thermal systems like heat exchangers can only be achieved if the second law of thermodynamics is accounted for. The performance of heat exchangers can be substantially improved by many passive heat transfer augmentation techniques. These letters permit to improve heat transfer rate and to increase exchange surface, but on the other side, they also increase the friction factor associated with the flow. This raises the question of how to employ these passive techniques in order to minimize the useful energy. The objective of this present study is to use a porous substrate attached to the walls as a passive enhancement technique in heat exchangers and to find the compromise between the hydrodynamic and thermal performances under turbulent flow conditions, by using a second law approach. A modified k- ε model is used to simulating the turbulent flow in the porous medium and the turbulent shear flow is accounted for in the entropy generation equation. A numerical modeling, based on the finite volume method is employed for discretizing the governing equations. Effects of several parameters are investigated such as the porous substrate properties and the flow conditions. Results show that under certain conditions of the porous layer thickness, its permeability, and its effective thermal conductivity the minimum rate of entropy production is obtained.Keywords: second law approach, annular heat exchanger, turbulent flow, porous medium, modified model, numerical analysis
Procedia PDF Downloads 2887567 Design and Implementation of Wave-Pipelined Circuit Using Reconfigurable Technique
Authors: Adhinarayanan Venkatasubramanian
Abstract:
For design of high speed digital circuit wave pipeline is the best approach this can be operated at higher operating frequencies by adjusting clock periods and skews so as latch the o/p of combinational logic circuit at the stable period. In this paper, there are two methods are proposed in automation task one is BIST (Built in self test) and second method is Reconfigurable technique. For the above two approaches dedicated AND gate (multiplier) by applying wave pipeline technique. BIST approach is implemented by Xilinx Spartan-II device. In reconfigurable technique done by ASIC. From the results, wave pipeline circuits are faster than nonpipeline circuit and area, power dissipation are reduced by reconfigurable technique.Keywords: SOC, wave-pipelining, FPGA, self-testing, reconfigurable, ASIC
Procedia PDF Downloads 427