Search results for: application lifecycle management
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 16204

Search results for: application lifecycle management

3184 Methods of Livable Goal-Oriented Master Urban Design: A Case Study on Zibo City

Authors: Xiaoping Zhang, Fengying Yan

Abstract:

The implementation of the 'Urban Design Management Measures' requires that the master urban design should aim at creating a livable urban space. However, to our best knowledge, the existing researches and practices of master urban design not only focus less on the livable space but also face a number of problems such as paying more attention to the image of the city, ignoring the people-oriented and lacking dynamic continuity. In order to make the master urban design can better guide the construction of city. Firstly, the paper proposes the livable city hierarchy system to meet the needs of different groups of people and then constructs the framework of livable goal-oriented master urban design based on the theory of livable content and the ideological origin of people-oriented. Secondly, the paper takes the master urban design practice of Zibo as a sample and puts forward the design strategy of strengthening the pattern, improve the quality of space, shape the feature, and establish a series of action plans based on the strategy of urban space development. Finally, the paper explores the method system of livable goal-oriented master urban design from the aspects of safety pattern, morphology pattern, neighborhood scale, open space, street space, public interface, style feature, public participation and action plans.

Keywords: livable, master urban design, public participation, zibo city

Procedia PDF Downloads 282
3183 Endometrioma Ethanol Sclerotherapy

Authors: Lamia Bensissaid

Abstract:

Goals: Endometriosis affects 6 to 10% of women of childbearing age. 17 to 44% of them have ovarian endometriomas. Medical and surgical treatments represent the two therapeutic axes with which PMA can be associated. Laparoscopic intraperitoneal ovarian cystectomy is described as the reference technique in the management of endometriomas by learned societies (CNGOF, ESHRE, NICE). However, it leads to a significant short-term reduction in the AMH level and the number of antral follicles, especially in cases of bilateral cystectomy, large cyst size or cystectomy after recurrence. Often, the disease is at an advanced stage with several surgical patients. Most have adhesions, which increase the risk of surgical complications and suboptimal resection and, therefore recurrence of the cyst. These results led to a change of opinion towards a conservative approach. Sclerotherapy is an old technique which acts by fibrinoid necrosis. It consists of injecting a sclerosing agent into the cyst cavity. Results : Recurrence was less than 15% for a 12-month follow-up; these rates are comparable to those of surgery. It does not seem to have a negative impact on ovarian reserve, but this is not sufficiently evaluated. It has an advantage in IVF pregnancy rates compared to cystectomy, particularly in cases of recurrent endometriomas. It has the advantages: · To be done on an outpatient basis. · To be inexpensive. · To avoid sometimes difficult and iterative surgery: · To allow an increase in pregnancy rates and the preservation of the ovarian reserve compared to iterative surgery. · of great interest in cases of bilateral endometriomas (kissing ovaries) or recurrent endometriomas. Conclusions: Ethanol sclerotherapy could be a good alternative to surgery.

Keywords: Endometrioma, Sclerotherapy, infertility, Ethanol

Procedia PDF Downloads 34
3182 Soil Organic Carbon Pool Assessment and Chemical Evaluation of Soils in Akure North and South Local Government Area of Ondo State

Authors: B. F. Dada, B. S. Ewulo, M. A. Awodun, S. O. Ajayi

Abstract:

Aggregate soil carbon distribution and stock in the soil in the form of a carbon pool is important for soil fertility and sequestration. The amount of carbon pool and other nutrients statues of the soil are to benefit plants, animal and the environment in the long run. This study was carried out at Akure North and South Local Government; the study area is one of the 18 Local Government Areas of Ondo State in the Southwest geo-political zone of Nigeria. The sites were divided into Map Grids and geo-referenced with Global Positioning System (GPS). Horizons were designated and morphological description carried out on the field. Pedons were characterized and classified according to USDA soil taxonomy. The local government area shares boundaries with; Ikere Local Government (LG) in the North, Ise Orun LG in the northwest, Ifedore LG in the northeast Akure South LG in the East, Ose LG in the South East, and Owo LG in the South. SOC-pool at Federal College of Agriculture topsoil horizon A2 is significantly higher than all horizons, 67.83 th⁻¹. The chemical properties of the pedons have shown that the soil is very strongly acidic to neutral reaction (4.68 – 6.73). The nutrients status of the soil topsoil A1 and A2 generally indicates that the soils have a low potential for retaining plant nutrients, and therefore call for adequate soil management.

Keywords: soil organic carbon (SOC), horizon, pedon, Akure

Procedia PDF Downloads 118
3181 Land Cover Classification System for the Estimation of Carbon Storage in Terrestrial Ecosystems

Authors: Lei Zhang

Abstract:

The carbon cycle greatly influences global change, and the land cover changes contribute to the status and rate of the carbon budget in ecosystems. This paper proposes a land cover classification system for mapping land cover, the national ecological environment assessment, and estimating carbon storage in ecosystems. The classification system consists of basic land cover classes at levels Ⅰ and Ⅱ and auxiliary features at level III. The basic 38 classes characterizing land cover features are derived from 19 criteria referring to composition, structure, pattern, phenology, etc. The basic classes reflect the status of carbon storage in ecosystems. The auxiliary classes at level III complement the attributes of higher levels by 9 criteria. The 5 environmental criteria of temperature, moisture, landform, aspect and slope mainly reflect the potential and intensity of carbon storage in ecosystems. The disturbance of vegetation succession caused by land use type influences the vegetation carbon budget. The other 3 vegetation cover criteria, growth period, and species characteristics further refine the vegetation types. The hierarchical structure of the land cover map (the classes of levels Ⅰ and Ⅱ) is independent of the products of level III, which is helpful for land cover product management and applications. The classification system has been adopted in the Chinese national land cover database for the carbon budget in ecosystems at a 30 m scale.

Keywords: classification system, land cover, ecosystem, carbon storage, object based

Procedia PDF Downloads 39
3180 X-Ray Detector Technology Optimization In CT Imaging

Authors: Aziz Ikhlef

Abstract:

Most of multi-slices CT scanners are built with detectors composed of scintillator - photodiodes arrays. The photodiodes arrays are mainly based on front-illuminated technology for detectors under 64 slices and on back-illuminated photodiode for systems of 64 slices or more. The designs based on back-illuminated photodiodes were being investigated for CT machines to overcome the challenge of the higher number of runs and connection required in front-illuminated diodes. In backlit diodes, the electronic noise has already been improved because of the reduction of the load capacitance due to the routing reduction. This translated by a better image quality in low signal application, improving low dose imaging in large patient population. With the fast development of multi-detector-rows CT (MDCT) scanners and the increasing number of examinations, the clinical community has raised significant concerns on radiation dose received by the patient in both medical and regulatory community. In order to reduce individual exposure and in response to the recommendations of the International Commission on Radiological Protection (ICRP) which suggests that all exposures should be kept as low as reasonably achievable (ALARA), every manufacturer is trying to implement strategies and solutions to optimize dose efficiency and image quality based on x-ray emission and scanning parameters. The added demands on the CT detector performance also comes from the increased utilization of spectral CT or dual-energy CT in which projection data of two different tube potentials are collected. One of the approaches utilizes a technology called fast-kVp switching in which the tube voltage is switched between 80kVp and 140kVp in fraction of a millisecond. To reduce the cross-contamination of signals, the scintillator based detector temporal response has to be extremely fast to minimize the residual signal from previous samples. In addition, this paper will present an overview of detector technologies and image chain improvement which have been investigated in the last few years to improve the signal-noise ratio and the dose efficiency CT scanners in regular examinations and in energy discrimination techniques. Several parameters of the image chain in general and in the detector technology contribute in the optimization of the final image quality. We will go through the properties of the post-patient collimation to improve the scatter-to-primary ratio, the scintillator material properties such as light output, afterglow, primary speed, crosstalk to improve the spectral imaging, the photodiode design characteristics and the data acquisition system (DAS) to optimize for crosstalk, noise and temporal/spatial resolution.

Keywords: computed tomography, X-ray detector, medical imaging, image quality, artifacts

Procedia PDF Downloads 236
3179 Vehicular Speed Detection Camera System Using Video Stream

Authors: C. A. Anser Pasha

Abstract:

In this paper, a new Vehicular Speed Detection Camera System that is applicable as an alternative to traditional radars with the same accuracy or even better is presented. The real-time measurement and analysis of various traffic parameters such as speed and number of vehicles are increasingly required in traffic control and management. Image processing techniques are now considered as an attractive and flexible method for automatic analysis and data collections in traffic engineering. Various algorithms based on image processing techniques have been applied to detect multiple vehicles and track them. The SDCS processes can be divided into three successive phases; the first phase is Objects detection phase, which uses a hybrid algorithm based on combining an adaptive background subtraction technique with a three-frame differencing algorithm which ratifies the major drawback of using only adaptive background subtraction. The second phase is Objects tracking, which consists of three successive operations - object segmentation, object labeling, and object center extraction. Objects tracking operation takes into consideration the different possible scenarios of the moving object like simple tracking, the object has left the scene, the object has entered the scene, object crossed by another object, and object leaves and another one enters the scene. The third phase is speed calculation phase, which is calculated from the number of frames consumed by the object to pass by the scene.

Keywords: radar, image processing, detection, tracking, segmentation

Procedia PDF Downloads 437
3178 Evaluation of Gesture-Based Password: User Behavioral Features Using Machine Learning Algorithms

Authors: Lakshmidevi Sreeramareddy, Komalpreet Kaur, Nane Pothier

Abstract:

Graphical-based passwords have existed for decades. Their major advantage is that they are easier to remember than an alphanumeric password. However, their disadvantage (especially recognition-based passwords) is the smaller password space, making them more vulnerable to brute force attacks. Graphical passwords are also highly susceptible to the shoulder-surfing effect. The gesture-based password method that we developed is a grid-free, template-free method. In this study, we evaluated the gesture-based passwords for usability and vulnerability. The results of the study are significant. We developed a gesture-based password application for data collection. Two modes of data collection were used: Creation mode and Replication mode. In creation mode (Session 1), users were asked to create six different passwords and reenter each password five times. In replication mode, users saw a password image created by some other user for a fixed duration of time. Three different duration timers, such as 5 seconds (Session 2), 10 seconds (Session 3), and 15 seconds (Session 4), were used to mimic the shoulder-surfing attack. After the timer expired, the password image was removed, and users were asked to replicate the password. There were 74, 57, 50, and 44 users participated in Session 1, Session 2, Session 3, and Session 4 respectfully. In this study, the machine learning algorithms have been applied to determine whether the person is a genuine user or an imposter based on the password entered. Five different machine learning algorithms were deployed to compare the performance in user authentication: namely, Decision Trees, Linear Discriminant Analysis, Naive Bayes Classifier, Support Vector Machines (SVMs) with Gaussian Radial Basis Kernel function, and K-Nearest Neighbor. Gesture-based password features vary from one entry to the next. It is difficult to distinguish between a creator and an intruder for authentication. For each password entered by the user, four features were extracted: password score, password length, password speed, and password size. All four features were normalized before being fed to a classifier. Three different classifiers were trained using data from all four sessions. Classifiers A, B, and C were trained and tested using data from the password creation session and the password replication with a timer of 5 seconds, 10 seconds, and 15 seconds, respectively. The classification accuracies for Classifier A using five ML algorithms are 72.5%, 71.3%, 71.9%, 74.4%, and 72.9%, respectively. The classification accuracies for Classifier B using five ML algorithms are 69.7%, 67.9%, 70.2%, 73.8%, and 71.2%, respectively. The classification accuracies for Classifier C using five ML algorithms are 68.1%, 64.9%, 68.4%, 71.5%, and 69.8%, respectively. SVMs with Gaussian Radial Basis Kernel outperform other ML algorithms for gesture-based password authentication. Results confirm that the shorter the duration of the shoulder-surfing attack, the higher the authentication accuracy. In conclusion, behavioral features extracted from the gesture-based passwords lead to less vulnerable user authentication.

Keywords: authentication, gesture-based passwords, machine learning algorithms, shoulder-surfing attacks, usability

Procedia PDF Downloads 82
3177 Corn Production in the Visayas: An Industry Study from 2002-2019

Authors: Julie Ann L. Gadin, Andrearose C. Igano, Carl Joseph S. Ignacio, Christopher C. Bacungan

Abstract:

Corn production has become an important and pervasive industry in the Visayas for many years. Its role as a substitute commodity to rice heightens demand for health-particular consumers. Unfortunately, the corn industry is confronted with several challenges, such as weak institutions. Considering these issues, the paper examined the factors that influence corn production in the three administrative regions in the Visayas, namely, Western Visayas, Central Visayas, and Eastern Visayas. The data used was retrieved from a variety of publicly available data sources such as the Philippine Statistics Authority, the Department of Agriculture, the Philippine Crop Insurance Corporation, and the International Disaster Database. Utilizing a dataset from 2002 to 2019, the indicators were tested using three multiple linear regression (MLR) models. Results showed that the land area harvested (p=0.02), and the value of corn production (p=0.00) are statistically significant variables that influence corn production in the Visayas. Given these findings, it is suggested that the policy of forest conversion and sustainable land management should be effective in enabling farmworkers to obtain land to grow corn crops, especially in rural regions. Furthermore, the Biofuels Act of 2006, the Livestock Industry Restructuring and Rationalization Act, and supported policy, Senate Bill No. 225, or an Act Establishing the Philippine Corn Research Institute and Appropriating Funds, should be enforced inclusively in order to improve the demand for the corn-allied industries which may lead to an increase in the value and volume of corn production in the Visayas.

Keywords: corn, industry, production, MLR, Visayas

Procedia PDF Downloads 160
3176 Groundwater Recharge Estimation of Fetam Catchment in Upper Blue Nile Basin North-Western Ethiopia

Authors: Mekonen G., Sileshi M., Melkamu M.

Abstract:

Recharge estimation is important for the assessment and management of groundwater resources effectively. This study applied the soil moisture balance and Baseflow separation methods to estimate groundwater recharge in the Fetam Catchment. It is one of the major catchments understudied from the different catchments in the upper Blue Nile River basin. Surface water has been subjected to high seasonal variation; due to this, groundwater is a primary option for drinking water supply to the community. This research has been conducted to estimate groundwater recharge by using fifteen years of River flow data for the Baseflow separation and ten years of daily meteorological data for the daily soil moisture balance recharge estimating method. The recharge rate by the two methods is 170.5 and 244.9mm/year daily soil moisture and baseflow separation method, respectively, and the average recharge is 207.7mm/year. The average value of annual recharge in the catchment is almost equal to the average recharge in the country, which is 200mm/year. So, each method has its own limitations, and taking the average value is preferable rather than taking a single value. Baseflow provides overestimated result compared to the average of the two, and soil moisture balance is the list estimator. The recharge estimation in the area also should be done by other recharge estimation methods.

Keywords: groundwater, recharge, baseflow separation, soil moisture balance, Fetam catchment

Procedia PDF Downloads 329
3175 Qualitative and Quantitative Characterization of Generated Waste in Nouri Petrochemical Complex, Assaluyeh, Iran

Authors: L. Heidari, M. Jalili Ghazizade

Abstract:

In recent years, different petrochemical complexes have been established to produce aromatic compounds. Among them, Nouri Petrochemical Complex (NPC) is the largest producer of aromatic raw materials in the world, and is located in south of Iran. Environmental concerns have been raised in this region due to generation of different types of solid waste generated in the process of aromatics production, and subsequently, industrial waste characterization has been thoroughly considered. The aim of this study is qualitative and quantitative characterization of industrial waste generated in the aromatics production process and determination of the best method for industrial waste management. For this purpose, all generated industrial waste during the production process was determined using a checklist. Four main industrial wastes were identified as follows: spent industrial soil, spent catalyst, spent molecular sieves and spent N-formyl morpholine (NFM) solvent. The amount of heavy metals and organic compounds in these wastes were further measured in order to identify the nature and toxicity of such a dangerous compound. Then industrial wastes were classified based on lab analysis results as well as using different international lists of hazardous waste identification such as EPA, UNEP and Basel Convention. Finally, the best method of waste disposal is selected based on environmental, economic and technical aspects. 

Keywords: aromatic compounds, industrial soil, molecular sieve, normal formyl morpholine solvent

Procedia PDF Downloads 209
3174 Students’ Speech Anxiety in Blended Learning

Authors: Mary Jane B. Suarez

Abstract:

Public speaking anxiety (PSA), also known as speech anxiety, is innumerably persistent in any traditional communication classes, especially for students who learn English as a second language. The speech anxiety intensifies when communication skills assessments have taken their toll in an online or a remote mode of learning due to the perils of the COVID-19 virus. Both teachers and students have experienced vast ambiguity on how to realize a still effective way to teach and learn speaking skills amidst the pandemic. Communication skills assessments like public speaking, oral presentations, and student reporting have defined their new meaning using Google Meet, Zoom, and other online platforms. Though using such technologies has paved for more creative ways for students to acquire and develop communication skills, the effectiveness of using such assessment tools stands in question. This mixed method study aimed to determine the factors that affected the public speaking skills of students in a communication class, to probe on the assessment gaps in assessing speaking skills of students attending online classes vis-à-vis the implementation of remote and blended modalities of learning, and to recommend ways on how to address the public speaking anxieties of students in performing a speaking task online and to bridge the assessment gaps based on the outcome of the study in order to achieve a smooth segue from online to on-ground instructions maneuvering towards a much better post-pandemic academic milieu. Using a convergent parallel design, both quantitative and qualitative data were reconciled by probing on the public speaking anxiety of students and the potential assessment gaps encountered in an online English communication class under remote and blended learning. There were four phases in applying the convergent parallel design. The first phase was the data collection, where both quantitative and qualitative data were collected using document reviews and focus group discussions. The second phase was data analysis, where quantitative data was treated using statistical testing, particularly frequency, percentage, and mean by using Microsoft Excel application and IBM Statistical Package for Social Sciences (SPSS) version 19, and qualitative data was examined using thematic analysis. The third phase was the merging of data analysis results to amalgamate varying comparisons between desired learning competencies versus the actual learning competencies of students. Finally, the fourth phase was the interpretation of merged data that led to the findings that there was a significantly high percentage of students' public speaking anxiety whenever students would deliver speaking tasks online. There were also assessment gaps identified by comparing the desired learning competencies of the formative and alternative assessments implemented and the actual speaking performances of students that showed evidence that public speaking anxiety of students was not properly identified and processed.

Keywords: blended learning, communication skills assessment, public speaking anxiety, speech anxiety

Procedia PDF Downloads 77
3173 Investigating the Algorithm to Maintain a Constant Speed in the Wankel Engine

Authors: Adam Majczak, Michał Bialy, Zbigniew Czyż, Zdzislaw Kaminski

Abstract:

Increasingly stringent emission standards for passenger cars require us to find alternative drives. The share of electric vehicles in the sale of new cars increases every year. However, their performance and, above all, range cannot be today successfully compared to those of cars with a traditional internal combustion engine. Battery recharging lasts hours, which can be hardly accepted due to the time needed to refill a fuel tank. Therefore, the ways to reduce the adverse features of cars equipped with electric motors only are searched for. One of the methods is a combination of an electric engine as a main source of power and a small internal combustion engine as an electricity generator. This type of drive enables an electric vehicle to achieve a radically increased range and low emissions of toxic substances. For several years, the leading automotive manufacturers like the Mazda and the Audi together with the best companies in the automotive industry, e.g., AVL have developed some electric drive systems capable of recharging themselves while driving, known as a range extender. An electricity generator is powered by a Wankel engine that has seemed to pass into history. This low weight and small engine with a rotating piston and a very low vibration level turned out to be an excellent source in such applications. Its operation as an energy source for a generator almost entirely eliminates its disadvantages like high fuel consumption, high emission of toxic substances, or short lifetime typical of its traditional application. The operation of the engine at a constant rotational speed enables a significant increase in its lifetime, and its small external dimensions enable us to make compact modules to drive even small urban cars like the Audi A1 or the Mazda 2. The algorithm to maintain a constant speed was investigated on the engine dynamometer with an eddy current brake and the necessary measuring apparatus. The research object was the Aixro XR50 rotary engine with the electronic power supply developed at the Lublin University of Technology. The load torque of the engine was altered during the research by means of the eddy current brake capable of giving any number of load cycles. The parameters recorded included speed and torque as well as a position of a throttle in an inlet system. Increasing and decreasing load did not significantly change engine speed, which means that control algorithm parameters are correctly selected. This work has been financed by the Polish Ministry of Science and Higher Education.

Keywords: electric vehicle, power generator, range extender, Wankel engine

Procedia PDF Downloads 132
3172 Specialized Instruction: Teaching and Leading Diverse Learners

Authors: Annette G. Walters Ph.D.

Abstract:

With a global shortage of qualified educational professionals, school systems continue to struggle with adequate staffing. How might learning communities meet the needs of all students, in particular those with specialized needs. While the task may seem foreboding and certain factors may seem divergent, all are connected in the education of students. Special education has a significant impact on the teaching and learning experience of all students in an educational community. Even when there are concerted efforts at embracing learners with diverse aptitude and abilities, there are often many important local factors that are misaligned, overlooked, or misunderstood. Working with learners with diverse abilities, often requires intentional services and supports for students to achieve success. Developing and implementing specialized instruction requires a multifaceted approach to supports the entire learning community, which includes educational providers, learners, and families, all while being mindful of fiscal and natural resources. This research explores the implications and complexities of special education instruction and specializing instruction, as well as leading and teaching diverse learners. This work is separated into three sections: the state of special education, teaching and leading diverse learners, and developing educational competencies through collaborative engagement. This structured analysis extrapolates historical and current research on special education practices and the role of educators in ensuring diverse students meet success.

Keywords: - diverse learners, - special education, - modification and supports, - curriculum and instruction, - classroom management, - formal and informal assessments

Procedia PDF Downloads 28
3171 Investigation of Cavitation in a Centrifugal Pump Using Synchronized Pump Head Measurements, Vibration Measurements and High-Speed Image Recording

Authors: Simon Caba, Raja Abou Ackl, Svend Rasmussen, Nicholas E. Pedersen

Abstract:

It is a challenge to directly monitor cavitation in a pump application during operation because of a lack of visual access to validate the presence of cavitation and its form of appearance. In this work, experimental investigations are carried out in an inline single-stage centrifugal pump with optical access. Hence, it gives the opportunity to enhance the value of CFD tools and standard cavitation measurements. Experiments are conducted using two impellers running in the same volute at 3000 rpm and the same flow rate. One of the impellers used is optimized for lower NPSH₃% by its blade design, whereas the other one is manufactured using a standard casting method. The cavitation is detected by pump performance measurements, vibration measurements and high-speed image recordings. The head drop and the pump casing vibration caused by cavitation are correlated with the visual appearance of the cavitation. The vibration data is recorded in an axial direction of the impeller using accelerometers recording at a sample rate of 131 kHz. The vibration frequency domain data (up to 20 kHz) and the time domain data are analyzed as well as the root mean square values. The high-speed recordings, focusing on the impeller suction side, are taken at 10,240 fps to provide insight into the flow patterns and the cavitation behavior in the rotating impeller. The videos are synchronized with the vibration time signals by a trigger signal. A clear correlation between cloud collapses and abrupt peaks in the vibration signal can be observed. The vibration peaks clearly indicate cavitation, especially at higher NPSHA values where the hydraulic performance is not affected. It is also observed that below a certain NPSHA value, the cavitation started in the inlet bend of the pump. Above this value, cavitation occurs exclusively on the impeller blades. The impeller optimized for NPSH₃% does show a lower NPSH₃% than the standard impeller, but the head drop starts at a higher NPSHA value and is more gradual. Instabilities in the head drop curve of the optimized impeller were observed in addition to a higher vibration level. Furthermore, the cavitation clouds on the suction side appear more unsteady when using the optimized impeller. The shape and location of the cavitation are compared to 3D fluid flow simulations. The simulation results are in good agreement with the experimental investigations. In conclusion, these investigations attempt to give a more holistic view on the appearance of cavitation by comparing the head drop, vibration spectral data, vibration time signals, image recordings and simulation results. Data indicates that a criterion for cavitation detection could be derived from the vibration time-domain measurements, which requires further investigation. Usually, spectral data is used to analyze cavitation, but these investigations indicate that the time domain could be more appropriate for some applications.

Keywords: cavitation, centrifugal pump, head drop, high-speed image recordings, pump vibration

Procedia PDF Downloads 159
3170 The Design of a Computer Simulator to Emulate Pathology Laboratories: A Model for Optimising Clinical Workflows

Authors: M. Patterson, R. Bond, K. Cowan, M. Mulvenna, C. Reid, F. McMahon, P. McGowan, H. Cormican

Abstract:

This paper outlines the design of a simulator to allow for the optimisation of clinical workflows through a pathology laboratory and to improve the laboratory’s efficiency in the processing, testing, and analysis of specimens. Often pathologists have difficulty in pinpointing and anticipating issues in the clinical workflow until tests are running late or in error. It can be difficult to pinpoint the cause and even more difficult to predict any issues which may arise. For example, they often have no indication of how many samples are going to be delivered to the laboratory that day or at a given hour. If we could model scenarios using past information and known variables, it would be possible for pathology laboratories to initiate resource preparations, e.g. the printing of specimen labels or to activate a sufficient number of technicians. This would expedite the clinical workload, clinical processes and improve the overall efficiency of the laboratory. The simulator design visualises the workflow of the laboratory, i.e. the clinical tests being ordered, the specimens arriving, current tests being performed, results being validated and reports being issued. The simulator depicts the movement of specimens through this process, as well as the number of specimens at each stage. This movement is visualised using an animated flow diagram that is updated in real time. A traffic light colour-coding system will be used to indicate the level of flow through each stage (green for normal flow, orange for slow flow, and red for critical flow). This would allow pathologists to clearly see where there are issues and bottlenecks in the process. Graphs would also be used to indicate the status of specimens at each stage of the process. For example, a graph could show the percentage of specimen tests that are on time, potentially late, running late and in error. Clicking on potentially late samples will display more detailed information about those samples, the tests that still need to be performed on them and their urgency level. This would allow any issues to be resolved quickly. In the case of potentially late samples, this could help to ensure that critically needed results are delivered on time. The simulator will be created as a single-page web application. Various web technologies will be used to create the flow diagram showing the workflow of the laboratory. JavaScript will be used to program the logic, animate the movement of samples through each of the stages and to generate the status graphs in real time. This live information will be extracted from an Oracle database. As well as being used in a real laboratory situation, the simulator could also be used for training purposes. ‘Bots’ would be used to control the flow of specimens through each step of the process. Like existing software agents technology, these bots would be configurable in order to simulate different situations, which may arise in a laboratory such as an emerging epidemic. The bots could then be turned on and off to allow trainees to complete the tasks required at that step of the process, for example validating test results.

Keywords: laboratory-process, optimization, pathology, computer simulation, workflow

Procedia PDF Downloads 257
3169 Revisiting the Surgical Approaches to Decompression in Quadrangular Space Syndrome: A Cadaveric Study

Authors: Sundip Charmode, Simmi Mehra, Sudhir Kushwaha, Shalom Philip, Pratik Amrutiya, Ranjna Jangal

Abstract:

Introduction: Quadrangular space syndrome involves compression of the axillary nerve and posterior circumflex humeral artery and its management in few cases, requires surgical decompression. The current study reviews the surgical approaches used in the decompression of neurovascular structures and presents our reflections and recommendations. Methods: Four human cadavers, in the Department of Anatomy were used for dissection of the Axillae and the Scapular region by the senior residents of the Department of Anatomy and Department of Orthopedics, who dissected quadrangular space in the eight upper limbs, using anterior and posterior surgical approaches. Observations: Posterior approach to identify the quadrangular space and secure its contents was recognized as the easier and much quicker method by both the Anatomy and Orthopedic residents, but it may result in increased postoperative morbidity. Whereas the anterior (Delto-pectoral) approach involves more skill but reduces postoperative morbidity. Conclusions: Anterior (Delto-pectoral) approach with suggested modifications can prove as an effective method in surgical decompression of quadrangular space syndrome. The authors suggest more cadaveric studies to facilitate anatomists and surgeons with the opportunities to practice and evaluate older and newer surgical approaches.

Keywords: surgical approach, anatomical approach, decompression, axillary nerve, quadrangular space

Procedia PDF Downloads 145
3168 Functions and Pathophysiology of the Ventricular System: Review of the Underlying Basic Physics

Authors: Mohamed Abdelrahman Abdalla

Abstract:

Apart from their function in producing CSF, the brain ventricles have been recognized as the mere remnant of the embryological neural tube with no clear role. The lack of proper definition of the function of the brain ventricles and the central spinal canal has made it difficult to ascertain the pathophysiology of its different disease conditions or to treat them. This study aims to review the simple physics that could explain the basic function of the CNS ventricular system and to suggest new ways of approaching its pathology. There are probably more physical factors to consider than only the pressure. Monro-Killie hypothesis focuses on volume and subsequently pressure to direct our surgical management in different disease conditions. However, the enlarged volume of the ventricles in normal pressure hydrocephalus does not move any blood or brain outside the skull. Also, in idiopathic intracranial hypertension, the very high intracranial pressure rarely causes brain herniation. On this note, the continuum of the intracranial cavity with the spinal canal makes it a whole unit and hence the defect in the theory. In this study, adding different factors to the equation like brain and CSF density and positions of the brain in space, in addition to the volume and pressure, aims to identify how the ventricles are important in the CNS homeostasis. In addition, increasing the variables that we analyze to treat different CSF pathological conditions should increase our understanding and hence accuracy of treatment of such conditions.

Keywords: communicating hydrocephalus, functions of the ventricles, idiopathic intracranial hypertension physics of CSF

Procedia PDF Downloads 76
3167 Overcoming Reading Barriers in an Inclusive Mathematics Classroom with Linguistic and Visual Support

Authors: A. Noll, J. Roth, M. Scholz

Abstract:

The importance of written language in a democratic society is non-controversial. Students with physical, learning, cognitive or developmental disabilities often have difficulties in understanding information which is presented in written language only. These students suffer from obstacles in diverse domains. In order to reduce such barriers in educational as well as in out-of-school areas, access to written information must be facilitated. Readability can be enhanced by linguistic simplifications like the application of easy-to-read language. Easy-to-read language shall help people with disabilities to participate socially and politically in society. The authors state, for example, that only short simple words should be used, whereas the occurrence of complex sentences should be avoided. So far, these guidelines were not empirically proved. Another way to reduce reading barriers is the use of visual support, for example, symbols. A symbol conveys, in contrast to a photo, a single idea or concept. Little empirical data about the use of symbols to foster the readability of texts exist. Nevertheless, a positive influence can be assumed, e.g., because of the multimedia principle. It indicates that people learn better from words and pictures than from words alone. A qualitative Interview and Eye-Tracking-Study, which was conducted by the authors, gives cause for the assumption that besides the illustration of single words, the visualization of complete sentences may be helpful. Thus, the effect of photos, which illustrate the content of complete sentences, is also investigated in this study. This leads us to the main research question which was focused on: Does the use of easy-to-read language and/or enriching text with symbols or photos facilitate pupils’ comprehension of learning tasks? The sample consisted of students with learning difficulties (N = 144) and students without SEN (N = 159). The students worked on the tasks, which dealt with introducing fractions, individually. While experimental group 1 received a linguistically simplified version of the tasks, experimental group 2 worked with a variation which was linguistically simplified and furthermore, the keywords of the tasks were visualized by symbols. Experimental group 3 worked on exercises which were simplified by easy-to-read-language and the content of the whole sentences was illustrated by photos. Experimental group 4 received a not simplified version. The participants’ reading ability and their IQ was elevated beforehand to build four comparable groups. There is a significant effect of the different setting on the students’ results F(3,140) = 2,932; p = 0,036*. A post-hoc-analyses with multiple comparisons shows that this significance results from the difference between experimental group 3 and 4. The students in the group easy-to-read language plus photos worked on the exercises significantly more successfully than the students who worked in the group with no simplifications. Further results which refer, among others, to the influence of the students reading ability will be presented at the ICERI 2018.

Keywords: inclusive education, mathematics education, easy-to-read language, photos, symbols, special educational needs

Procedia PDF Downloads 131
3166 The Direct Deconvolution Model for the Large Eddy Simulation of Turbulence

Authors: Ning Chang, Zelong Yuan, Yunpeng Wang, Jianchun Wang

Abstract:

Large eddy simulation (LES) has been extensively used in the investigation of turbulence. LES calculates the grid-resolved large-scale motions and leaves small scales modeled by sub lfilterscale (SFS) models. Among the existing SFS models, the deconvolution model has been used successfully in the LES of the engineering flows and geophysical flows. Despite the wide application of deconvolution models, the effects of subfilter scale dynamics and filter anisotropy on the accuracy of SFS modeling have not been investigated in depth. The results of LES are highly sensitive to the selection of fi lters and the anisotropy of the grid, which has been overlooked in previous research. In the current study, two critical aspects of LES are investigated. Firstly, we analyze the influence of sub-fi lter scale (SFS) dynamics on the accuracy of direct deconvolution models (DDM) at varying fi lter-to-grid ratios (FGR) in isotropic turbulence. An array of invertible filters are employed, encompassing Gaussian, Helmholtz I and II, Butterworth, Chebyshev I and II, Cauchy, Pao, and rapidly decaying filters. The signi ficance of FGR becomes evident, as it acts as a pivotal factor in error control for precise SFS stress prediction. When FGR is set to 1, the DDM models cannot accurately reconstruct the SFS stress due to the insufficient resolution of SFS dynamics. Notably, prediction capabilities are enhanced at an FGR of 2, resulting in accurate SFS stress reconstruction, except for cases involving Helmholtz I and II fi lters. A remarkable precision close to 100% is achieved at an FGR of 4 for all DDM models. Additionally, the further exploration extends to the fi lter anisotropy to address its impact on the SFS dynamics and LES accuracy. By employing dynamic Smagorinsky model (DSM), dynamic mixed model (DMM), and direct deconvolution model (DDM) with the anisotropic fi lter, aspect ratios (AR) ranging from 1 to 16 in LES fi lters are evaluated. The findings highlight the DDM's pro ficiency in accurately predicting SFS stresses under highly anisotropic filtering conditions. High correlation coefficients exceeding 90% are observed in the a priori study for the DDM's reconstructed SFS stresses, surpassing those of the DSM and DMM models. However, these correlations tend to decrease as lter anisotropy increases. In the a posteriori studies, the DDM model consistently outperforms the DSM and DMM models across various turbulence statistics, encompassing velocity spectra, probability density functions related to vorticity, SFS energy flux, velocity increments, strain-rate tensors, and SFS stress. It is observed that as fi lter anisotropy intensify , the results of DSM and DMM become worse, while the DDM continues to deliver satisfactory results across all fi lter-anisotropy scenarios. The fi ndings emphasize the DDM framework's potential as a valuable tool for advancing the development of sophisticated SFS models for LES of turbulence.

Keywords: deconvolution model, large eddy simulation, subfilter scale modeling, turbulence

Procedia PDF Downloads 48
3165 Mathematical Study for Traffic Flow and Traffic Density in Kigali Roads

Authors: Kayijuka Idrissa

Abstract:

This work investigates a mathematical study for traffic flow and traffic density in Kigali city roads and the data collected from the national police of Rwanda in 2012. While working on this topic, some mathematical models were used in order to analyze and compare traffic variables. This work has been carried out on Kigali roads specifically at roundabouts from Kigali Business Center (KBC) to Prince House as our study sites. In this project, we used some mathematical tools to analyze the data collected and to understand the relationship between traffic variables. We applied the Poisson distribution method to analyze and to know the number of accidents occurred in this section of the road which is from KBC to Prince House. The results show that the accidents that occurred in 2012 were at very high rates due to the fact that this section has a very narrow single lane on each side which leads to high congestion of vehicles, and consequently, accidents occur very frequently. Using the data of speeds and densities collected from this section of road, we found that the increment of the density results in a decrement of the speed of the vehicle. At the point where the density is equal to the jam density the speed becomes zero. The approach is promising in capturing sudden changes on flow patterns and is open to be utilized in a series of intelligent management strategies and especially in noncurrent congestion effect detection and control.

Keywords: statistical methods, traffic flow, Poisson distribution, car moving technics

Procedia PDF Downloads 254
3164 Standalone Docking Station with Combined Charging Methods for Agricultural Mobile Robots

Authors: Leonor Varandas, Pedro D. Gaspar, Martim L. Aguiar

Abstract:

One of the biggest concerns in the field of agriculture is around the energy efficiency of robots that will perform agriculture’s activity and their charging methods. In this paper, two different charging methods for agricultural standalone docking stations are shown that will take into account various variants as field size and its irregularities, work’s nature to which the robot will perform, deadlines that have to be respected, among others. Its features also are dependent on the orchard, season, battery type and its technical specifications and cost. First charging base method focuses on wireless charging, presenting more benefits for small field. The second charging base method relies on battery replacement being more suitable for large fields, thus avoiding the robot stop for recharge. Existing many methods to charge a battery, the CC CV was considered the most appropriate for either simplicity or effectiveness. The choice of the battery for agricultural purposes is if most importance. While the most common battery used is Li-ion battery, this study also discusses the use of graphene-based new type of batteries with 45% over capacity to the Li-ion one. A Battery Management Systems (BMS) is applied for battery balancing. All these approaches combined showed to be a promising method to improve a lot of technical agricultural work, not just in terms of plantation and harvesting but also about every technique to prevent harmful events like plagues and weeds or even to reduce crop time and cost.

Keywords: agricultural mobile robot, charging methods, battery replacement method, wireless charging method

Procedia PDF Downloads 124
3163 Embracing the Uniqueness and Potential of Each Child: Moving Theory to Practice

Authors: Joy Chadwick

Abstract:

This Study of Teaching and Learning (SoTL) research focused on the experiences of teacher candidates involved in an inclusive education methods course within a four-year direct entry Bachelor of Education program. The placement of this course within the final fourteen-week practicum semester is designed to facilitate deeper theory-practice connections between effective inclusive pedagogical knowledge and the real life of classroom teaching. The course focuses on supporting teacher candidates to understand that effective instruction within an inclusive classroom context must be intentional, responsive, and relational. Diversity is situated not as exceptional but rather as expected. This interpretive qualitative study involved the analysis of twenty-nine teacher candidate reflective journals and six individual teacher candidate semi-structured interviews. The journal entries were completed at the start of the semester and at the end of the semester with the intent of having teacher candidates reflect on their beliefs of what it means to be an effective inclusive educator and how the course and practicum experiences impacted their understanding and approaches to teaching in inclusive classrooms. The semi-structured interviews provided further depth and context to the journal data. The journals and interview transcripts were coded and themed using NVivo software. The findings suggest that instructional frameworks such as universal design for learning (UDL), differentiated instruction (DI), response to intervention (RTI), social emotional learning (SEL), and self-regulation supported teacher candidate’s abilities to meet the needs of their students more effectively. Course content that focused on specific exceptionalities also supported teacher candidates to be proactive rather than reactive when responding to student learning challenges. Teacher candidates also articulated the importance of reframing their perspective about students in challenging moments and that seeing the individual worth of each child was integral to their approach to teaching. A persisting question for teacher educators exists as to what pedagogical knowledge and understanding is most relevant in supporting future teachers to be effective at planning for and embracing the diversity of student needs within classrooms today. This research directs us to consider the critical importance of addressing personal attributes and mindsets of teacher candidates regarding children as well as considering instructional frameworks when designing coursework. Further, the alignment of an inclusive education course during a teaching practicum allows for an iterative approach to learning. The practical application of course concepts while teaching in a practicum allows for a deeper understanding of instructional frameworks, thus enhancing the confidence of teacher candidates. Research findings have implications for teacher education programs as connected to inclusive education methods courses, practicum experiences, and overall teacher education program design.

Keywords: inclusion, inclusive education, pre-service teacher education, practicum experiences, teacher education

Procedia PDF Downloads 40
3162 Students’ Online Forum Activities and Social Network Analysis in an E-Learning Environment

Authors: P. L. Cheng, I. N. Umar

Abstract:

Online discussion forum is a popular e-learning technique that allows participants to interact and construct knowledge. This study aims to examine the levels of participation, categories of participants and the structure of their interactions in a forum. A convenience sampling of one course coordinator and 23 graduate students was selected in this study. The forums’ log file and the Social Network Analysis software were used in this study. The analysis reveals 610 activities (including viewing forum’s topic, viewing discussion thread, posting a new thread, replying to other participants’ post, updating an existing thread and deleting a post) performed by them in this forum, with an average of 3.83 threads posted. Also, this forum consists of five at-risk participants, six bridging participants, four isolated participants and five leaders of information. In addition, the network density value is 0.15 and there exist five reciprocal interactions in this forum. The closeness value varied between 28 and 68 while the eigen vector centrality value varied between 0.008 and 0.39. The finding indicates that the participants tend to listen more rather than express their opinions in the forum. It was also revealed that those who actively provide supports in the discussion forum were not the same people who received the most responses from their peers. This study found that cliques do not exist in the forum and the participants are not selective to whom they response to, rather, it was based on the content of the posts made by their peers. Based upon the findings, further analysis with different method and population, larger sample size and a longer time frame are recommended.

Keywords: e-learning, learning management system, online forum, social network analysis

Procedia PDF Downloads 366
3161 Central Palmar Necrosis Following Steroid Injections for the Treatment of Carpal Tunnel Syndrome: A Case Report

Authors: M. Ridwanul Hassan, Samuel George

Abstract:

Aims: Steroid injections are commonly used as a diagnostic tool or an alternative to surgical management of carpal tunnel syndrome (CTS) and are generally safe. Ischaemia is a rare complication with very few cases reported in the literature. Methods: We report a case of a 50-year-old female that presented with a necrotic wound to her left palm one month after a steroid injection into the carpal tunnel. She had a 2-year history of CTS in her left hand that was treated with six previous steroid injections in primary care during this period. The wound evolved from a blister to a necrotic ulcer which led to a painful, hollow defect in the centre of her palm. She did not report any history of trauma, nor did she have any co-morbidities. Clinical photographs were taken. Results: On examination, she had a 0.5 cmx1 cm defect in the palm of her left hand down to aponeurosis. There was purulent discharge in the wound with surrounding erythema but no spreading cellulitis. She had full function of her fingers but was very tender on movements and at rest. She was admitted for intravenous antibiotics and underwent a debridement, washout, and carpal tunnel release the next day. The defect was packed to heal by secondary intention and has now fully healed one month following her operation. Conclusions: This is an extremely rare complication of steroid injections to the carpal tunnel and may have been avoided by earlier referral for surgery rather than treatment using multiple steroid injections.

Keywords: hand surgery, complication, rare, carpal tunnel syndrome

Procedia PDF Downloads 89
3160 Physical Contact Modulation of Macrophage-Mediated Anti-Inflammatory Response in Osteoimmune Microenvironment by Pollen-Like Nanoparticles

Authors: Qing Zhang, Janak L. Pathak, Macro N. Helder, Richard T. Jaspers, Yin Xiao

Abstract:

Introduction: Nanomaterial-based bone regeneration is greatly influenced by the immune microenvironment. Tissue-engineered nanomaterials mediate the inflammatory response of macrophages to regulate bone regeneration. Silica nanoparticles have been widely used in tissue engineering-related preclinical studies. However, the effect of topological features on the surface of silica nanoparticles on the immune response of macrophages remains unknown. Purposes: The aims of this research are to compare the influences of normal and pollen-like silica nano-surface topography on macrophage immune responses and to obtain insight into their potential regulatory mechanisms. Method: Macrophages (RAW 264.7 cells) were exposed to mesoporous silica nanoparticles with normal morphology (MSNs) and pollen-like morphology (PMSNs). RNA-seq, RT-qPCR, and LSCM were used to assess the changes in expression levels of immune response-related genes and proteins. SEM and TEM were executed to evaluate the contact and adherence of silica nanoparticles by macrophages. For the assessment of the immunomodulation-mediated osteogenic potential, BMSCs were cultured with conditioned medium (CM) from LPS pre-stimulated macrophage cultures treated with MSNs or PMSNs. Osteoimmunomodulatory potential of MSNs and PMSNs in vivo was tested in a mouse cranial bone osteolysis model. Results: The results of the RNA-seq, RT-qPCR, and LSCM assays showed that PMSNs inhibited the expression of pro-inflammatory genes and proteins in macrophages. SEM images showed distinct macrophage membrane surface binding patterns of MSNs and PMSNs. MSNs were more evenly dispersed across the macrophage cell membrane, while PMSNs were aggregated. PMSNs-induced macrophage anti-inflammatory response was associated with upregulation of the cell surface receptor CD28 and inhibition of ERK phosphorylation. TEM images showed that both MSNs and PMSNs could be phagocytosed by macrophages, and inhibiting nanoparticle phagocytosis did not affect the expression of anti-inflammatory genes and proteins. Moreover, PMSNs-induced conditioned medium from macrophages enhanced BMP-2 expression and osteogenic differentiation mBMSCs. Similarly, PMSNs prevented LPS-induced bone resorption via downregulation of inflammatory reaction. Conclusions: PMSNs can promote bone regeneration by modulating osteoimmunological processes through surface topography. The study offers insights into how surface physical contact cues can modulate the regulation of osteoimmunology and provides a basis for the application of nanoparticles with pollen-like morphology to affect immunomodulation in bone tissue engineering and regeneration.

Keywords: physical contact, osteoimmunology, macrophages, silica nanoparticles, surface morphology, membrane receptor, osteogenesis, inflammation

Procedia PDF Downloads 31
3159 A Hybrid Algorithm Based on Greedy Randomized Adaptive Search Procedure and Chemical Reaction Optimization for the Vehicle Routing Problem with Hard Time Windows

Authors: Imen Boudali, Marwa Ragmoun

Abstract:

The Vehicle Routing Problem with Hard Time Windows (VRPHTW) is a basic distribution management problem that models many real-world problems. The objective of the problem is to deliver a set of customers with known demands on minimum-cost vehicle routes while satisfying vehicle capacity and hard time windows for customers. In this paper, we propose to deal with our optimization problem by using a new hybrid stochastic algorithm based on two metaheuristics: Chemical Reaction Optimization (CRO) and Greedy Randomized Adaptive Search Procedure (GRASP). The first method is inspired by the natural process of chemical reactions enabling the transformation of unstable substances with excessive energy to stable ones. During this process, the molecules interact with each other through a series of elementary reactions to reach minimum energy for their existence. This property is embedded in CRO to solve the VRPHTW. In order to enhance the population diversity throughout the search process, we integrated the GRASP in our method. Simulation results on the base of Solomon’s benchmark instances show the very satisfactory performances of the proposed approach.

Keywords: Benchmark Problems, Combinatorial Optimization, Vehicle Routing Problem with Hard Time Windows, Meta-heuristics, Hybridization, GRASP, CRO

Procedia PDF Downloads 383
3158 Long Standing Orbital Floor Fracture Repair: Case Report

Authors: Hisham A. Hashem, Sameh Galal, Bassem M. Moeshed

Abstract:

A 36 years old male patient presented to our unit with a history of motor-car accident from 7 months complaining of disfigurement and double vision. On examination and investigations, there was an orbital floor fracture in the left eye with inferior rectus muscle entrapment causing diplopia, dystopia and enophthalmos. Under general anesthesia, a sub-ciliary incision was performed, and the orbital floor fracture was repaired with a double layer Medpor sheet (30x50x15) with removing and freeing fibrosis that was present and freeing of the inferior rectus muscle. Remarkable improvement of the dystopia was noticed, however, there was a residual diplopia in upgaze and enophthalmos. He was then referred to a strabismologist, which upon examination found left hypotropia of 8 ΔD corrected by 8 ΔD base up prism and positive forced duction test on elevation and pseudoptosis. Under local anesthesia, a limbal incision approach with hangback 4mm recession of inferior rectus muscle was performed after identifying an inferior rectus muscle structure. Improvement was noted shortly postoperative with correction of both diplopia and pseudoptosis. Follow up after 1, 4 and 8 months was done showing a stable condition. Delayed surgery in cases of orbital floor fracture may still hold good results provided proper assessment of the case with management of each sign separately.

Keywords: diplopia, dystopia, late surgery, orbital floor fracture

Procedia PDF Downloads 208
3157 Hybrid Solutions in Physicochemical Processes for the Removal of Turbidity in Andean Reservoirs

Authors: María Cárdenas Gaudry, Gonzalo Ramces Fano Miranda

Abstract:

Sediment removal is very important in the purification of water, not only for reasons of visual perception but also because of its association with odor and taste problems. The Cuchoquesera reservoir, which is in the Andean region of Ayacucho (Peru) at an altitude of 3,740 meters above sea level, visually presents suspended particles and organic impurities indicating that it contains water of dubious quality to deduce that it is suitable for direct consumption of human beings. In order to quantitatively know the degree of impurities, water quality monitoring was carried out from February to August 2018, in which four sampling stations were established in the reservoir. The selected measured parameters were electrical conductivity, total dissolved solids, pH, color, turbidity, and sludge volume. The indicators of the studied parameters exceed the permissible limits except for electrical conductivity (190 μS/cm) and total dissolved solids (255 mg/L). In this investigation, the best combination and the optimal doses of reagents were determined that allowed the removal of sediments from the waters of the Cuchoquesera reservoir, through the physicochemical process of coagulation-flocculation. In order to improve this process during the rainy season, six combinations of reagents were evaluated, made up of three coagulants (ferric chloride, ferrous sulfate, and aluminum sulfate) and two natural flocculants: prickly pear powder (Opuntia ficus-indica) and tara gum (Caesalpinia spinoza). For each combination of reagents, jar tests were developed following the central composite experimental design (CCED), where the design factors were the doses of coagulant and flocculant and the initial turbidity. The results of the jar tests were adjusted to mathematical models, obtaining that to treat the water from the Cuchoquesera reservoir, with a turbidity of 150 UTN and a color of 137 U Pt-Co, 27.9 mg/L of the coagulant aluminum sulfate with 3 mg/L of the natural tara gum flocculant to produce a purified water quality of 1.7 UTN of turbidity and 3.2 U Pt-Co of apparent color. The estimated cost of the dose of coagulant and flocculant found was 0.22 USD/m³. This is how “grey-green” technologies can be used as a combination in nature-based solutions in water treatment, in this case, to achieve potability, making it more sustainable, especially economically, if green technology is available at the site of application of the nature-based hybrid solution. This research is a demonstration of the compatibility of natural coagulants/flocculants with other treatment technologies in the integrated/hybrid treatment process, such as the possibility of hybridizing natural coagulants with other types of coagulants.

Keywords: prickly pear powder, tara gum, nature-based solutions, aluminum sulfate, jar test, turbidity, coagulation, flocculation

Procedia PDF Downloads 82
3156 Pinch Technology for Minimization of Water Consumption at a Refinery

Authors: W. Mughees, M. Alahmad

Abstract:

Water is the most significant entity that controls local and global development. For the Gulf region, especially Saudi Arabia, with its limited potable water resources, the potential of the fresh water problem is highly considerable. In this research, the study involves the design and analysis of pinch-based water/wastewater networks. Multiple water/wastewater networks were developed using pinch analysis involving direct recycle/material recycle method. Property-integration technique was adopted to carry out direct recycle method. Particularly, a petroleum refinery was considered as a case study. In direct recycle methodology, minimum water discharge and minimum fresh water resource targets were estimated. Re-design (or retrofitting) of water allocation in the networks was undertaken. Chemical Oxygen Demand (COD) and hardness properties were taken as pollutants. This research was based on single and double contaminant approach for COD and hardness and the amount of fresh water was reduced from 340.0 m3/h to 149.0 m3/h (43.8%), 208.0 m3/h (61.18%) respectively. While regarding double contaminant approach, reduction in fresh water demand was 132.0 m3/h (38.8%). The required analysis was also carried out using mathematical programming technique. Operating software such as LINGO was used for these studies which have verified the graphical method results in a valuable and accurate way. Among the multiple water networks, the one possible water allocation network was developed based on mass exchange.

Keywords: minimization, water pinch, water management, pollution prevention

Procedia PDF Downloads 452
3155 Substation Automation, Digitization, Cyber Risk and Chain Risk Management Reliability

Authors: Serzhan Ashirov, Dana Nour, Rafat Rob, Khaled Alotaibi

Abstract:

There has been a fast growth in the introduction and use of communications, information, monitoring, and sensing technologies. The new technologies are making their way to the Industrial Control Systems as embedded in products, software applications, IT services, or commissioned to enable integration and automation of increasingly global supply chains. As a result, the lines that separated the physical, digital, and cyber world have diminished due to the vast implementation of the new, disruptive digital technologies. The variety and increased use of these technologies introduce many cybersecurity risks affecting cyber-resilience of the supply chain, both in terms of the product or service delivered to a customer and members of the supply chain operation. US department of energy considers supply chain in the IR4 space to be the weakest link in cybersecurity. The IR4 identified the digitization of the field devices, followed by digitalization that eventually moved through the digital transformation space with little care for the new introduced cybersecurity risks. This paper will examine the best methodologies for securing the electrical substations from cybersecurity attacks due to supply chain risks, and due to digitization effort. SCADA systems are the most vulnerable part of the power system infrastructure due to digitization and due to the weakness and vulnerabilities in the supply chain security. The paper will discuss in details how create a secure supply chain methodology, secure substations, and mitigate the risks due to digitization

Keywords: cybersecurity, supply chain methodology, secure substation, digitization

Procedia PDF Downloads 43