Search results for: information extraction evaluation method
31390 The Flavonoids for a Plant Grows in the Arid and Semi-Arid Zone of the Northern Sahara of Algeria - Atriplex halimus L.
Authors: O. Smara, H. Dendougui, B. Legseir
Abstract:
Atriplex halimus L. is particularly well adapted to arid and salt-affected areas. In this species, salinity resistance is often attributed to the presence of vesiculated hairs covering leaf surface and containing a large amount of salt. Atriplex halimus L. (Chenopodiaceae) is a perennial shrub native to the Mediterranean basin with excellent tolerance to drought and salinity. The species is present in semiarid to subhumid areas of the north Mediterranean and in arid zones from North Africa and the eastern Mediterranean. The main aim of this study was to identify a medicinal plant used in the Ouargla (Est-southern Algeria) for the treatment of several human pathologies. This plant is an important source for livestock in nitrogenous matter, it is an effective and relatively inexpensive tool in the fight against erosion and desertification and rehabilitation of degraded lands. Phytochemical investigation is applied to the majority of extracts of the powder of the aerial parts of Atriplex halimus L. Different chromatographic methods after liquid-liquid extraction are used; it is the thin layer chromatography (TLC) and paper using multiple systems and chemical revelations. This study followed by an evaluation by the phenol assay the Folin-Ciocalteu method, using gallic acid as a reference for phenols and quercetin for flavonols. Some polar extracts showed an interesting result better than the less polar extracts.Keywords: Atriples halimus L., chenopodiaceae, flavonoids, phenols
Procedia PDF Downloads 30431389 Bitplanes Gray-Level Image Encryption Approach Using Arnold Transform
Authors: Ali Abdrhman M. Ukasha
Abstract:
Data security needed in data transmission, storage, and communication to ensure the security. The single step parallel contour extraction (SSPCE) method is used to create the edge map as a key image from the different Gray level/Binary image. Performing the X-OR operation between the key image and each bit plane of the original image for image pixel values change purpose. The Arnold transform used to changes the locations of image pixels as image scrambling process. Experiments have demonstrated that proposed algorithm can fully encrypt 2D Gary level image and completely reconstructed without any distortion. Also shown that the analyzed algorithm have extremely large security against some attacks like salt & pepper and JPEG compression. Its proof that the Gray level image can be protected with a higher security level. The presented method has easy hardware implementation and suitable for multimedia protection in real time applications such as wireless networks and mobile phone services.Keywords: SSPCE method, image compression-salt- peppers attacks, bitplanes decomposition, Arnold transform, lossless image encryption
Procedia PDF Downloads 43531388 Multi-Criteria Optimal Management Strategy for in-situ Bioremediation of LNAPL Contaminated Aquifer Using Particle Swarm Optimization
Authors: Deepak Kumar, Jahangeer, Brijesh Kumar Yadav, Shashi Mathur
Abstract:
In-situ remediation is a technique which can remediate either surface or groundwater at the site of contamination. In the present study, simulation optimization approach has been used to develop management strategy for remediating LNAPL (Light Non-Aqueous Phase Liquid) contaminated aquifers. Benzene, toluene, ethyl benzene and xylene are the main component of LNAPL contaminant. Collectively, these contaminants are known as BTEX. In in-situ bioremediation process, a set of injection and extraction wells are installed. Injection wells supply oxygen and other nutrient which convert BTEX into carbon dioxide and water with the help of indigenous soil bacteria. On the other hand, extraction wells check the movement of plume along downstream. In this study, optimal design of the system has been done using PSO (Particle Swarm Optimization) algorithm. A comprehensive management strategy for pumping of injection and extraction wells has been done to attain a maximum allowable concentration of 5 ppm and 4.5 ppm. The management strategy comprises determination of pumping rates, the total pumping volume and the total running cost incurred for each potential injection and extraction well. The results indicate a high pumping rate for injection wells during the initial management period since it facilitates the availability of oxygen and other nutrients necessary for biodegradation, however it is low during the third year on account of sufficient oxygen availability. This is because the contaminant is assumed to have biodegraded by the end of the third year when the concentration drops to a permissible level.Keywords: groundwater, in-situ bioremediation, light non-aqueous phase liquid, BTEX, particle swarm optimization
Procedia PDF Downloads 44531387 A Systematic Review of Process Research in Software Engineering
Authors: Tulasi Rayasa, Phani Kumar Pullela
Abstract:
A systematic review is a research method that involves collecting and evaluating the information on a specific topic in order to provide a comprehensive and unbiased review. This type of review aims to improve the software development process by ensuring that the research is thorough and accurate. To ensure objectivity, it is important to follow systematic guidelines and consider multiple sources, such as literature reviews, interviews, and surveys. The evaluation process should also be streamlined by incorporating research from journals and other sources, such as grey literature. The main goal of a systematic review is to identify the consistency of current models in the field of computer application and software engineering.Keywords: computer application, software engineering, process research, data science
Procedia PDF Downloads 9931386 A Comparative Study on the Impact of Global Warming of Applying Low Carbon Factor Concrete Products
Authors: Su-Hyun Cho, Chang-U Chae
Abstract:
Environmental impact assessment techniques have been developed as a result of the worldwide efforts to reduce the environmental impact of global warming. By using the quantification method in the construction industry, it is now possible to manage the greenhouse gas is to systematically evaluate the impact on the environment over the entire construction process. In particular, the proportion of greenhouse gas emissions at the production stage of construction material occupied is high, and efforts are needed in particular in the construction field. In this study, intended for concrete products for the construction materials, by using the LCA evaluation method, we compared the results of environmental impact assessment and carbon emissions of developing products that have been applied low-carbon technologies compared to existing products. As a results, by introducing a raw material of industrial waste, showed carbon reduction. Through a comparison of the carbon emission reduction effect of low-carbon technologies, it is intended to provide academic data for the evaluation of greenhouse gases in the construction sector and the development of low-carbon technologies of the future.Keywords: CO₂ emissions, CO₂ reduction, ready-mixed concrete, environmental impact assessment
Procedia PDF Downloads 47931385 Seismic Safety Evaluation of Weir Structures Using the Finite and Infinite Element Method
Authors: Ho Young Son, Bu Seog Ju, Woo Young Jung
Abstract:
This study presents the seismic safety evaluation of weir structure subjected to strong earthquake ground motions, as a flood defense structure in civil engineering structures. The seismic safety analysis procedure was illustrated through development of Finite Element (FE) and InFinite Element (IFE) method in ABAQUS platform. The IFE model was generated by CINPS4, 4-node linear one-way infinite model as a sold continuum infinite element in foundation areas of the weir structure and then nonlinear FE model using friction model for soil-structure interactions was applied in this study. In order to understand the complex behavior of weir structures, nonlinear time history analysis was carried out. Consequently, it was interesting to note that the compressive stress gave more vulnerability to the weir structure, in comparison to the tensile stress, during an earthquake. The stress concentration of the weir structure was shown at the connection area between the weir body and stilling basin area. The stress both tension and compression was reduced in IFE model rather than FE model of weir structures.Keywords: seismic, numerical analysis, FEM, weir, boundary condition
Procedia PDF Downloads 45231384 Determination of Complexity Level in Merged Irregular Transposition Cipher
Authors: Okike Benjamin, Garba Ejd
Abstract:
Today, it has been observed security of information along the superhighway is often compromised by those who are not authorized to have access to such information. In order to ensure the security of information along the superhighway, such information should be encrypted by some means to conceal the real meaning of the information. There are many encryption techniques out there in the market. However, some of these encryption techniques are often easily decrypted by adversaries. The researcher has decided to develop an encryption technique that may be more difficult to decrypt. This may be achieved by splitting the message to be encrypted into parts and encrypting each part separately and swapping the positions before transmitting the message along the superhighway. The method is termed Merged Irregular Transposition Cipher. Also, the research would determine the complexity level in respect to the number of splits of the message.Keywords: transposition cipher, merged irregular cipher, encryption, complexity level
Procedia PDF Downloads 34431383 Automatic Segmentation of the Clean Speech Signal
Authors: M. A. Ben Messaoud, A. Bouzid, N. Ellouze
Abstract:
Speech Segmentation is the measure of the change point detection for partitioning an input speech signal into regions each of which accords to only one speaker. In this paper, we apply two features based on multi-scale product (MP) of the clean speech, namely the spectral centroid of MP, and the zero crossings rate of MP. We focus on multi-scale product analysis as an important tool for segmentation extraction. The multi-scale product is based on making the product of the speech wavelet transform coefficients at three successive dyadic scales. We have evaluated our method on the Keele database. Experimental results show the effectiveness of our method presenting a good performance. It shows that the two simple features can find word boundaries, and extracted the segments of the clean speech.Keywords: multiscale product, spectral centroid, speech segmentation, zero crossings rate
Procedia PDF Downloads 49931382 Power Quality Evaluation of Electrical Distribution Networks
Authors: Mohamed Idris S. Abozaed, Suliman Mohamed Elrajoubi
Abstract:
Researches and concerns in power quality gained significant momentum in the field of power electronics systems over the last two decades globally. This sudden increase in the number of concerns over power quality problems is a result of the huge increase in the use of non-linear loads. In this paper, power quality evaluation of some distribution networks at Misurata - Libya has been done using a power quality and energy analyzer (Fluke 437 Series II). The results of this evaluation are used to minimize the problems of power quality. The analysis shows the main power quality problems that exist and the level of awareness of power quality issues with the aim of generating a start point which can be used as guidelines for researchers and end users in the field of power systems.Keywords: power quality disturbances, power quality evaluation, statistical analysis, electrical distribution networks
Procedia PDF Downloads 53331381 Early Intervention for Preschool Children of Parents with Mental Illness: The Evaluation of a Resource for Service Providers
Authors: Stella Laletas, Andrea Reupert, Melinda Goodyear, Bradley Morgan
Abstract:
Background: Many people with a mental illness have young children. Research has shown that early childhood is a particularly vulnerable time for children whose parents have a mental illness. Moreover, repeated research has demonstrated the effectiveness of a multiagency approach to family focused practice for improving parental functioning and preventing adverse outcomes in children whose parents have a mental illness, particularly in the early years of a child’s life. However, there is a paucity of professional development resources for professionals who work with families where a parent has a mental illness and has young children. Significance of the study: This study will make a contribution to addressing knowledge gaps around resource development and workforce needs for early childhood and mental health professionals working with young children where a parent has a mental illness. Objective: This presentation describes a newly developed resource, 'Pathways of Care', specifically designed for early childhood educators and mental health workers, alongside pilot evaluation data regarding its effectiveness. ‘Pathways of Care’ aims to promote collaborative practice and present early identification and referral processes for workers in this sector. The resource was developed by the Children of Parents with a Mental Illness (COPMI) National Initiative which is funded by the Australian Government. Method: Using a mixed method design, the effectiveness of the training resource is also presented. Fifteen workers completed the Family Focus Mental Health Practice Questionnaire pre and post using the resource, to measure confidence and practice change; semi-structured interviews were also conducted with eight of these same workers to further explore the utility of the resource. Findings: The findings indicated the resource was effective in increasing knowledge and confidence, particularly for new and/or inexperienced staff. Examples of how the resource was used in practice by various professions emerged from the interview data. Conclusions: Collaborative practice, early identification and intervention in early childhood can potentially play a key role in altering the life trajectory of children who are at risk. This information has important implications for workforce development and staff training in both the early childhood and mental health sectors. Implications for policy and future research are discussed.Keywords: parents with mental ilnesses, early intervention, evaluation, preschool children
Procedia PDF Downloads 45231380 Similarity Based Membership of Elements to Uncertain Concept in Information System
Authors: M. Kamel El-Sayed
Abstract:
The process of determining the degree of membership for an element to an uncertain concept has been found in many ways, using equivalence and symmetry relations in information systems. In the case of similarity, these methods did not take into account the degree of symmetry between elements. In this paper, we use a new definition for finding the membership based on the degree of symmetry. We provide an example to clarify the suggested methods and compare it with previous methods. This method opens the door to more accurate decisions in information systems.Keywords: information system, uncertain concept, membership function, similarity relation, degree of similarity
Procedia PDF Downloads 22331379 Blockchain Technology Security Evaluation: Voting System Based on Blockchain
Authors: Omid Amini
Abstract:
Nowadays, technology plays the most important role in the life of human beings because people use technology to share data and to communicate with each other, but the challenge is the security of this data. For instance, as more people turn to technology in the world, more data is generated, and more hackers try to steal or infiltrate data. In addition, the data is under the control of the central authority, which can trigger the challenge of losing information and changing information; this can create widespread anxiety for different people in different communities. In this paper, we sought to investigate Blockchain technology that can guarantee information security and eliminate the challenge of central authority access to information. Now a day, people are suffering from the current voting system. This means that the lack of transparency in the voting system is a big problem for society and the government in most countries, but blockchain technology can be the best alternative to the previous voting system methods because it removes the most important challenge for voting. According to the results, this research can be a good start to getting acquainted with this new technology, especially on the security part and familiarity with how to use a voting system based on blockchain in the world. At the end of this research, it is concluded that the use of blockchain technology can solve the major security problem and lead to a secure and transparent election.Keywords: blockchain, technology, security, information, voting system, transparency
Procedia PDF Downloads 13231378 The Extraction and Stripping of Hg(II) from Produced Water via Hollow Fiber Contactor
Authors: Dolapop Sribudda, Ura Pancharoen
Abstract:
The separation of Hg(II) from produced water by hollow fiber contactors (HFC) was investigation. This system included of two hollow fiber modules in the series connecting. The first module used for the extraction reaction and the second module for stripping reaction. Aliquat336 extractant was fed from the organic reservoirs into the shell side of the first hollow fiber module and continuous to the shell side of the second module. The organic liquid was continuously feed recirculate and back to the reservoirs. The feed solution was pumped into the lumen (tube side) of the first hollow fiber module. Simultaneously, the stripping solution was pumped in the same way in tube side of the second module. The feed and stripping solution was fed which had a counter current flow. Samples were kept in the outlet of feed and stripping solution for 1 hour and characterized concentration of Hg(II) by Inductively Couple Plasma Atomic Emission Spectroscopy (ICP-AES). Feed solution was produced water from natural gulf of Thailand. The extractant was Aliquat336 dissolved in kerosene diluent. Stripping solution used was nitric acid (HNO3) and thiourea (NH2CSNH2). The effect of carrier concentration and type of stripping solution were investigated. Results showed that the best condition were 10 % (v/v) Aliquat336 and 1.0 M NH2CSNH2. At the optimum condition, the extraction and stripping of Hg(II) were 98% and 44.2%, respectively.Keywords: Hg(II), hollow fiber contactor, produced water, wastewater treatment
Procedia PDF Downloads 40331377 An Improved Image Steganography Technique Based on Least Significant Bit Insertion
Authors: Olaiya Folorunsho, Comfort Y. Daramola, Joel N. Ugwu, Lawrence B. Adewole, Olufisayo S. Ekundayo
Abstract:
In today world, there is a tremendous rise in the usage of internet due to the fact that almost all the communication and information sharing is done over the web. Conversely, there is a continuous growth of unauthorized access to confidential data. This has posed a challenge to information security expertise whose major goal is to curtail the menace. One of the approaches to secure the safety delivery of data/information to the rightful destination without any modification is steganography. Steganography is the art of hiding information inside an embedded information. This research paper aimed at designing a secured algorithm with the use of image steganographic technique that makes use of Least Significant Bit (LSB) algorithm for embedding the data into the bit map image (bmp) in order to enhance security and reliability. In the LSB approach, the basic idea is to replace the LSB of the pixels of the cover image with the Bits of the messages to be hidden without destroying the property of the cover image significantly. The system was implemented using C# programming language of Microsoft.NET framework. The performance evaluation of the proposed system was experimented by conducting a benchmarking test for analyzing the parameters like Mean Squared Error (MSE) and Peak Signal to Noise Ratio (PSNR). The result showed that image steganography performed considerably in securing data hiding and information transmission over the networks.Keywords: steganography, image steganography, least significant bits, bit map image
Procedia PDF Downloads 26631376 Status of Communication and Swallowing Therapy in Patient with a Tracheostomy
Authors: Ya-Hui Wang
Abstract:
Lower speech therapy rate of tracheostomized patient was noted in comparison with previous researches. This study is aim to shed light on the referral status of speech therapy in those patients in Taiwan. This study developed an analysis for the size and key characteristics of the population of tracheostomized in-patient in the Taiwan. Method: We analyzed National Healthcare Insurance data (The Collaboration Center of Health Information Application, CCHIA) from Jan 1 2010 to Dec 31 2010. Result: over ages 3, number of tracheostomized in-patient is directly proportional to age. A high service loading was observed in North region in comparison with other regions. Only 4.87% of the tracheostomized in-patients were referred for speech therapy, and 1.9% for swallow examination, 2.5% for communication evaluation.Keywords: refer, speech therapy, training, rehabilitation
Procedia PDF Downloads 44031375 The Modification of Convolutional Neural Network in Fin Whale Identification
Authors: Jiahao Cui
Abstract:
In the past centuries, due to climate change and intense whaling, the global whale population has dramatically declined. Among the various whale species, the fin whale experienced the most drastic drop in number due to its popularity in whaling. Under this background, identifying fin whale calls could be immensely beneficial to the preservation of the species. This paper uses feature extraction to process the input audio signal, then a network based on AlexNet and three networks based on the ResNet model was constructed to classify fin whale calls. A mixture of the DOSITS database and the Watkins database was used during training. The results demonstrate that a modified ResNet network has the best performance considering precision and network complexity.Keywords: convolutional neural network, ResNet, AlexNet, fin whale preservation, feature extraction
Procedia PDF Downloads 12231374 Reverse Logistics Information Management Using Ontological Approach
Authors: F. Lhafiane, A. Elbyed, M. Bouchoum
Abstract:
Reverse Logistics (RL) Process is considered as complex and dynamic network that involves many stakeholders such as: suppliers, manufactures, warehouse, retails, and costumers, this complexity is inherent in such process due to lack of perfect knowledge or conflicting information. Ontologies, on the other hand, can be considered as an approach to overcome the problem of sharing knowledge and communication among the various reverse logistics partners. In this paper, we propose a semantic representation based on hybrid architecture for building the Ontologies in an ascendant way, this method facilitates the semantic reconciliation between the heterogeneous information systems (ICT) that support reverse logistics Processes and product data.Keywords: Reverse Logistics, information management, heterogeneity, ontologies, semantic web
Procedia PDF Downloads 49231373 Age Estimation from Teeth among North Indian Population: Comparison and Reliability of Qualitative and Quantitative Methods
Authors: Jasbir Arora, Indu Talwar, Daisy Sahni, Vidya Rattan
Abstract:
Introduction: Age estimation is a crucial step to build the identity of a person, both in case of deceased and alive. In adults, age can be estimated on the basis of six regressive (Attrition, Secondary dentine, Dentine transparency, Root resorption, Cementum apposition and Periodontal Disease) changes in teeth qualitatively using scoring system and quantitatively by micrometric method. The present research was designed to establish the reliability of qualitative (method 1) and quantitative (method 2) of age estimation among North Indians and to compare the efficacy of these two methods. Method: 250 single-rooted extracted teeth (18-75 yrs.) were collected from Department of Oral Health Sciences, PGIMER, Chandigarh. Before extraction, periodontal score of each tooth was noted. Labiolingual sections were prepared and examined under light microscope for regressive changes. Each parameter was scored using Gustafson’s 0-3 point score system (qualitative), and total score was calculated. For quantitative method, each regressive change was measured quantitatively in form of 18 micrometric parameters under microscope with the help of measuring eyepiece. Age was estimated using linear and multiple regression analysis in Gustafson’s method and Kedici’s method respectively. Estimated age was compared with actual age on the basis of absolute mean error. Results: In pooled data, by Gustafson’s method, significant correlation (r= 0.8) was observed between total score and actual age. Total score generated an absolute mean error of ±7.8 years. Whereas, for Kedici’s method, a value of correlation coefficient of r=0.5 (p<0.01) was observed between all the eighteen micrometric parameters and known age. Using multiple regression equation, age was estimated, and an absolute mean error of age was found to be ±12.18 years. Conclusion: Gustafson’s (qualitative) method was found to be a better predictor for age estimation among North Indians.Keywords: forensic odontology, age estimation, North India, teeth
Procedia PDF Downloads 24231372 Teaching the Tacit Nuances of Japanese Onomatopoeia through an E-Learning System: An Evaluation Approach of Narrative Interpretation
Authors: Xiao-Yan Li, Takashi Hashimoto, Guanhong Li, Shuo Yang
Abstract:
In Japanese, onomatopoeia is an important element in the lively expression of feelings and experiences. It is very difficult for students of Japanese to acquire onomatopoeia, especially, its nuances. In this paper, based on traditional L2 learning theories, we propose a new method to improve the efficiency of teaching the nuances – both explicit and tacit - to non-native speakers of Japanese. The method for teaching the tacit nuances of onomatopoeia consists of three elements. First is to teach the formal rules representing the explicit nuances of onomatopoeic words. Second is to have the students create new onomatopoeic words by utilizing those formal rules. The last element is to provide feedback by evaluating the onomatopoeias created. Our previous study used five-grade relative estimation. However students were confused about the five-grade system, because they could not understand the evaluation criteria only based on a figure. In this new system, then, we built an evaluation database through native speakers’ narrative interpretation. We asked Japanese native speakers to describe their awareness of the nuances of onomatopoeia in writing. Then they voted on site and defined priorities for showing to learners on the system. To verify the effectiveness of the proposed method and the learning system, we conducted a preliminary experiment involving two groups of subjects. While Group A got feedback about the appropriateness of their onomatopoeic constructions from the native speakers’ narrative interpretation, Group B got feedback just in the form of the five-grade relative estimation. A questionnaire survey administered to all of the learners clarified our learning system availability and also identified areas that should be improved. Repetitive learning of word-formation rules, creating new onomatopoeias and gaining new awareness from narrative interpretation is the total process used to teach the explicit and tacit nuances of onomatopoeia.Keywords: onomatopoeia, tacit nuance, narrative interpretation, e-learning system, second language teaching
Procedia PDF Downloads 39631371 Towards Integrating Statistical Color Features for Human Skin Detection
Authors: Mohd Zamri Osman, Mohd Aizaini Maarof, Mohd Foad Rohani
Abstract:
Human skin detection recognized as the primary step in most of the applications such as face detection, illicit image filtering, hand recognition and video surveillance. The performance of any skin detection applications greatly relies on the two components: feature extraction and classification method. Skin color is the most vital information used for skin detection purpose. However, color feature alone sometimes could not handle images with having same color distribution with skin color. A color feature of pixel-based does not eliminate the skin-like color due to the intensity of skin and skin-like color fall under the same distribution. Hence, the statistical color analysis will be exploited such mean and standard deviation as an additional feature to increase the reliability of skin detector. In this paper, we studied the effectiveness of statistical color feature for human skin detection. Furthermore, the paper analyzed the integrated color and texture using eight classifiers with three color spaces of RGB, YCbCr, and HSV. The experimental results show that the integrating statistical feature using Random Forest classifier achieved a significant performance with an F1-score 0.969.Keywords: color space, neural network, random forest, skin detection, statistical feature
Procedia PDF Downloads 46231370 Six Tropical Medicinal Plants Effects in the Treatment of Prostate Diseases in Forty Different Patients
Authors: T. Nalowa, L. Foncha, S. Eposi
Abstract:
Prostate enlargement, prostate cancer are major global health problems affecting many men as they advance in age. It is highly recommended to encourage older men to get Prostate Specific Antigen test screening frequently. Conventional treatments like radiation, chemotherapy are associated with many side effects. And this situation is a call for concern. Traditional medicine is affordable, easily prepared with little or no side effects and it contains many phytochemicals. The study aims to find the cure for prostate cancer and prostate enlargement by extracting products from plant tissues of specific herbs to determine anti-inflammatory, anti-cancer, and anti-hematuria properties. Descriptive statistical analysis was applied to describe the data process. The commonly used method of preparation was extraction. Overall, 40 patients were classified based on their medical conditions on their underlying user report. Rural communities in Fako are rich sources of plants with medicinal properties. The used plants consequently provide basic information and aid to investigate the cure of prostate cancer and prostate enlargement, with great significance.Keywords: cancer, enlargement, metastases, prostate
Procedia PDF Downloads 7531369 Reduction of False Positives in Head-Shoulder Detection Based on Multi-Part Color Segmentation
Authors: Lae-Jeong Park
Abstract:
The paper presents a method that utilizes figure-ground color segmentation to extract effective global feature in terms of false positive reduction in the head-shoulder detection. Conventional detectors that rely on local features such as HOG due to real-time operation suffer from false positives. Color cue in an input image provides salient information on a global characteristic which is necessary to alleviate the false positives of the local feature based detectors. An effective approach that uses figure-ground color segmentation has been presented in an effort to reduce the false positives in object detection. In this paper, an extended version of the approach is presented that adopts separate multipart foregrounds instead of a single prior foreground and performs the figure-ground color segmentation with each of the foregrounds. The multipart foregrounds include the parts of the head-shoulder shape and additional auxiliary foregrounds being optimized by a search algorithm. A classifier is constructed with the feature that consists of a set of the multiple resulting segmentations. Experimental results show that the presented method can discriminate more false positive than the single prior shape-based classifier as well as detectors with the local features. The improvement is possible because the presented approach can reduce the false positives that have the same colors in the head and shoulder foregrounds.Keywords: pedestrian detection, color segmentation, false positive, feature extraction
Procedia PDF Downloads 28131368 Endocardial Ultrasound Segmentation using Level Set method
Authors: Daoudi Abdelaziz, Mahmoudi Saïd, Chikh Mohamed Amine
Abstract:
This paper presents a fully automatic segmentation method of the left ventricle at End Systolic (ES) and End Diastolic (ED) in the ultrasound images by means of an implicit deformable model (level set) based on Geodesic Active Contour model. A pre-processing Gaussian smoothing stage is applied to the image, which is essential for a good segmentation. Before the segmentation phase, we locate automatically the area of the left ventricle by using a detection approach based on the Hough Transform method. Consequently, the result obtained is used to automate the initialization of the level set model. This initial curve (zero level set) deforms to search the Endocardial border in the image. On the other hand, quantitative evaluation was performed on a data set composed of 15 subjects with a comparison to ground truth (manual segmentation).Keywords: level set method, transform Hough, Gaussian smoothing, left ventricle, ultrasound images.
Procedia PDF Downloads 46531367 A Fast and Robust Protocol for Reconstruction and Re-Enactment of Historical Sites
Authors: Sanaa I. Abu Alasal, Madleen M. Esbeih, Eman R. Fayyad, Rami S. Gharaibeh, Mostafa Z. Ali, Ahmed A. Freewan, Monther M. Jamhawi
Abstract:
This research proposes a novel reconstruction protocol for restoring missing surfaces and low-quality edges and shapes in photos of artifacts at historical sites. The protocol starts with the extraction of a cloud of points. This extraction process is based on four subordinate algorithms, which differ in the robustness and amount of resultant. Moreover, they use different -but complementary- accuracy to some related features and to the way they build a quality mesh. The performance of our proposed protocol is compared with other state-of-the-art algorithms and toolkits. The statistical analysis shows that our algorithm significantly outperforms its rivals in the resultant quality of its object files used to reconstruct the desired model.Keywords: meshes, point clouds, surface reconstruction protocols, 3D reconstruction
Procedia PDF Downloads 45731366 Electrical Decomposition of Time Series of Power Consumption
Authors: Noura Al Akkari, Aurélie Foucquier, Sylvain Lespinats
Abstract:
Load monitoring is a management process for energy consumption towards energy savings and energy efficiency. Non Intrusive Load Monitoring (NILM) is one method of load monitoring used for disaggregation purposes. NILM is a technique for identifying individual appliances based on the analysis of the whole residence data retrieved from the main power meter of the house. Our NILM framework starts with data acquisition, followed by data preprocessing, then event detection, feature extraction, then general appliance modeling and identification at the final stage. The event detection stage is a core component of NILM process since event detection techniques lead to the extraction of appliance features. Appliance features are required for the accurate identification of the household devices. In this research work, we aim at developing a new event detection methodology with accurate load disaggregation to extract appliance features. Time-domain features extracted are used for tuning general appliance models for appliance identification and classification steps. We use unsupervised algorithms such as Dynamic Time Warping (DTW). The proposed method relies on detecting areas of operation of each residential appliance based on the power demand. Then, detecting the time at which each selected appliance changes its states. In order to fit with practical existing smart meters capabilities, we work on low sampling data with a frequency of (1/60) Hz. The data is simulated on Load Profile Generator software (LPG), which was not previously taken into consideration for NILM purposes in the literature. LPG is a numerical software that uses behaviour simulation of people inside the house to generate residential energy consumption data. The proposed event detection method targets low consumption loads that are difficult to detect. Also, it facilitates the extraction of specific features used for general appliance modeling. In addition to this, the identification process includes unsupervised techniques such as DTW. To our best knowledge, there exist few unsupervised techniques employed with low sampling data in comparison to the many supervised techniques used for such cases. We extract a power interval at which falls the operation of the selected appliance along with a time vector for the values delimiting the state transitions of the appliance. After this, appliance signatures are formed from extracted power, geometrical and statistical features. Afterwards, those formed signatures are used to tune general model types for appliances identification using unsupervised algorithms. This method is evaluated using both simulated data on LPG and real-time Reference Energy Disaggregation Dataset (REDD). For that, we compute performance metrics using confusion matrix based metrics, considering accuracy, precision, recall and error-rate. The performance analysis of our methodology is then compared with other detection techniques previously used in the literature review, such as detection techniques based on statistical variations and abrupt changes (Variance Sliding Window and Cumulative Sum).Keywords: electrical disaggregation, DTW, general appliance modeling, event detection
Procedia PDF Downloads 7831365 Analytical Tools for Multi-Residue Analysis of Some Oxygenated Metabolites of PAHs (Hydroxylated, Quinones) in Sediments
Authors: I. Berger, N. Machour, F. Portet-Koltalo
Abstract:
Polycyclic aromatic hydrocarbons (PAHs) are toxic and carcinogenic pollutants produced in majority by incomplete combustion processes in industrialized and urbanized areas. After being emitted in atmosphere, these persistent contaminants are deposited to soils or sediments. Even if persistent, some can be partially degraded (photodegradation, biodegradation, chemical oxidation) and they lead to oxygenated metabolites (oxy-PAHs) which can be more toxic than their parent PAH. Oxy-PAHs are less measured than PAHs in sediments and this study aims to compare different analytical tools in order to extract and quantify a mixture of four hydroxylated PAHs (OH-PAHs) and four carbonyl PAHs (quinones) in sediments. Methodologies: Two analytical systems – HPLC with on-line UV and fluorescence detectors (HPLC-UV-FLD) and GC coupled to a mass spectrometer (GC-MS) – were compared to separate and quantify oxy-PAHs. Microwave assisted extraction (MAE) was optimized to extract oxy-PAHs from sediments. Results: First OH-PAHs and quinones were analyzed in HPLC with on-line UV and fluorimetric detectors. OH-PAHs were detected with the sensitive FLD, but the non-fluorescent quinones were detected with UV. The limits of detection (LOD)s obtained were in the range (2-3)×10-4 mg/L for OH-PAHs and (2-3)×10-3 mg/L for quinones. Second, even if GC-MS is not well adapted to the analysis of the thermodegradable OH-PAHs and quinones without any derivatization step, it was used because of the advantages of the detector in terms of identification and of GC in terms of efficiency. Without derivatization, only two of the four quinones were detected in the range 1-10 mg/L (LODs=0.3-1.2 mg/L) and LODs were neither very satisfying for the four OH-PAHs (0.18-0.6 mg/L). So two derivatization processes were optimized, comparing to literature: one for silylation of OH-PAHs, one for acetylation of quinones. Silylation using BSTFA/TCMS 99/1 was enhanced using a mixture of catalyst solvents (pyridine/ethyle acetate) and finding the appropriate reaction duration (5-60 minutes). Acetylation was optimized at different steps of the process, including the initial volume of compounds to derivatize, the added amounts of Zn (0.1-0.25 g), the nature of the derivatization product (acetic anhydride, heptafluorobutyric acid…) and the liquid/liquid extraction at the end of the process. After derivatization, LODs were decreased by a factor 3 for OH-PAHs and by a factor 4 for quinones, all the quinones being now detected. Thereafter, quinones and OH-PAHs were extracted from spiked sediments using microwave assisted extraction (MAE) followed by GC-MS analysis. Several mixtures of solvents of different volumes (10-25 mL) and using different extraction temperatures (80-120°C) were tested to obtain the best recovery yields. Satisfactory recoveries could be obtained for quinones (70-96%) and for OH-PAHs (70-104%). Temperature was a critical factor which had to be controlled to avoid oxy-PAHs degradation during the MAE extraction process. Conclusion: Even if MAE-GC-MS was satisfactory to analyze these oxy-PAHs, MAE optimization has to be carried on to obtain a most appropriate extraction solvent mixture, allowing a direct injection in the HPLC-UV-FLD system, which is more sensitive than GC-MS and does not necessitate a previous long derivatization step.Keywords: derivatizations for GC-MS, microwave assisted extraction, on-line HPLC-UV-FLD, oxygenated PAHs, polluted sediments
Procedia PDF Downloads 28731364 Lead in The Soil-Plant System Following Aged Contamination from Ceramic Wastes
Authors: F. Pedron, M. Grifoni, G. Petruzzelli, M. Barbafieri, I. Rosellini, B. Pezzarossa
Abstract:
Lead contamination of agricultural land mainly vegetated with perennial ryegrass (Lolium perenne) has been investigated. The metal derived from the discharge of sludge from a ceramic industry in the past had used lead paints. The results showed very high values of lead concentration in many soil samples. In order to assess the lead soil contamination, a sequential extraction with H2O, KNO3, EDTA was performed, and the chemical forms of lead in the soil were evaluated. More than 70% of lead was in a potentially bioavailable form. Analysis of Lolium perenne showed elevated lead concentration. A Freundlich-like model was used to describe the transferability of the metal from the soil to the plant.Keywords: bioavailability, Freundlich-like equation, sequential extraction, soil lead contamination
Procedia PDF Downloads 31031363 Mobile Augmented Reality for Collaboration in Operation
Authors: Chong-Yang Qiao
Abstract:
Mobile augmented reality (MAR) tracking targets from the surroundings and aids operators for interactive data and procedures visualization, potential equipment and system understandably. Operators remotely communicate and coordinate with each other for the continuous tasks, information and data exchange between control room and work-site. In the routine work, distributed control system (DCS) monitoring and work-site manipulation require operators interact in real-time manners. The critical question is the improvement of user experience in cooperative works through applying Augmented Reality in the traditional industrial field. The purpose of this exploratory study is to find the cognitive model for the multiple task performance by MAR. In particular, the focus will be on the comparison between different tasks and environment factors which influence information processing. Three experiments use interface and interaction design, the content of start-up, maintenance and stop embedded in the mobile application. With the evaluation criteria of time demands and human errors, and analysis of the mental process and the behavior action during the multiple tasks, heuristic evaluation was used to find the operators performance with different situation factors, and record the information processing in recognition, interpretation, judgment and reasoning. The research will find the functional properties of MAR and constrain the development of the cognitive model. Conclusions can be drawn that suggest MAR is easy to use and useful for operators in the remote collaborative works.Keywords: mobile augmented reality, remote collaboration, user experience, cognition model
Procedia PDF Downloads 19731362 Evaluation Synthesis of Private Sector Engagement in International Development
Authors: Valerie Habbel, Magdalena Orth, Johanna Richter, Steffen Schimko
Abstract:
Cooperation between development actors and the private sector is becoming increasingly important, as it is expected to mobilize additional resources to achieve the Sustainable Development Goals (SDGs), among other things. However, whether the goals of cooperation are achieved has so far only been explored in evaluations and studies of individual projects and instruments. The evaluation synthesis attempts to close this gap by systematically analyzing existing evidence (evaluations and academic studies) from national and international development cooperation on private sector engagement. Overall, the evaluations and studies considered report mainly positive effects on investors and donors, intermediaries, partner countries, and target groups. However, various analyses, including on the quality of the evaluations, point to a positive bias in the results. The evaluation synthesis makes recommendations on the definition of indicators, the measurement and evaluation of impacts and additionality, knowledge management, and the consideration of transaction costs in cooperation with private actors.Keywords: evaluation synthesis, private sector engagement, international development, sustainable development
Procedia PDF Downloads 21031361 The Relationship between Job Stress and Handover Effectiveness of Nurses
Authors: Rujnan Tuna, Ayse Cil Akinci
Abstract:
Work life takes up an important place in human life, and an employed person faces many stimuli from internal and external environments and is affected by them in a positive or negative way. Also, the handover process, which is the process of sharing information about the patient with other health professionals, is an important criterion to maintain patient care and enhance the quality of care provided. Handover is a key component for sustaining daily basic clinical practices and is also essential to maintain the safe patient care. This investigation followed a descriptive and correlation design in order to establish job stress and the handover efficiency of nurses and the relationship in between. The study was conducted with 192 nurses working in a public hospital in Istanbul between January and March 2017. Descriptive information form, Job Stressors Scale, and Handover Evaluation Scale were used to collect the data of the study. The data were analyzed by using IBM SPSS Statistics 22.0 statistical software. Approvals from participants, managers of institution, and ethics committee were taken for the study. As a result of the research, it was found that job stress was above the median value, and the highest score in the ‘work role conflict’ subdimension. Also, it was found that the effectiveness of the nurses' handover effectiviness was above the median value and the highest score in the ‘quality of information’ subdimension. In the study, there was a negatively weak correlation between ‘work role overload’ subdimension of Job Stressors Scale and ‘interaction and support’ subdimension of Handover Evaluation Scale. There is a need for further study in order to maintain patient safety.Keywords: handover, job stress, nurse, patient
Procedia PDF Downloads 168