Search results for: entity extraction
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 2269

Search results for: entity extraction

1549 Association between Neurofibromatosis Type 1 and Breast Sarcoma: A Case Report

Authors: Ines Zemni, Maher Slimane, Jamel Ben Hassouna, Khaled Rahal

Abstract:

Background: Neurofibromatosis type 1 (NF1) is a genetic disease, which is associated with an increased risk of developing different malignancies including breast cancer. The association between NF1 band breast sarcoma is a rare entity. Herein we present a 25-year-old woman with NF1 who had fibrosarcoma of the left breast. Case presentation: The patient has multiple thoraco-abdominal 'café au lait' spots. Clinical examination showed a lump of the left breast measuring 9 cm of diameter, which was noticed for 6 months. There was a left inguinal mass of 6 cm of diameter. The patient underwent first a left lumpectomy. Histopathological exam revealed a high-grade fibrosarcoma of the left breast measuring 7.5 cm. Three months later, the patient underwent a left mastectomy and excision of the inguinal mass, which was a neurofibroma. An adjuvant chemotherapy and radiation therapy were indicated, but not applied because of the timeout. The patient is now alive after a follow up of 6 years, with no loco-regional recurrence or metastasis. Conclusion: The relationship between NF1 and breast cancer need to be more clarified by further studies. Establishing a specific screening program of these patients may help to make an earlier diagnosis of breast cancer.

Keywords: neurofibromatosis, breast, sarcoma, cancer

Procedia PDF Downloads 121
1548 An Analysis of Eco-efficiency and GHG Emission of Olive Oil Production in Northeast of Portugal

Authors: M. Feliciano, F. Maia, A. Gonçalves

Abstract:

Olive oil production sector plays an important role in Portuguese economy. It had a major growth over the last decade, increasing its weight in the overall national exports. International market penetration for Mediterranean traditional products is increasingly more demanding, especially in the Northern European markets, where consumers are looking for more sustainable products. Trying to support this growing demand this study addresses olive oil production under the environmental and eco-efficiency perspectives. The analysis considers two consecutive product life cycle stages: olive trees farming; and olive oil extraction in mills. Addressing olive farming, data collection covered two different organizations: a middle-size farm (~12ha) (F1) and a large-size farm (~100ha) (F2). Results from both farms show that olive collection activities are responsible for the largest amounts of Green House Gases (GHG) emissions. In this activities, estimate for the Carbon Footprint per olive was higher in F2 (188g CO2e/kgolive) than in F1 (148g CO2e/kgolive). Considering olive oil extraction, two different mills were considered: one using a two-phase system (2P) and other with a three-phase system (3P). Results from the study of two mills show that there is a much higher use of water in 3P. Energy intensity (EI) is similar in both mills. When evaluating the GHG generated, two conditions are evaluated: a biomass neutral condition resulting on a carbon footprint higher in 3P (184g CO2e/Lolive oil) than in 2P (92g CO2e/Lolive oil); and a non-neutral biomass condition in which 2P increase its carbon footprint to 273g CO2e/Lolive oil. When addressing the carbon footprint of possible combinations among studied subsystems, results suggest that olive harvesting is the major source for GHG.

Keywords: carbon footprint, environmental indicators, farming subsystem, industrial subsystem, olive oil

Procedia PDF Downloads 287
1547 Secure Text Steganography for Microsoft Word Document

Authors: Khan Farhan Rafat, M. Junaid Hussain

Abstract:

Seamless modification of an entity for the purpose of hiding a message of significance inside its substance in a manner that the embedding remains oblivious to an observer is known as steganography. Together with today's pervasive registering frameworks, steganography has developed into a science that offers an assortment of strategies for stealth correspondence over the globe that must, however, need a critical appraisal from security breach standpoint. Microsoft Word is amongst the preferably used word processing software, which comes as a part of the Microsoft Office suite. With a user-friendly graphical interface, the richness of text editing, and formatting topographies, the documents produced through this software are also most suitable for stealth communication. This research aimed not only to epitomize the fundamental concepts of steganography but also to expound on the utilization of Microsoft Word document as a carrier for furtive message exchange. The exertion is to examine contemporary message hiding schemes from security aspect so as to present the explorative discoveries and suggest enhancements which may serve a wellspring of information to encourage such futuristic research endeavors.

Keywords: hiding information in plain sight, stealth communication, oblivious information exchange, conceal, steganography

Procedia PDF Downloads 241
1546 Modal Density Influence on Modal Complexity Quantification in Dynamic Systems

Authors: Fabrizio Iezzi, Claudio Valente

Abstract:

The viscous damping in dynamic systems can be proportional or non-proportional. In the first case, the mode shapes are real whereas in the second case they are complex. From an engineering point of view, the complexity of the mode shapes is important in order to quantify the non-proportional damping. Different indices exist to provide estimates of the modal complexity. These indices are or not zero, depending whether the mode shapes are not or are complex. The modal density problem arises in the experimental identification when the dynamic systems have close modal frequencies. Depending on the entity of this closeness, the mode shapes can hold fictitious imaginary quantities that affect the values of the modal complexity indices. The results are the failing in the identification of the real or complex mode shapes and then of the proportional or non-proportional damping. The paper aims to show the influence of the modal density on the values of these indices in case of both proportional and non-proportional damping. Theoretical and pseudo-experimental solutions are compared to analyze the problem according to an appropriate mechanical system.

Keywords: complex mode shapes, dynamic systems identification, modal density, non-proportional damping

Procedia PDF Downloads 387
1545 The OLOS® Way to Cultural Heritage: User Interface with Anthropomorphic Characteristics

Authors: Daniele Baldacci, Remo Pareschi

Abstract:

Augmented Reality and Augmented Intelligence are radically changing information technology. The path that starts from the keyboard and then, passing through milestones such as Siri, Alexa and other vocal avatars, reaches a more fluid and natural communication with computers, thus converting the dichotomy between man and machine into a harmonious interaction, now heads unequivocally towards a new IT paradigm, where holographic computing will play a key role. The OLOS® platform contributes substantially to this trend in that it infuses computers with human features, by transferring the gestures and expressions of persons of flesh and bones to anthropomorphic holographic interfaces which in turn will use them to interact with real-life humans. In fact, we could say, boldly but with a solid technological background to back the statement, that OLOS® gives reality to an altogether new entity, placed at the exact boundary between nature and technology, namely the holographic human being. Holographic humans qualify as the perfect carriers for the virtual reincarnation of characters handed down from history and tradition. Thus, they provide for an innovative and highly immersive way of experiencing our cultural heritage as something alive and pulsating in the present.

Keywords: digital cinematography, human-computer interfaces, holographic simulation, interactive museum exhibits

Procedia PDF Downloads 115
1544 Power Quality Modeling Using Recognition Learning Methods for Waveform Disturbances

Authors: Sang-Keun Moon, Hong-Rok Lim, Jin-O Kim

Abstract:

This paper presents a Power Quality (PQ) modeling and filtering processes for the distribution system disturbances using recognition learning methods. Typical PQ waveforms with mathematical applications and gathered field data are applied to the proposed models. The objective of this paper is analyzing PQ data with respect to monitoring, discriminating, and evaluating the waveform of power disturbances to ensure the system preventative system failure protections and complex system problem estimations. Examined signal filtering techniques are used for the field waveform noises and feature extractions. Using extraction and learning classification techniques, the efficiency was verified for the recognition of the PQ disturbances with focusing on interactive modeling methods in this paper. The waveform of selected 8 disturbances is modeled with randomized parameters of IEEE 1159 PQ ranges. The range, parameters, and weights are updated regarding field waveform obtained. Along with voltages, currents have same process to obtain the waveform features as the voltage apart from some of ratings and filters. Changing loads are causing the distortion in the voltage waveform due to the drawing of the different patterns of current variation. In the conclusion, PQ disturbances in the voltage and current waveforms indicate different types of patterns of variations and disturbance, and a modified technique based on the symmetrical components in time domain was proposed in this paper for the PQ disturbances detection and then classification. Our method is based on the fact that obtained waveforms from suggested trigger conditions contain potential information for abnormality detections. The extracted features are sequentially applied to estimation and recognition learning modules for further studies.

Keywords: power quality recognition, PQ modeling, waveform feature extraction, disturbance trigger condition, PQ signal filtering

Procedia PDF Downloads 186
1543 Digital Twin for a Floating Solar Energy System with Experimental Data Mining and AI Modelling

Authors: Danlei Yang, Luofeng Huang

Abstract:

The integration of digital twin technology with renewable energy systems offers an innovative approach to predicting and optimising performance throughout the entire lifecycle. A digital twin is a continuously updated virtual replica of a real-world entity, synchronised with data from its physical counterpart and environment. Many digital twin companies today claim to have mature digital twin products, but their focus is primarily on equipment visualisation. However, the core of a digital twin should be its model, which can mirror, shadow, and thread with the real-world entity, which is still underdeveloped. For a floating solar energy system, a digital twin model can be defined in three aspects: (a) the physical floating solar energy system along with environmental factors such as solar irradiance and wave dynamics, (b) a digital model powered by artificial intelligence (AI) algorithms, and (c) the integration of real system data with the AI-driven model and a user interface. The experimental setup for the floating solar energy system, is designed to replicate real-ocean conditions of floating solar installations within a controlled laboratory environment. The system consists of a water tank that simulates an aquatic surface, where a floating catamaran structure supports a solar panel. The solar simulator is set up in three positions: one directly above and two inclined at a 45° angle in front and behind the solar panel. This arrangement allows the simulation of different sun angles, such as sunrise, midday, and sunset. The solar simulator is positioned 400 mm away from the solar panel to maintain consistent solar irradiance on its surface. Stability for the floating structure is achieved through ropes attached to anchors at the bottom of the tank, which simulates the mooring systems used in real-world floating solar applications. The floating solar energy system's sensor setup includes various devices to monitor environmental and operational parameters. An irradiance sensor measures solar irradiance on the photovoltaic (PV) panel. Temperature sensors monitor ambient air and water temperatures, as well as the PV panel temperature. Wave gauges measure wave height, while load cells capture mooring force. Inclinometers and ultrasonic sensors record heave and pitch amplitudes of the floating system’s motions. An electric load measures the voltage and current output from the solar panel. All sensors collect data simultaneously. Artificial neural network (ANN) algorithms are central to developing the digital model, which processes historical and real-time data, identifies patterns, and predicts the system’s performance in real time. The data collected from various sensors are partly used to train the digital model, with the remaining data reserved for validation and testing. The digital twin model combines the experimental setup with the ANN model, enabling monitoring, analysis, and prediction of the floating solar energy system's operation. The digital model mirrors the functionality of the physical setup, running in sync with the experiment to provide real-time insights and predictions. It provides useful industrial benefits, such as informing maintenance plans as well as design and control strategies for optimal energy efficiency. In long term, this digital twin will help improve overall solar energy yield whilst minimising the operational costs and risks.

Keywords: digital twin, floating solar energy system, experiment setup, artificial intelligence

Procedia PDF Downloads 8
1542 Data Mining Spatial: Unsupervised Classification of Geographic Data

Authors: Chahrazed Zouaoui

Abstract:

In recent years, the volume of geospatial information is increasing due to the evolution of communication technologies and information, this information is presented often by geographic information systems (GIS) and stored on of spatial databases (BDS). The classical data mining revealed a weakness in knowledge extraction at these enormous amounts of data due to the particularity of these spatial entities, which are characterized by the interdependence between them (1st law of geography). This gave rise to spatial data mining. Spatial data mining is a process of analyzing geographic data, which allows the extraction of knowledge and spatial relationships from geospatial data, including methods of this process we distinguish the monothematic and thematic, geo- Clustering is one of the main tasks of spatial data mining, which is registered in the part of the monothematic method. It includes geo-spatial entities similar in the same class and it affects more dissimilar to the different classes. In other words, maximize intra-class similarity and minimize inter similarity classes. Taking account of the particularity of geo-spatial data. Two approaches to geo-clustering exist, the dynamic processing of data involves applying algorithms designed for the direct treatment of spatial data, and the approach based on the spatial data pre-processing, which consists of applying clustering algorithms classic pre-processed data (by integration of spatial relationships). This approach (based on pre-treatment) is quite complex in different cases, so the search for approximate solutions involves the use of approximation algorithms, including the algorithms we are interested in dedicated approaches (clustering methods for partitioning and methods for density) and approaching bees (biomimetic approach), our study is proposed to design very significant to this problem, using different algorithms for automatically detecting geo-spatial neighborhood in order to implement the method of geo- clustering by pre-treatment, and the application of the bees algorithm to this problem for the first time in the field of geo-spatial.

Keywords: mining, GIS, geo-clustering, neighborhood

Procedia PDF Downloads 375
1541 A Neuron Model of Facial Recognition and Detection of an Authorized Entity Using Machine Learning System

Authors: J. K. Adedeji, M. O. Oyekanmi

Abstract:

This paper has critically examined the use of Machine Learning procedures in curbing unauthorized access into valuable areas of an organization. The use of passwords, pin codes, user’s identification in recent times has been partially successful in curbing crimes involving identities, hence the need for the design of a system which incorporates biometric characteristics such as DNA and pattern recognition of variations in facial expressions. The facial model used is the OpenCV library which is based on the use of certain physiological features, the Raspberry Pi 3 module is used to compile the OpenCV library, which extracts and stores the detected faces into the datasets directory through the use of camera. The model is trained with 50 epoch run in the database and recognized by the Local Binary Pattern Histogram (LBPH) recognizer contained in the OpenCV. The training algorithm used by the neural network is back propagation coded using python algorithmic language with 200 epoch runs to identify specific resemblance in the exclusive OR (XOR) output neurons. The research however confirmed that physiological parameters are better effective measures to curb crimes relating to identities.

Keywords: biometric characters, facial recognition, neural network, OpenCV

Procedia PDF Downloads 256
1540 Solid Phase Micro-Extraction/Gas Chromatography-Mass Spectrometry Study of Volatile Compounds from Strawberry Tree and Autumn Heather Honeys

Authors: Marinos Xagoraris, Elisavet Lazarou, Eleftherios Alissandrakis, Christos S. Pappas, Petros A. Tarantilis

Abstract:

Strawberry tree (Arbutus unedo L.) and autumn heather (Erica manipuliflora Salisb.) are important beekeeping plants of Greece. Six monofloral honeys (four strawberry tree, two autumn heather) were analyzed by means of Solid Phase Micro-Extraction (SPME, 60 min, 60 oC) followed by Gas Chromatography coupled to Mass Spectrometry (GC-MS) for the purpose of assessing the botanical origin. A Divinylbenzene/Carboxen/Polydimethylsiloxane (DVB/CAR/PDMS) fiber was employed, and benzophenone was used as internal standard. The volatile compounds with higher concentration (μg/ g of honey expressed as benzophenone) from strawberry tree honey samples, were α-isophorone (2.50-8.12); 3,4,5-trimethyl-phenol (0.20-4.62); 2-hydroxy-isophorone (0.06-0.53); 4-oxoisophorone (0.38-0.46); and β-isophorone (0.02-0.43). Regarding heather honey samples, the most abundant compounds were 1-methoxy-4-propyl-benzene (1.22-1.40); p-anisaldehyde (0.97-1.28); p-anisic acid (0.35-0.58); 2-furaldehyde (0.52-0.57); and benzaldehyde (0.41-0.56). Norisoprenoids are potent floral markers for strawberry-tree honey. β-isophorone is found exclusively in the volatile fraction of this type of honey, while also α-isophorone, 4-oxoisophorone and 2-hydroxy-isophorone could be considered as additional marker compounds. The analysis of autumn heather honey revealed that phenolic compounds are the most abundant and p-anisaldehyde; 1-methoxy-4-propyl-benzene; and p-anisic acid could serve as potent marker compounds. In conclusion, marker compounds for the determination of the botanical origin for these honeys could be identified as several norisoprenoids and phenolic components were found exclusively or in higher concentrations compared to common Greek honey varieties.

Keywords: SPME/GC-MS, volatile compounds, heather honey, strawberry tree honey

Procedia PDF Downloads 200
1539 A Q-Methodology Approach for the Evaluation of Land Administration Mergers

Authors: Tsitsi Nyukurayi Muparari, Walter Timo De Vries, Jaap Zevenbergen

Abstract:

The nature of Land administration accommodates diversity in terms of both spatial data handling activities and the expertise involved, which supposedly aims to satisfy the unpredictable demands of land data and the diverse demands of the customers arising from the land. However, it is known that strategic decisions of restructuring are in most cases repelled in favour of complex structures that strive to accommodate professional diversity and diverse roles in the field of Land administration. Yet despite of this widely accepted knowledge, there is scanty theoretical knowledge concerning the psychological methodologies that can extract the deeper perceptions from the diverse spatial expertise in order to explain the invisible control arm of the polarised reception of the ideas of change. This paper evaluates Q methodology in the context of a cadastre and land registry merger (under one agency) using the Swedish cadastral system as a case study. Precisely, the aim of this paper is to evaluate the effectiveness of Q methodology towards modelling the diverse psychological perceptions of spatial professionals who are in a widely contested decision of merging the cadastre and land registry components of Land administration using the Swedish cadastral system as a case study. An empirical approach that is prescribed by Q methodology starts with the concourse development, followed by the design of statements and q sort instrument, selection of the participants, the q-sorting exercise, factor extraction by PQMethod and finally narrative development by logic of abduction. The paper uses 36 statements developed from a dominant competing value theory that stands out on its reliability and validity, purposively selects 19 participants to do the Qsorting exercise, proceeds with factor extraction from the diversity using varimax rotation and judgemental rotation provided by PQMethod and effect the narrative construction using the logic abduction. The findings from the diverse perceptions from cadastral professionals in the merger decision of land registry and cadastre components in Sweden’s mapping agency (Lantmäteriet) shows that focus is rather inclined on the perfection of the relationship between the legal expertise and technical spatial expertise. There is much emphasis on tradition, loyalty and communication attributes which concern the organisation’s internal environment rather than innovation and market attributes that reveals customer behavior and needs arising from the changing humankind-land needs. It can be concluded that Q methodology offers effective tools that pursues a psychological approach for the evaluation and gradations of the decisions of strategic change through extracting the local perceptions of spatial expertise.

Keywords: cadastre, factor extraction, land administration merger, land registry, q-methodology, rotation

Procedia PDF Downloads 194
1538 Negative RT-PCR in a Newborn Infected with Zika Virus: A Case Report

Authors: Vallejo Michael, Acuña Edgar, Roa Juan David, Peñuela Rosa, Parra Alejandra, Casallas Daniela, Rodriguez Sheyla

Abstract:

Congenital Zika Virus Syndrome is an entity composed by a variety of birth defects presented in newborns that have been exposed to the Zika Virus during pregnancy. The syndrome characteristic features are severe microcephaly, cerebral tissue abnormalities, ophthalmological abnormalities such as uveitis and chorioretinitis, arthrogryposis, clubfoot deformity and muscular tone abnormalities. The confirmatory test is the Reverse transcription polymerase chain reaction (RT-PCR) associated to the physical findings. Here we present the case of a newborn with microcephaly whose mother presented a confirmed Zika Virus infection during the third trimester of pregnancy, despite of the evident findings and the history of Zika infection the RT-PCR in amniotic and cerebrospinal fluid of the newborn was negative. RT-PCR has demonstrated a low sensibility in samples with low viral loads, reason why, we propose a clinical diagnosis in patients with clinical history of Zika Virus infection during pregnancy accompanied by evident clinical manifestations of the child.

Keywords: congenital, Zika virus, microcephaly, reverse transcriptase polymerase chain reaction

Procedia PDF Downloads 211
1537 The Role of E-Learning in Science, Technology, Engineering, and Math Education

Authors: Annette McArthur

Abstract:

The traditional model of teaching and learning, where ICT sits as a separate entity is not a model for a 21st century school. It is imperative that teaching and learning embraces technological advancements. The challenge in schools lies in shifting the mindset of teachers so they see ICT as integral to their teaching, learning and curriculum rather than a separate E-Learning curriculum stream. This research project investigates how the effective, planned, intentional integration of ICT into a STEM curriculum, can enable the shift in the teacher mindset. The project incorporated: • Developing a professional coaching relationship with key STEM teachers. • Facilitating staff professional development involving student centered project based learning pedagogy in the context of a STEM curriculum. • Facilitating staff professional development involving digital literacy. • Establishing a professional community where collaboration; sharing and reflection were part of the culture of the STEM community. • Facilitating classroom support for the effective delivery innovative STEM curriculum. • Developing STEM learning spaces where technologies were used to empower and engage learners to participate in student-centered, project-based learning.

Keywords: e-learning, ICT, project based learning, STEM

Procedia PDF Downloads 300
1536 Assessing Acute Toxicity and Endocrine Disruption Potential of Selected Packages Internal Layers Extracts

Authors: N. Szczepanska, B. Kudlak, G. Yotova, S. Tsakovski, J. Namiesnik

Abstract:

In the scientific literature related to the widely understood issue of packaging materials designed to have contact with food (food contact materials), there is much information on raw materials used for their production, as well as their physiochemical properties, types, and parameters. However, not much attention is given to the issues concerning migration of toxic substances from packaging and its actual influence on the health of the final consumer, even though health protection and food safety are the priority tasks. The goal of this study was to estimate the impact of particular foodstuff packaging type, food production, and storage conditions on the degree of leaching of potentially toxic compounds and endocrine disruptors to foodstuffs using the acute toxicity test Microtox and XenoScreen YES YAS assay. The selected foodstuff packaging materials were metal cans used for fish storage and tetrapak. Five stimulants respectful to specific kinds of food were chosen in order to assess global migration: distilled water for aqueous foods with a pH above 4.5; acetic acid at 3% in distilled water for acidic aqueous food with pH below 4.5; ethanol at 5% for any food that may contain alcohol; dimethyl sulfoxide (DMSO) and artificial saliva were used in regard to the possibility of using it as an simulation medium. For each packaging three independent variables (temperature and contact time) factorial design simulant was performed. Xenobiotics migration from epoxy resins was studied at three different temperatures (25°C, 65°C, and 121°C) and extraction time of 12h, 48h and 2 weeks. Such experimental design leads to 9 experiments for each food simulant as conditions for each experiment are obtained by combination of temperature and contact time levels. Each experiment was run in triplicate for acute toxicity and in duplicate for estrogen disruption potential determination. Multi-factor analysis of variation (MANOVA) was used to evaluate the effects of the three main factors solvent, temperature (temperature regime for cup), contact time and their interactions on the respected dependent variable (acute toxicity or estrogen disruption potential). From all stimulants studied the most toxic were can and tetrapak lining acetic acid extracts that are indication for significant migration of toxic compounds. This migration increased with increase of contact time and temperature and justified the hypothesis that food products with low pH values cause significant damage internal resin filling. Can lining extracts of all simulation medias excluding distilled water and artificial saliva proved to contain androgen agonists even at 25°C and extraction time of 12h. For tetrapak extracts significant endocrine potential for acetic acid, DMSO and saliva were detected.

Keywords: food packaging, extraction, migration, toxicity, biotest

Procedia PDF Downloads 181
1535 Synthesis and Pharmaco-Potential Evaluation of Quinoline Hybrids

Authors: Paul Awolade, Parvesh Singh

Abstract:

The global threat of pathogenic resistance to available therapeutic agents has become a menace to clinical practice, public health and man’s existence inconsequential. This has therefore led to an exigency in the development of new molecular scaffolds with profound activity profiles. In this vein, a versatile synthetic tool for accessing new molecules by incorporating two or more pharmacophores into a single entity with the unique ability to be recognized by multiple receptors hence leading to an improved bioactivity, known as molecular hybridization, has been explored with tremendous success. Accordingly, aware of the similarity in pharmacological activity spectrum of quinoline and 1,2,3-triazole pharmacophores such as; anti-Alzheimer, anticancer, anti-HIV, antimalarial and antimicrobial to mention but a few, the present study sets out to synthesize hybrids of quinoline and 1,2,3-triazole. The hybrids were accessed via click chemistry using copper catalysed azide-alkyne 1,3-dipolar cycloaddition reaction. All synthesized compounds were evaluated for their pharmaco-potential in an antimicrobial assay out of which the 3-OH derivative emerged as the most active with MIC value of 4 μg/mL against Cryptococcus neoformans; a value superior to standard Fluconazole and comparable to Amphotericin B. Structures of synthesized hybrids were elucidated using appropriate spectroscopic techniques (1H, 13C and 2D NMR, FT-IR and HRMS).

Keywords: bioisostere, click chemistry, molecular hybridization, quinoline, 1, 2, 3-triazole

Procedia PDF Downloads 130
1534 Lung Cancer Detection and Multi Level Classification Using Discrete Wavelet Transform Approach

Authors: V. Veeraprathap, G. S. Harish, G. Narendra Kumar

Abstract:

Uncontrolled growth of abnormal cells in the lung in the form of tumor can be either benign (non-cancerous) or malignant (cancerous). Patients with Lung Cancer (LC) have an average of five years life span expectancy provided diagnosis, detection and prediction, which reduces many treatment options to risk of invasive surgery increasing survival rate. Computed Tomography (CT), Positron Emission Tomography (PET), and Magnetic Resonance Imaging (MRI) for earlier detection of cancer are common. Gaussian filter along with median filter used for smoothing and noise removal, Histogram Equalization (HE) for image enhancement gives the best results without inviting further opinions. Lung cavities are extracted and the background portion other than two lung cavities is completely removed with right and left lungs segmented separately. Region properties measurements area, perimeter, diameter, centroid and eccentricity measured for the tumor segmented image, while texture is characterized by Gray-Level Co-occurrence Matrix (GLCM) functions, feature extraction provides Region of Interest (ROI) given as input to classifier. Two levels of classifications, K-Nearest Neighbor (KNN) is used for determining patient condition as normal or abnormal, while Artificial Neural Networks (ANN) is used for identifying the cancer stage is employed. Discrete Wavelet Transform (DWT) algorithm is used for the main feature extraction leading to best efficiency. The developed technology finds encouraging results for real time information and on line detection for future research.

Keywords: artificial neural networks, ANN, discrete wavelet transform, DWT, gray-level co-occurrence matrix, GLCM, k-nearest neighbor, KNN, region of interest, ROI

Procedia PDF Downloads 153
1533 Web Data Scraping Technology Using Term Frequency Inverse Document Frequency to Enhance the Big Data Quality on Sentiment Analysis

Authors: Sangita Pokhrel, Nalinda Somasiri, Rebecca Jeyavadhanam, Swathi Ganesan

Abstract:

Tourism is a booming industry with huge future potential for global wealth and employment. There are countless data generated over social media sites every day, creating numerous opportunities to bring more insights to decision-makers. The integration of Big Data Technology into the tourism industry will allow companies to conclude where their customers have been and what they like. This information can then be used by businesses, such as those in charge of managing visitor centers or hotels, etc., and the tourist can get a clear idea of places before visiting. The technical perspective of natural language is processed by analysing the sentiment features of online reviews from tourists, and we then supply an enhanced long short-term memory (LSTM) framework for sentiment feature extraction of travel reviews. We have constructed a web review database using a crawler and web scraping technique for experimental validation to evaluate the effectiveness of our methodology. The text form of sentences was first classified through Vader and Roberta model to get the polarity of the reviews. In this paper, we have conducted study methods for feature extraction, such as Count Vectorization and TFIDF Vectorization, and implemented Convolutional Neural Network (CNN) classifier algorithm for the sentiment analysis to decide the tourist’s attitude towards the destinations is positive, negative, or simply neutral based on the review text that they posted online. The results demonstrated that from the CNN algorithm, after pre-processing and cleaning the dataset, we received an accuracy of 96.12% for the positive and negative sentiment analysis.

Keywords: counter vectorization, convolutional neural network, crawler, data technology, long short-term memory, web scraping, sentiment analysis

Procedia PDF Downloads 88
1532 Case Report: Rare Case of Endometrial Stromal Sarcoma with Omental Metastasis in a 19-Year Old Girl

Authors: Mukurdipi Ray, Seema Singh

Abstract:

Extrauterine endometrial stromal sarcoma (ESS) is a rare entity and typified by delayed recurrence of primary ESS. Here, we present an unusual case of uterine ESS in a woman with a history of hysterectomy. A 19-year-old girl, underwent a hysterectomy and bilateral salpingo-oophorectomy for uterine ESS 12 months ago and now after remaining disease free for nine months ago she presented with ascites along with pelvic and peritoneal mass. Intraoperatively, the large omental mass was found, and optimal cytoreduction with total omentomy (supracolic and infracolic ) total peritonectomy and hyperthermic intraperitoneal chemotherapy (HIPEC) was offered to the patient. Final histopathology report showed the involvement of only omentum by ESS cells. Immunohistochemistry (IHC) and receptor study were done and it was positive for CD-10 and desmin and negative for CK- 7. This case highlights the rarity of extrauterine ESS in the omentum with a known history of primary uterine ESS which was treated successfully with the above-mentioned procedure. Though active and long-term surveillance is recommended to monitor for late recurrences.

Keywords: endrometrial stromal sarcoma, complete cytoreduction, hyperthermic intra peritoneal chemotherapy, total omentectomy

Procedia PDF Downloads 207
1531 Multi-source Question Answering Framework Using Transformers for Attribute Extraction

Authors: Prashanth Pillai, Purnaprajna Mangsuli

Abstract:

Oil exploration and production companies invest considerable time and efforts to extract essential well attributes (like well status, surface, and target coordinates, wellbore depths, event timelines, etc.) from unstructured data sources like technical reports, which are often non-standardized, multimodal, and highly domain-specific by nature. It is also important to consider the context when extracting attribute values from reports that contain information on multiple wells/wellbores. Moreover, semantically similar information may often be depicted in different data syntax representations across multiple pages and document sources. We propose a hierarchical multi-source fact extraction workflow based on a deep learning framework to extract essential well attributes at scale. An information retrieval module based on the transformer architecture was used to rank relevant pages in a document source utilizing the page image embeddings and semantic text embeddings. A question answering framework utilizingLayoutLM transformer was used to extract attribute-value pairs incorporating the text semantics and layout information from top relevant pages in a document. To better handle context while dealing with multi-well reports, we incorporate a dynamic query generation module to resolve ambiguities. The extracted attribute information from various pages and documents are standardized to a common representation using a parser module to facilitate information comparison and aggregation. Finally, we use a probabilistic approach to fuse information extracted from multiple sources into a coherent well record. The applicability of the proposed approach and related performance was studied on several real-life well technical reports.

Keywords: natural language processing, deep learning, transformers, information retrieval

Procedia PDF Downloads 193
1530 The Simplicity of the Future: Plain Methods of Setting up a Company under the Freedom of Enterprise

Authors: Renata Hrecska

Abstract:

This research aims to present today's corporate law reforms in the micro, small and medium-sized enterprise sector. The UN Commission on International Trade Law (UNCITRAL) currently deals with emerging issues in the sector in its Working Group I that has specifically focused on possible company law simplifications, including the creation of a fully unique company, the UNCITRAL Limited Liability Organization. However, beyond the work at the UN, the different states has also been focusing on simplification efforts and demands in the sphere of commercial law. We can observe that e.g. Slovakia, Serbia, Poland, Croatia, Hungary, Romania and France are undergoing legal reforms aimed at restructuring the sector through simplification of registration or operation. An important objective of the research is to examine where the boundary is for the legal entity to be more transparent and accountable, while the legislator wants to bring the possibility of establishing a company closer to the citizen. The research material presents the advantages and disadvantages of different initiatives with comparative legal instruments and draws conclusions on the possible future vision. The researcher herself attended some of the meetings of the relevant UNCITRAL working group as a national delegated expert, giving her a personal insight into the UNLLO discourse.

Keywords: commercial law, company formation, MSME, UNCITRAL

Procedia PDF Downloads 118
1529 Efficient Video Compression Technique Using Convolutional Neural Networks and Generative Adversarial Network

Authors: P. Karthick, K. Mahesh

Abstract:

Video has become an increasingly significant component of our digital everyday contact. With the advancement of greater contents and shows of the resolution, its significant volume poses serious obstacles to the objective of receiving, distributing, compressing, and revealing video content of high quality. In this paper, we propose the primary beginning to complete a deep video compression model that jointly upgrades all video compression components. The video compression method involves splitting the video into frames, comparing the images using convolutional neural networks (CNN) to remove duplicates, repeating the single image instead of the duplicate images by recognizing and detecting minute changes using generative adversarial network (GAN) and recorded with long short-term memory (LSTM). Instead of the complete image, the small changes generated using GAN are substituted, which helps in frame level compression. Pixel wise comparison is performed using K-nearest neighbours (KNN) over the frame, clustered with K-means, and singular value decomposition (SVD) is applied for each and every frame in the video for all three color channels [Red, Green, Blue] to decrease the dimension of the utility matrix [R, G, B] by extracting its latent factors. Video frames are packed with parameters with the aid of a codec and converted to video format, and the results are compared with the original video. Repeated experiments on several videos with different sizes, duration, frames per second (FPS), and quality results demonstrate a significant resampling rate. On average, the result produced had approximately a 10% deviation in quality and more than 50% in size when compared with the original video.

Keywords: video compression, K-means clustering, convolutional neural network, generative adversarial network, singular value decomposition, pixel visualization, stochastic gradient descent, frame per second extraction, RGB channel extraction, self-detection and deciding system

Procedia PDF Downloads 187
1528 Perception-Oriented Model Driven Development for Designing Data Acquisition Process in Wireless Sensor Networks

Authors: K. Indra Gandhi

Abstract:

Wireless Sensor Networks (WSNs) have always been characterized for application-specific sensing, relaying and collection of information for further analysis. However, software development was not considered as a separate entity in this process of data collection which has posed severe limitations on the software development for WSN. Software development for WSN is a complex process since the components involved are data-driven, network-driven and application-driven in nature. This implies that there is a tremendous need for the separation of concern from the software development perspective. A layered approach for developing data acquisition design based on Model Driven Development (MDD) has been proposed as the sensed data collection process itself varies depending upon the application taken into consideration. This work focuses on the layered view of the data acquisition process so as to ease the software point of development. A metamodel has been proposed that enables reusability and realization of the software development as an adaptable component for WSN systems. Further, observing users perception indicates that proposed model helps in improving the programmer's productivity by realizing the collaborative system involved.

Keywords: data acquisition, model-driven development, separation of concern, wireless sensor networks

Procedia PDF Downloads 434
1527 ARABEX: Automated Dotted Arabic Expiration Date Extraction using Optimized Convolutional Autoencoder and Custom Convolutional Recurrent Neural Network

Authors: Hozaifa Zaki, Ghada Soliman

Abstract:

In this paper, we introduced an approach for Automated Dotted Arabic Expiration Date Extraction using Optimized Convolutional Autoencoder (ARABEX) with bidirectional LSTM. This approach is used for translating the Arabic dot-matrix expiration dates into their corresponding filled-in dates. A custom lightweight Convolutional Recurrent Neural Network (CRNN) model is then employed to extract the expiration dates. Due to the lack of available dataset images for the Arabic dot-matrix expiration date, we generated synthetic images by creating an Arabic dot-matrix True Type Font (TTF) matrix to address this limitation. Our model was trained on a realistic synthetic dataset of 3287 images, covering the period from 2019 to 2027, represented in the format of yyyy/mm/dd. We then trained our custom CRNN model using the generated synthetic images to assess the performance of our model (ARABEX) by extracting expiration dates from the translated images. Our proposed approach achieved an accuracy of 99.4% on the test dataset of 658 images, while also achieving a Structural Similarity Index (SSIM) of 0.46 for image translation on our dataset. The ARABEX approach demonstrates its ability to be applied to various downstream learning tasks, including image translation and reconstruction. Moreover, this pipeline (ARABEX+CRNN) can be seamlessly integrated into automated sorting systems to extract expiry dates and sort products accordingly during the manufacturing stage. By eliminating the need for manual entry of expiration dates, which can be time-consuming and inefficient for merchants, our approach offers significant results in terms of efficiency and accuracy for Arabic dot-matrix expiration date recognition.

Keywords: computer vision, deep learning, image processing, character recognition

Procedia PDF Downloads 82
1526 Improving Fingerprinting-Based Localization System Using Generative AI

Authors: Getaneh Berie Tarekegn

Abstract:

A precise localization system is crucial for many artificial intelligence Internet of Things (AI-IoT) applications in the era of smart cities. Their applications include traffic monitoring, emergency alarming, environmental monitoring, location-based advertising, intelligent transportation, and smart health care. The most common method for providing continuous positioning services in outdoor environments is by using a global navigation satellite system (GNSS). Due to nonline-of-sight, multipath, and weather conditions, GNSS systems do not perform well in dense urban, urban, and suburban areas.This paper proposes a generative AI-based positioning scheme for large-scale wireless settings using fingerprinting techniques. In this article, we presented a semi-supervised deep convolutional generative adversarial network (S-DCGAN)-based radio map construction method for real-time device localization. It also employed a reliable signal fingerprint feature extraction method with t-distributed stochastic neighbor embedding (t-SNE), which extracts dominant features while eliminating noise from hybrid WLAN and long-term evolution (LTE) fingerprints. The proposed scheme reduced the workload of site surveying required to build the fingerprint database by up to 78.5% and significantly improved positioning accuracy. The results show that the average positioning error of GAILoc is less than 0.39 m, and more than 90% of the errors are less than 0.82 m. According to numerical results, SRCLoc improves positioning performance and reduces radio map construction costs significantly compared to traditional methods.

Keywords: location-aware services, feature extraction technique, generative adversarial network, long short-term memory, support vector machine

Procedia PDF Downloads 59
1525 Organization Structure of Towns and Villages System in County Area Based on Fractal Theory and Gravity Model: A Case Study of Suning, Hebei Province, China

Authors: Liuhui Zhu, Peng Zeng

Abstract:

With the rapid development in China, the urbanization has entered the transformation and promotion stage, and its direction of development has shifted to overall regional synergy. China has a large number of towns and villages, with comparative small scale and scattered distribution, which always support and provide resources to cities leading to urban-rural opposition, so it is difficult to achieve common development in a single town or village. In this context, the regional development should focus more on towns and villages to form a synergetic system, joining the regional association with cities. Thus, the paper raises the question about how to effectively organize towns and villages system to regulate the resource allocation and improve the comprehensive value of the regional area. To answer the question, it is necessary to find a suitable research unit and analysis of its present situation of towns and villages system for optimal development. By combing relevant researches and theoretical models, the county is the most basic administrative unit in China, which can directly guide and regulate the development of towns and villages, so the paper takes county as the research unit. Following the theoretical concept of ‘three structures and one network’, the paper concludes the research framework to analyse the present situation of towns and villages system, including scale structure, functional structure, spatial structure, and organization network. The analytical methods refer to the fractal theory and gravity model, using statistics and spatial data. The scale structure analyzes rank-size dimensions and uses the principal component method to calculate the comprehensive scale of towns and villages. The functional structure analyzes the functional types and industrial development of towns and villages. The spatial structure analyzes the aggregation dimension, network dimension, and correlation dimension of spatial elements to represent the overall spatial relationships. In terms of organization network, from the perspective of entity and ono-entity, the paper analyzes the transportation network and gravitational network. Based on the present situation analysis, the optimization strategies are proposed in order to achieve a synergetic relationship between towns and villages in the county area. The paper uses Suning county in the Beijing-Tianjin-Hebei region as a case study to apply the research framework and methods and then proposes the optimization orientations. The analysis results indicate that: (1) The Suning county is lack of medium-scale towns to transfer effect from towns to villages. (2) The distribution of gravitational centers is uneven, and the effect of gravity is limited only for nearby towns and villages. The gravitational network is not complete, leading to economic activities scattered and isolated. (3) The overall development of towns and villages system is immature, staying at ‘single heart and multi-core’ stage, and some specific optimization strategies are proposed. This study provides a regional view for the development of towns and villages and concludes the research framework and methods of towns and villages system for forming an effective synergetic relationship between them, contributing to organize resources and stimulate endogenous motivation, and form counter magnets to join the urban-rural integration.

Keywords: towns and villages system, organization structure, county area, fractal theory, gravity model

Procedia PDF Downloads 136
1524 Quality Assessment of New Zealand Mānuka Honeys Using Hyperspectral Imaging Combined with Deep 1D-Convolutional Neural Networks

Authors: Hien Thi Dieu Truong, Mahmoud Al-Sarayreh, Pullanagari Reddy, Marlon M. Reis, Richard Archer

Abstract:

New Zealand mānuka honey is a honeybee product derived mainly from Leptospermum scoparium nectar. The potent antibacterial activity of mānuka honey derives principally from methylglyoxal (MGO), in addition to the hydrogen peroxide and other lesser activities present in all honey. MGO is formed from dihydroxyacetone (DHA) unique to L. scoparium nectar. Mānuka honey also has an idiosyncratic phenolic profile that is useful as a chemical maker. Authentic mānuka honey is highly valuable, but almost all honey is formed from natural mixtures of nectars harvested by a hive over a time period. Once diluted by other nectars, mānuka honey irrevocably loses value. We aimed to apply hyperspectral imaging to honey frames before bulk extraction to minimise the dilution of genuine mānuka by other honey and ensure authenticity at the source. This technology is non-destructive and suitable for an industrial setting. Chemometrics using linear Partial Least Squares (PLS) and Support Vector Machine (SVM) showed limited efficacy in interpreting chemical footprints due to large non-linear relationships between predictor and predictand in a large sample set, likely due to honey quality variability across geographic regions. Therefore, an advanced modelling approach, one-dimensional convolutional neural networks (1D-CNN), was investigated for analysing hyperspectral data for extraction of biochemical information from honey. The 1D-CNN model showed superior prediction of honey quality (R² = 0.73, RMSE = 2.346, RPD= 2.56) to PLS (R² = 0.66, RMSE = 2.607, RPD= 1.91) and SVM (R² = 0.67, RMSE = 2.559, RPD=1.98). Classification of mono-floral manuka honey from multi-floral and non-manuka honey exceeded 90% accuracy for all models tried. Overall, this study reveals the potential of HSI and deep learning modelling for automating the evaluation of honey quality in frames.

Keywords: mānuka honey, quality, purity, potency, deep learning, 1D-CNN, chemometrics

Procedia PDF Downloads 139
1523 Structured-Ness and Contextual Retrieval Underlie Language Comprehension

Authors: Yao-Ying Lai, Maria Pinango, Ashwini Deo

Abstract:

While grammatical devices are essential to language processing, how comprehension utilizes cognitive mechanisms is less emphasized. This study addresses this issue by probing the complement coercion phenomenon: an entity-denoting complement following verbs like begin and finish receives an eventive interpretation. For example, (1) “The queen began the book” receives an agentive reading like (2) “The queen began [reading/writing/etc.…] the book.” Such sentences engender additional processing cost in real-time comprehension. The traditional account attributes this cost to an operation that coerces the entity-denoting complement to an event, assuming that these verbs require eventive complements. However, in closer examination, examples like “Chapter 1 began the book” undermine this assumption. An alternative, Structured Individual (SI) hypothesis, proposes that the complement following aspectual verbs (AspV; e.g. begin, finish) is conceptualized as a structured individual, construed as an axis along various dimensions (e.g. spatial, eventive, temporal, informational). The composition of an animate subject and an AspV such as (1) engenders an ambiguity between an agentive reading along the eventive dimension like (2), and a constitutive reading along the informational/spatial dimension like (3) “[The story of the queen] began the book,” in which the subject is interpreted as a subpart of the complement denotation. Comprehenders need to resolve the ambiguity by searching contextual information, resulting in additional cost. To evaluate the SI hypothesis, a questionnaire was employed. Method: Target AspV sentences such as “Shakespeare began the volume.” were preceded by one of the following types of context sentence: (A) Agentive-biasing, in which an event was mentioned (…writers often read…), (C) Constitutive-biasing, in which a constitutive meaning was hinted (Larry owns collections of Renaissance literature.), (N) Neutral context, which allowed both interpretations. Thirty-nine native speakers of English were asked to (i) rate each context-target sentence pair from a 1~5 scale (5=fully understandable), and (ii) choose possible interpretations for the target sentence given the context. The SI hypothesis predicts that comprehension is harder for the Neutral condition, as compared to the biasing conditions because no contextual information is provided to resolve an ambiguity. Also, comprehenders should obtain the specific interpretation corresponding to the context type. Results: (A) Agentive-biasing and (C) Constitutive-biasing were rated higher than (N) Neutral conditions (p< .001), while all conditions were within the acceptable range (> 3.5 on the 1~5 scale). This suggests that when lacking relevant contextual information, semantic ambiguity decreases comprehensibility. The interpretation task shows that the participants selected the biased agentive/constitutive reading for condition (A) and (C) respectively. For the Neutral condition, the agentive and constitutive readings were chosen equally often. Conclusion: These findings support the SI hypothesis: the meaning of AspV sentences is conceptualized as a parthood relation involving structured individuals. We argue that semantic representation makes reference to spatial structured-ness (abstracted axis). To obtain an appropriate interpretation, comprehenders utilize contextual information to enrich the conceptual representation of the sentence in question. This study connects semantic structure to human’s conceptual structure, and provides a processing model that incorporates contextual retrieval.

Keywords: ambiguity resolution, contextual retrieval, spatial structured-ness, structured individual

Procedia PDF Downloads 333
1522 Impact of Profitability, Slack Resources and Natural Disasters on China's Corporate Philanthropic Practices

Authors: Nabeel Safdar, Qian Aimin

Abstract:

Corporate philanthropy is important, as the donations have been considered as a source to improve the image of business entity in modern era of high competition. We used data on annual basis from 2000 to 2014 for 1,248 firms listed at Shanghai and Shenzhen stock exchanges. Results for giving firms reveal that there is curve linear relation of profitability and CP, as profitable firms utilize cash in an efficient way and have fewer amounts of slack resource and tradeoff among stakeholder and agency cost made it more justifiable. We found that more profitability does not mean that the cash flows are available, actually good performing firms or profitable firm also good at cash management. Cash is utilized in an effective way by profitable firms, and have fewer extents of slack resources which generate curvilinear relationship of profitability with Corporate Philanthropy. We found that the trend of Corporate Philanthropy also got affected due to natural disasters. Analysis made by innovation, slack resources and directors salary revealed the positive significant relationship. It is not compulsory that firm should be only profitable for engaging in philanthropy rather they should have abundant slack resources to donate.

Keywords: corporate philanthropy, free cash flows, natural disasters, profitability

Procedia PDF Downloads 311
1521 21st Century Provocation: Modern Slavery, the Implications for Individuals on the Autism Spectrum

Authors: Christina Surmei

Abstract:

Autism Spectrum Disorder (ASD) is defined as a diverse range of developmental conditions that affect an individual’s functionality. ASD is not linear, and individuals can present with deficits in social interaction, communication, and demonstrate limited, repetitive patterns of behaviour, interests, or activities. These characteristics may be observed in a variety of ways and range from mild to severe. ASD may include autism disorder, pervasive developmental disorder not otherwise specified, Asperger’s, or other related pervasive developmental disorders. Modern slavery is defined as 'situations of exploitation that a person cannot refuse or leave because of threats, violence, coercion, and abuse of power or deception'. A review of the literature investigated the prevalence of research regarding ASD and modern slavery. Two universal search engines and five online journals were used as the apparatuses of inquiry. The results revealed two editorials, one study, and one act, totaling four publications attesting to ASD and modern slavery as a joint entity. This is representative of a vast absence of research. However, as individual entities research on autism and modern slavery is in a general high occurrence. This paper has identified a significant gap in research on ASD and modern slavery, and initiates the dialogue to unpack a significant global issue in society today.

Keywords: autism spectrum, education, modern slavery, support

Procedia PDF Downloads 168
1520 Purification of Zr from Zr-Hf Resources Using Crystallization in HF-HCl Solvent Mixture

Authors: Kenichi Hirota, Jifeng Wang, Sadao Araki, Koji Endo, Hideki Yamamoto

Abstract:

Zirconium (Zr) has been used as a fuel cladding tube for nuclear reactors, because of the excellent corrosion resistance and the low adsorptive material for neutron. Generally speaking, the natural resource of Zr is often containing Hf that has similar properties. The content of Hf in the Zr resources is about 2~4 wt%. In the industrial use, the content of Hf in Zr resources should be lower than the 100 ppm. However, the separation of Zr and Hf is not so easy, because of similar chemical and physical properties such as melting point, boiling point and things. Solvent extraction method has been applied for the separation of Zr and Hf from Zr natural resources. This method can separate Hf with high efficiency (Hf < 100ppm), however, it needs much amount of organic solvents for solvent extraction and the cost of its disposal treatment is high. Therefore, we attached attention for the fractional crystallization. This separation method depends on the solubility difference of Zr and Hf in the solvent. In this work, hexafluorozirconate (hafnate) (K2Zr(Hf)F6) was used as model compound. Solubility of K2ZrF6 in water showed lower than that of K2HfF6. By repeating of this treatment, it is possible to purify Zr, practically. In this case, 16-18 times of recrystallization stages were needed for its high purification. The improvement of the crystallization process was carried out in this work. Water, hydrofluoric acid (HF) and hydrofluoric acid (HF) +hydrochloric acid (HCl) mixture were chosen as solvent for dissolution of Zr and Hf. In the experiment, 10g of K2ZrF6 was added to each solvent of 100mL. Each solution was heated for 1 hour at 353K. After 1h of this operation, they were cooled down till 293K, and were held for 5 hours at 273K. Concentration of Zr or Hf was measured using ICP analysis. It was found that Hf was separated from Zr-Hf mixed compound with high efficiency, when HF-HCl solution was used for solvent of crystallization. From the comparison of the particle size of each crystal by SEM, it was confirmed that the particle diameter of the crystal showed smaller size with decreasing of Hf content. This paper concerned with purification of Zr from Zr-Hf mixture using crystallization method.

Keywords: crystallization, zirconium, hafnium, separation

Procedia PDF Downloads 438