Search results for: mapping code ultrawideband
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 2495

Search results for: mapping code ultrawideband

2255 Myanmar Character Recognition Using Eight Direction Chain Code Frequency Features

Authors: Kyi Pyar Zaw, Zin Mar Kyu

Abstract:

Character recognition is the process of converting a text image file into editable and searchable text file. Feature Extraction is the heart of any character recognition system. The character recognition rate may be low or high depending on the extracted features. In the proposed paper, 25 features for one character are used in character recognition. Basically, there are three steps of character recognition such as character segmentation, feature extraction and classification. In segmentation step, horizontal cropping method is used for line segmentation and vertical cropping method is used for character segmentation. In the Feature extraction step, features are extracted in two ways. The first way is that the 8 features are extracted from the entire input character using eight direction chain code frequency extraction. The second way is that the input character is divided into 16 blocks. For each block, although 8 feature values are obtained through eight-direction chain code frequency extraction method, we define the sum of these 8 feature values as a feature for one block. Therefore, 16 features are extracted from that 16 blocks in the second way. We use the number of holes feature to cluster the similar characters. We can recognize the almost Myanmar common characters with various font sizes by using these features. All these 25 features are used in both training part and testing part. In the classification step, the characters are classified by matching the all features of input character with already trained features of characters.

Keywords: chain code frequency, character recognition, feature extraction, features matching, segmentation

Procedia PDF Downloads 320
2254 Preliminary Seismic Hazard Mapping of Papua New Guinea

Authors: Hadi Ghasemi, Mark Leonard, Spiliopoulos Spiro, Phil Cummins, Mathew Moihoi, Felix Taranu, Eric Buri, Chris Mckee

Abstract:

In this study the level of seismic hazard in terms of Peak Ground Acceleration (PGA) was calculated for return period of 475 years, using modeled seismic sources and assigned ground-motion equations. The calculations were performed for bedrock site conditions (Vs30=760 m/s). From the results it is evident that the seismic hazard reaches its maximum level (i.e. PGA≈1g for 475 yr return period) at the Huon Peninsula and southern New Britain regions. Disaggregation analysis revealed that moderate to large earthquakes occurring along the New Britain Trench mainly control the level of hazard at these locations. The open-source computer program OpenQuake developed by Global Earthquake Model foundation was used for the seismic hazard computations. It should be emphasized that the presented results are still preliminary and should not be interpreted as our final assessment of seismic hazard in PNG.

Keywords: probabilistic seismic hazard assessment, Papua New Guinea, building code, OpenQuake

Procedia PDF Downloads 556
2253 Mapping of Siltations of AlKhod Dam, Muscat, Sultanate of Oman Using Low-Cost Multispectral Satellite Data

Authors: Sankaran Rajendran

Abstract:

Remote sensing plays a vital role in mapping of resources and monitoring of environments of the earth. In the present research study, mapping and monitoring of clay siltations occurred in the Alkhod Dam of Muscat, Sultanate of Oman are carried out using low-cost multispectral Landsat and ASTER data. The dam is constructed across the Wadi Samail catchment for ground water recharge. The occurrence and spatial distribution of siltations in the dam are studied with five years of interval from the year 1987 of construction to 2014. The deposits are mainly due to the clay, sand, and silt occurrences derived from the weathering rocks of ophiolite sequences occurred in the Wadi Samail catchment. The occurrences of clays are confirmed by minerals identification using ASTER VNIR-SWIR spectral bands and Spectral Angle Mapper supervised image processing method. The presence of clays and their spatial distribution are verified in the field. The study recommends the technique and the low-cost satellite data to similar region of the world.

Keywords: Alkhod Dam, ASTER siltation, Landsat, remote sensing, Oman

Procedia PDF Downloads 437
2252 The Connection between the Schwartz Theory of Basic Values and Ethical Principles in Clinical Psychology

Authors: Matej Stritesky

Abstract:

The research deals with the connection between the Schwartz Theory of Basic Values and the ethical principles in psychology, on which the meta-code of ethics the European Federation of Psychological Associations is based. The research focuses on ethically problematic situations in clinical psychology in the Czech Republic. Based on the analysis of papers that identified ethically problematic situations faced by clinical psychologists, a questionnaire of ethically problematic situations in clinical psychology (EPSCP) was created for the purposes of the research. The questionnaire was created to represent situations that correspond to the 4 principles on which the meta-code of ethics the European Federation of Psychological Associations is based. The questionnaire EPSCP consists of descriptions of 32 situations that respondents evaluate on a scale from 1 (psychologist's behaviour is ethically perfectly fine) to 10 (psychologist's behaviour is ethically completely unacceptable). The EPSCP questionnaire, together with Schwartz's PVQ questionnaire, will be presented to 60 psychology students. The relationship between principles in clinical psychology and the values on Schwartz´s value continuum will be described using multidimensional scaling. A positive correlation is assumed between the higher-order value of openness to change and problematic ethical situations related to the principle of integrity; a positive correlation between the value of the higher order of self-transcendence and the principle of respect and responsibility; a positive correlation between the value of the higher order of conservation and the principle of competence; and negative correlation between the value of the higher order of ego strengthening and sensitivity to ethically problematic situations. The research also includes an experimental part. The first half of the students are presented with the code of ethics of the Czech Association of Clinical Psychologists before completing the questionnaires, and to the second half of the students is the code of ethics presented after completing the questionnaires. In addition to reading the code of ethics, students describe the three rules of the code of ethics that they consider most important and state why they chose these rules. The output of the experimental part will be to determine whether the presentation of the code of ethics leads to greater sensitivity to ethically problematic situations.

Keywords: clinical psychology, ethically problematic situations in clinical psychology, ethical principles in psychology, Schwartz theory of basic values

Procedia PDF Downloads 112
2251 Contribution of Spatial Teledetection to the Geological Mapping of the Imiter Buttonhole: Application to the Mineralized Structures of the Principal Corps B3 (CPB3) of the Imiter Mine (Anti-atlas, Morocco)

Authors: Bouayachi Ali, Alikouss Saida, Baroudi Zouhir, Zerhouni Youssef, Zouhair Mohammed, El Idrissi Assia, Essalhi Mourad

Abstract:

The world-class Imiter silver deposit is located on the northern flank of the Precambrian Imiter buttonhole. This deposit is formed by epithermal veins hosted in the sandstone-pelite formations of the lower complex and in the basic conglomerates of the upper complex, these veins are controlled by a regional scale fault cluster, oriented N70°E to N90°E. The present work on the contribution of remote sensing on the geological mapping of the Imiter buttonhole and application to the mineralized structures of the Principal Corps B3. Mapping on satellite images is a very important tool in mineral prospecting. It allows the localization of the zones of interest in order to orientate the field missions by helping the localization of the major structures which facilitates the interpretation, the programming and the orientation of the mining works. The predictive map also allows for the correction of field mapping work, especially the direction and dimensions of structures such as dykes, corridors or scrapings. The use of a series of processing such as SAM, PCA, MNF and unsupervised and supervised classification on a Landsat 8 satellite image of the study area allowed us to highlight the main facies of the Imite area. To improve the exploration research, we used another processing that allows to realize a spatial distribution of the alteration mineral indices, and the application of several filters on the different bands to have lineament maps.

Keywords: principal corps B3, teledetection, Landsat 8, Imiter II, silver mineralization, lineaments

Procedia PDF Downloads 95
2250 Detecting Heartbeat Architectural Tactic in Source Code Using Program Analysis

Authors: Ananta Kumar Das, Sujit Kumar Chakrabarti

Abstract:

Architectural tactics such as heartbeat, ping-echo, encapsulate, encrypt data are techniques that are used to achieve quality attributes of a system. Detecting architectural tactics has several benefits: it can aid system comprehension (e.g., legacy systems) and in the estimation of quality attributes such as safety, security, maintainability, etc. Architectural tactics are typically spread over the source code and are implicit. For large codebases, manual detection is often not feasible. Therefore, there is a need for automated methods of detection of architectural tactics. This paper presents a formalization of the heartbeat architectural tactic and a program analytic approach to detect this tactic in source code. The experiment of the proposed method is done on a set of Java applications. The outcome of the experiment strongly suggests that the method compares well with a manual approach in terms of its sensitivity and specificity, and far supersedes a manual exercise in terms of its scalability.

Keywords: software architecture, architectural tactics, detecting architectural tactics, program analysis, AST, alias analysis

Procedia PDF Downloads 160
2249 Prompt Design for Code Generation in Data Analysis Using Large Language Models

Authors: Lu Song Ma Li Zhi

Abstract:

With the rapid advancement of artificial intelligence technology, large language models (LLMs) have become a milestone in the field of natural language processing, demonstrating remarkable capabilities in semantic understanding, intelligent question answering, and text generation. These models are gradually penetrating various industries, particularly showcasing significant application potential in the data analysis domain. However, retraining or fine-tuning these models requires substantial computational resources and ample downstream task datasets, which poses a significant challenge for many enterprises and research institutions. Without modifying the internal parameters of the large models, prompt engineering techniques can rapidly adapt these models to new domains. This paper proposes a prompt design strategy aimed at leveraging the capabilities of large language models to automate the generation of data analysis code. By carefully designing prompts, data analysis requirements can be described in natural language, which the large language model can then understand and convert into executable data analysis code, thereby greatly enhancing the efficiency and convenience of data analysis. This strategy not only lowers the threshold for using large models but also significantly improves the accuracy and efficiency of data analysis. Our approach includes requirements for the precision of natural language descriptions, coverage of diverse data analysis needs, and mechanisms for immediate feedback and adjustment. Experimental results show that with this prompt design strategy, large language models perform exceptionally well in multiple data analysis tasks, generating high-quality code and significantly shortening the data analysis cycle. This method provides an efficient and convenient tool for the data analysis field and demonstrates the enormous potential of large language models in practical applications.

Keywords: large language models, prompt design, data analysis, code generation

Procedia PDF Downloads 39
2248 Skin-Dose Mapping for Patients Undergoing Interventional Radiology Procedures: Clinical Experimentations versus a Mathematical Model

Authors: Aya Al Masri, Stefaan Carpentier, Fabrice Leroy, Thibault Julien, Safoin Aktaou, Malorie Martin, Fouad Maaloul

Abstract:

Introduction: During an 'Interventional Radiology (IR)' procedure, the patient's skin-dose may become very high for a burn, necrosis and ulceration to appear. In order to prevent these deterministic effects, an accurate calculation of the patient skin-dose mapping is essential. For most machines, the 'Dose Area Product (DAP)' and fluoroscopy time are the only information available for the operator. These two parameters are a very poor indicator of the peak skin dose. We developed a mathematical model that reconstructs the magnitude (delivered dose), shape, and localization of each irradiation field on the patient skin. In case of critical dose exceeding, the system generates warning alerts. We present the results of its comparison with clinical studies. Materials and methods: Two series of comparison of the skin-dose mapping of our mathematical model with clinical studies were performed: 1. At a first time, clinical tests were performed on patient phantoms. Gafchromic films were placed on the table of the IR machine under of PMMA plates (thickness = 20 cm) that simulate the patient. After irradiation, the film darkening is proportional to the radiation dose received by the patient's back and reflects the shape of the X-ray field. After film scanning and analysis, the exact dose value can be obtained at each point of the mapping. Four experimentation were performed, constituting a total of 34 acquisition incidences including all possible exposure configurations. 2. At a second time, clinical trials were launched on real patients during real 'Chronic Total Occlusion (CTO)' procedures for a total of 80 cases. Gafchromic films were placed at the back of patients. We performed comparisons on the dose values, as well as the distribution, and the shape of irradiation fields between the skin dose mapping of our mathematical model and Gafchromic films. Results: The comparison between the dose values shows a difference less than 15%. Moreover, our model shows a very good geometric accuracy: all fields have the same shape, size and location (uncertainty < 5%). Conclusion: This study shows that our model is a reliable tool to warn physicians when a high radiation dose is reached. Thus, deterministic effects can be avoided.

Keywords: clinical experimentation, interventional radiology, mathematical model, patient's skin-dose mapping.

Procedia PDF Downloads 140
2247 Simulation of Reflectometry in Alborz Tokamak

Authors: S. Kohestani, R. Amrollahi, P. Daryabor

Abstract:

Microwave diagnostics such as reflectometry are receiving growing attention in magnetic confinement fusionresearch. In order to obtain the better understanding of plasma confinement physics, more detailed measurements on density profile and its fluctuations might be required. A 2D full-wave simulation of ordinary mode propagation has been written in an effort to model effects seen in reflectometry experiment. The code uses the finite-difference-time-domain method with a perfectly-matched-layer absorption boundary to solve Maxwell’s equations.The code has been used to simulate the reflectometer measurement in Alborz Tokamak.

Keywords: reflectometry, simulation, ordinary mode, tokamak

Procedia PDF Downloads 420
2246 Redefining Infrastructure as Code Orchestration Using AI

Authors: Georges Bou Ghantous

Abstract:

This research delves into the transformative impact of Artificial Intelligence (AI) on Infrastructure as Code (IaaC) practices, specifically focusing on the redefinition of infrastructure orchestration. By harnessing AI technologies such as machine learning algorithms and predictive analytics, organizations can achieve unprecedented levels of efficiency and optimization in managing their infrastructure resources. AI-driven IaaC introduces proactive decision-making through predictive insights, enabling organizations to anticipate and address potential issues before they arise. Dynamic resource scaling, facilitated by AI, ensures that infrastructure resources can seamlessly adapt to fluctuating workloads and changing business requirements. Through case studies and best practices, this paper sheds light on the tangible benefits and challenges associated with AI-driven IaaC transformation, providing valuable insights for organizations navigating the evolving landscape of digital infrastructure management.

Keywords: artificial intelligence, infrastructure as code, efficiency optimization, predictive insights, dynamic resource scaling, proactive decision-making

Procedia PDF Downloads 34
2245 Numerical Modelling of Shear Zone and Its Implications on Slope Instability at Letšeng Diamond Open Pit Mine, Lesotho

Authors: M. Ntšolo, D. Kalumba, N. Lefu, G. Letlatsa

Abstract:

Rock mass damage due to shear tectonic activity has been investigated largely in geoscience where fluid transport is of major interest. However, little has been studied on the effect of shear zones on rock mass behavior and its impact on stability of rock slopes. At Letšeng Diamonds open pit mine in Lesotho, the shear zone composed of sheared kimberlite material, calcite and altered basalt is forming part of the haul ramp into the main pit cut 3. The alarming rate at which the shear zone is deteriorating has triggered concerns about both local and global stability of pit the walls. This study presents the numerical modelling of the open pit slope affected by shear zone at Letšeng Diamond Mine (LDM). Analysis of the slope involved development of the slope model by using a two-dimensional finite element code RS2. Interfaces between shear zone and host rock were represented by special joint elements incorporated in the finite element code. The analysis of structural geological mapping data provided a good platform to understand the joint network. Major joints including shear zone were incorporated into the model for simulation. This approach proved successful by demonstrating that continuum modelling can be used to evaluate evolution of stresses, strain, plastic yielding and failure mechanisms that are consistent with field observations. Structural control due to geological shear zone structure proved to be important in its location, size and orientation. Furthermore, the model analyzed slope deformation and sliding possibility along shear zone interfaces. This type of approach can predict shear zone deformation and failure mechanism, hence mitigation strategies can be deployed for safety of human lives and property within mine pits.

Keywords: numerical modeling, open pit mine, shear zone, slope stability

Procedia PDF Downloads 299
2244 Feature Engineering Based Detection of Buffer Overflow Vulnerability in Source Code Using Deep Neural Networks

Authors: Mst Shapna Akter, Hossain Shahriar

Abstract:

One of the most important challenges in the field of software code audit is the presence of vulnerabilities in software source code. Every year, more and more software flaws are found, either internally in proprietary code or revealed publicly. These flaws are highly likely exploited and lead to system compromise, data leakage, or denial of service. C and C++ open-source code are now available in order to create a largescale, machine-learning system for function-level vulnerability identification. We assembled a sizable dataset of millions of opensource functions that point to potential exploits. We developed an efficient and scalable vulnerability detection method based on deep neural network models that learn features extracted from the source codes. The source code is first converted into a minimal intermediate representation to remove the pointless components and shorten the dependency. Moreover, we keep the semantic and syntactic information using state-of-the-art word embedding algorithms such as glove and fastText. The embedded vectors are subsequently fed into deep learning networks such as LSTM, BilSTM, LSTM-Autoencoder, word2vec, BERT, and GPT-2 to classify the possible vulnerabilities. Furthermore, we proposed a neural network model which can overcome issues associated with traditional neural networks. Evaluation metrics such as f1 score, precision, recall, accuracy, and total execution time have been used to measure the performance. We made a comparative analysis between results derived from features containing a minimal text representation and semantic and syntactic information. We found that all of the deep learning models provide comparatively higher accuracy when we use semantic and syntactic information as the features but require higher execution time as the word embedding the algorithm puts on a bit of complexity to the overall system.

Keywords: cyber security, vulnerability detection, neural networks, feature extraction

Procedia PDF Downloads 89
2243 Spatial Orientation of Land Use Activities along Buffalo River Estuary: A Study in Buffalo City Metropolitan Municipality, Eastern Cape South Africa

Authors: A. Ngunga, M. K. Soviti, S. Nakin

Abstract:

South Africa is one of the developing countries rich in estuary ecosystem. Previous studies have identified many impacts of land use activities on the pollution status of the estuaries. These land use activity and related practices are often blamed for the many pollution problems affecting the estuaries. For example, the estuarine ecosystems on a global scale are experiencing vast transformations from anthropogenic influences; Buffalo River Estuary is one of the influenced estuaries whereby the sources of pollution are unknown. These problems consequently lead to the degradation of the estuaries. The aim of the research was to establish the factors that have the potential to impact pollution status of Buffalo river estuary. Study focuses on Identifying and mapping land use activities along Buffalo River Estuary. Questionnaire survey, structured interviews, direct observation, GPS survey and ArcGIS mapping were the methods used for data collection in the area, and results were analyzed and presented by ANOVA and Microsoft Excel statistical methods. The results showed that harbour is the main source of pollution, in Buffalo River Estuary, through Ballast water discharge. Therefore that requires more concern for protecting and cleaning the estuary.

Keywords: estuary, land-use activities, pollution, mapping, water pollution

Procedia PDF Downloads 194
2242 Underrepresentation of Right Middle Cerebral Infarct: A Statistical Parametric Mapping

Authors: Wi-Sun Ryu, Eun-Kee Bae

Abstract:

Prior studies have shown that patients with right hemispheric stroke are likely to seek medical service compared with those with left hemispheric stroke. However, the underlying mechanism for this phenomenon is unknown. In the present study, we generated lesion probability maps in a patient with right and left middle cerebral artery infarct and statistically compared. We found that precentral gyrus-Brodmann area 44, a language area in the left hemisphere - involvement was significantly higher in patients with left hemispheric stroke. This finding suggests that a language dysfunction was more noticeable, thereby taking more patients to hospitals.

Keywords: cerebral infarct, brain MRI, statistical parametric mapping, middle cerebral infarct

Procedia PDF Downloads 338
2241 Clinch Process Simulation Using Diffuse Elements

Authors: Benzegaou Ali, Brani Benabderrahmane

Abstract:

This work describes a numerical study of the TOX–clinching process using diffuse elements. A computer code baptized SEMA "Static Explicit Method Analysis" is developed to simulate the clinch joining process. The FE code is based on an Updated Lagrangian scheme. The used resolution method is based on an explicit static approach. The integration of the elasto-plastic behavior law is realized with an algorithm of Simo and Taylor. The tools are represented by plane facets.

Keywords: diffuse elements, numerical simulation, clinching, contact, large deformation

Procedia PDF Downloads 363
2240 Dissection of Genomic Loci for Yellow Vein Mosaic Virus Resistance in Okra (Abelmoschus esculentas)

Authors: Rakesh Kumar Meena, Tanushree Chatterjee

Abstract:

Okra (Abelmoschus esculentas L. Moench) or lady’s finger is an important vegetable crop belonging to the Malvaceae family. Unfortunately, production and productivity of Okra are majorly affected by Yellow Vein mosaic virus (YVMV). The AO: 189 (resistant parent) X AO: 191(susceptible parent) used for the development of mapping population. The mapping population has 143 individuals (F₂:F₃). Population was characterized by physiological and pathological observations. Screening of 360 DNA markers was performed to survey for parental polymorphism between the contrasting parents’, i.e., AO: 189 and AO: 191. Out of 360; 84 polymorphic markers were used for genotyping of the mapping population. Total markers were distributed into four linkage groups (LG1, LG2, LG3, and LG4). LG3 covered the longest span (106.8cM) with maximum number of markers (27) while LG1 represented the smallest linkage group in terms of length (71.2cM). QTL identification using the composite interval mapping approach detected two prominent QTLs, QTL1 and QTL2 for resistance against YVMV disease. These QTLs were placed between the marker intervals of NBS-LRR72-Path02 and NBS-LRR06- NBS-LRR65 on linkage group 02 and linkage group 04 respectively. The LOD values of QTL1 and QTL2 were 5.7 and 6.8 which accounted for 19% and 27% of the total phenotypic variation, respectively. The findings of this study provide two linked markers which can be used as efficient diagnostic tools to distinguish between YVMV resistant and susceptible Okra cultivars/genotypes. Lines identified as highly resistant against YVMV infection can be used as donor lines for this trait. This will be instrumental in accelerating the trait improvement program in Okra and will substantially reduce the yield losses due to this viral disease.

Keywords: Okra, yellow vein mosaic virus, resistant, linkage map, QTLs

Procedia PDF Downloads 215
2239 Low-Cost Parking Lot Mapping and Localization for Home Zone Parking Pilot

Authors: Hongbo Zhang, Xinlu Tang, Jiangwei Li, Chi Yan

Abstract:

Home zone parking pilot (HPP) is a fast-growing segment in low-speed autonomous driving applications. It requires the car automatically cruise around a parking lot and park itself in a range of up to 100 meters inside a recurrent home/office parking lot, which requires precise parking lot mapping and localization solution. Although Lidar is ideal for SLAM, the car OEMs favor a low-cost fish-eye camera based visual SLAM approach. Recent approaches have employed segmentation models to extract semantic features and improve mapping accuracy, but these AI models are memory unfriendly and computationally expensive, making deploying on embedded ADAS systems difficult. To address this issue, we proposed a new method that utilizes object detection models to extract robust and accurate parking lot features. The proposed method could reduce computational costs while maintaining high accuracy. Once combined with vehicles’ wheel-pulse information, the system could construct maps and locate the vehicle in real-time. This article will discuss in detail (1) the fish-eye based Around View Monitoring (AVM) with transparent chassis images as the inputs, (2) an Object Detection (OD) based feature point extraction algorithm to generate point cloud, (3) a low computational parking lot mapping algorithm and (4) the real-time localization algorithm. At last, we will demonstrate the experiment results with an embedded ADAS system installed on a real car in the underground parking lot.

Keywords: ADAS, home zone parking pilot, object detection, visual SLAM

Procedia PDF Downloads 67
2238 Regional Review of Outcome of Cervical Smears Reported with Cytological Features of Non Cervical Glandular Neoplasia

Authors: Uma Krishnamoorthy, Vivienne Beavers, Janet Marshall

Abstract:

Introduction: Cervical cytology showing features raising the suspicion of non cervical glandular neoplasia are reported as code 0 under the United Kingdom National Health Service Cervical screening programme ( NHSCSP). As the suspicion is regarding non cervical neoplasia, smear is reported as normal and patient informed that cervical screening result is normal. GP receives copy of results where it states further referral is indicated in small font within text of report. Background: There were several incidents of delayed diagnosis of endometrial cancer in Lancashire which prompted this Northwest Regional review to enable an understanding of underlying pathology outcome of code zero smears to raise awareness and also to review whether further action on wording of smear results was indicated to prevent such delay. Methodology: All Smears reported at the Manchester cytology centre who process cytology for Lancashire population from March 2013 to March 2014 were reviewed and histological diagnosis outcome of women in whom smear was reported as code zero was reviewed retrospectively . Results: Total smears reported by the cytology centre during this period was approximately 109400. Reports issued with result code 0 among this during this time period was 49.Results revealed that among three fourth (37) of women with code zero smear (N=49), evidence of underlying pathology of non cervical origin was confirmed. Of this, 73 % (36) were due to endometrial pathology with 49 % (24) endometrial carcinoma, 12 % (6)polyp, 4 % atypical endometrial hyperplasia (2), 6 % endometrial hyperplasia without atypia (3), and 2 % adenomyosis (1 case) and 2 % ( 1 case) due to ovarian adenocarcinoma. Conclusion: This review demonstrated that more than half (51 %) of women with a code 0 smear report were diagnosed with underlying carcinoma and 75 % had a confirmed underlying pathology contributory to code 0 smear findings. Recommendations and Action Plan: A local rapid access referral and management pathway for this group of women was implemented as a result of this in our unit. The findings and Pathway were shared with other regional units served by the cytology centre through the Pan Lancashire cervical screening board and through the Cytology centre. Locally, the smear report wording was updated to include a rubber stamp/ print in "Red Bold letters" stating that " URGENT REFERRAL TO GYNAECOLOGY IS INDICATED". Findings were also shared through the Pan Lancashire board with National cervical screening programme board, and revisions to wording of code zero smear reports to highlight the need for Urgent referral has now been agreed at National level to be implemented.

Keywords: code zero smears, endometrial cancer, non cervical glandular neoplasia, ovarian cancer

Procedia PDF Downloads 297
2237 Steady State Modeling and Simulation of an Industrial Steam Boiler

Authors: Amina Lyria Deghal Cheridi, Abla Chaker, Ahcene Loubar

Abstract:

Relap5 system code is one among powerful tools, which is used in the area of design and safety evaluation. This work aims to simulate the behavior of a radiant steam boiler at the steady-state conditions using Relap5 code system. To perform this study, a detailed Relap5 model is built including all the parts of the steam boiler. The control and regulation systems are also considered. To reproduce the most important parameters and phenomena with an acceptable accuracy and fidelity, a strong qualification work is undertaken concerning the facility nodalization. It consists of making a comparison between the code results and the plant available data in steady-state operation mode. Therefore, the model qualification results at the steady-state are in good agreement with the steam boiler experimental data. The steam boiler Relap5 model has proved satisfactory; and the model was capable of predicting the main thermal-hydraulic steady-state conditions of the steam boiler.

Keywords: industrial steam boiler, model qualification, natural circulation, relap5/mod3.2, steady state simulation

Procedia PDF Downloads 271
2236 Running the Athena Vortex Lattice Code in JAVA through the Java Native Interface

Authors: Paul Okonkwo, Howard Smith

Abstract:

This paper describes a methodology to integrate the Athena Vortex Lattice Aerodynamic Software for automated operation in a multivariate optimisation of the Blended Wing Body Aircraft. The Athena Vortex Lattice code developed at the Massachusetts Institute of Technology allows for the aerodynamic analysis of aircraft using the vortex lattice method. Ordinarily, the Athena Vortex Lattice operation requires a text file containing the aircraft geometry to be loaded into the AVL solver in order to determine the aerodynamic forces and moments. However, automated operation will be required to enable integration into a multidisciplinary optimisation framework. Automated AVL operation within the JAVA design environment will nonetheless require a modification and recompilation of AVL source code into an executable file capable of running on windows and other platforms without the –X11 libraries. This paper describes the procedure for the integrating the FORTRAN written AVL software for automated operation within the multivariate design synthesis optimisation framework for the conceptual design of the BWB aircraft.

Keywords: aerodynamics, automation, optimisation, AVL, JNI

Procedia PDF Downloads 565
2235 Landslide Susceptibility Mapping: A Comparison between Logistic Regression and Multivariate Adaptive Regression Spline Models in the Municipality of Oudka, Northern of Morocco

Authors: S. Benchelha, H. C. Aoudjehane, M. Hakdaoui, R. El Hamdouni, H. Mansouri, T. Benchelha, M. Layelmam, M. Alaoui

Abstract:

The logistic regression (LR) and multivariate adaptive regression spline (MarSpline) are applied and verified for analysis of landslide susceptibility map in Oudka, Morocco, using geographical information system. From spatial database containing data such as landslide mapping, topography, soil, hydrology and lithology, the eight factors related to landslides such as elevation, slope, aspect, distance to streams, distance to road, distance to faults, lithology map and Normalized Difference Vegetation Index (NDVI) were calculated or extracted. Using these factors, landslide susceptibility indexes were calculated by the two mentioned methods. Before the calculation, this database was divided into two parts, the first for the formation of the model and the second for the validation. The results of the landslide susceptibility analysis were verified using success and prediction rates to evaluate the quality of these probabilistic models. The result of this verification was that the MarSpline model is the best model with a success rate (AUC = 0.963) and a prediction rate (AUC = 0.951) higher than the LR model (success rate AUC = 0.918, rate prediction AUC = 0.901).

Keywords: landslide susceptibility mapping, regression logistic, multivariate adaptive regression spline, Oudka, Taounate

Procedia PDF Downloads 188
2234 Quick Similarity Measurement of Binary Images via Probabilistic Pixel Mapping

Authors: Adnan A. Y. Mustafa

Abstract:

In this paper we present a quick technique to measure the similarity between binary images. The technique is based on a probabilistic mapping approach and is fast because only a minute percentage of the image pixels need to be compared to measure the similarity, and not the whole image. We exploit the power of the Probabilistic Matching Model for Binary Images (PMMBI) to arrive at an estimate of the similarity. We show that the estimate is a good approximation of the actual value, and the quality of the estimate can be improved further with increased image mappings. Furthermore, the technique is image size invariant; the similarity between big images can be measured as fast as that for small images. Examples of trials conducted on real images are presented.

Keywords: big images, binary images, image matching, image similarity

Procedia PDF Downloads 196
2233 Analysis of Noise Environment and Acoustics Material in Residential Building

Authors: Heruanda Alviana Giska Barabah, Hilda Rasnia Hapsari

Abstract:

Acoustic phenomena create an acoustic interpretation condition that describes the characteristics of the environment. In urban areas, the tendency of heterogeneous and simultaneous human activity form a soundscape that is different from other regions, one of the characteristics of urban areas that developing the soundscape is the presence of vertical model houses or residential building. Activities both within the building and surrounding environment are able to make the soundscape with certain characteristics. The acoustics comfort of residential building becomes an important aspect, those demand lead the building features become more diverse. Initial steps in mapping acoustic conditions in a soundscape are important, this is the method to determine uncomfortable condition. Noise generated by road traffic, railway, and plane is an important consideration, especially for urban people, therefore the proper design of the building becomes very important as an effort to bring appropriate acoustics comfort. In this paper the authors developed noise mapping on the location of the residential building. Mapping done by taking some point referring to the noise source. The mapping result become the basis for modeling the acoustics wave interacted with the building model. Material selection is done based on literature study and modeling simulation using Insul by considering the absorption coefficient and Sound Transmission Class. The analysis of acoustics rays is ray tracing method using Comsol simulator software that can show the movement of acoustics rays and their interaction with a boundary. The result of this study can be used to consider boundary material in residential building as well as consideration for improving the acoustic quality in the acoustics zones that are formed.

Keywords: residential building, noise, absorption coefficient, sound transmission class, ray tracing

Procedia PDF Downloads 247
2232 Ontology Mapping with R-GNN for IT Infrastructure: Enhancing Ontology Construction and Knowledge Graph Expansion

Authors: Andrey Khalov

Abstract:

The rapid growth of unstructured data necessitates advanced methods for transforming raw information into structured knowledge, particularly in domain-specific contexts such as IT service management and outsourcing. This paper presents a methodology for automatically constructing domain ontologies using the DOLCE framework as the base ontology. The research focuses on expanding ITIL-based ontologies by integrating concepts from ITSMO, followed by the extraction of entities and relationships from domain-specific texts through transformers and statistical methods like formal concept analysis (FCA). In particular, this work introduces an R-GNN-based approach for ontology mapping, enabling more efficient entity extraction and ontology alignment with existing knowledge bases. Additionally, the research explores transfer learning techniques using pre-trained transformer models (e.g., DeBERTa-v3-large) fine-tuned on synthetic datasets generated via large language models such as LLaMA. The resulting ontology, termed IT Ontology (ITO), is evaluated against existing methodologies, highlighting significant improvements in precision and recall. This study advances the field of ontology engineering by automating the extraction, expansion, and refinement of ontologies tailored to the IT domain, thus bridging the gap between unstructured data and actionable knowledge.

Keywords: ontology mapping, knowledge graphs, R-GNN, ITIL, NER

Procedia PDF Downloads 15
2231 ROOP: Translating Sequential Code Fragments to Distributed Code Fragments Using Deep Reinforcement Learning

Authors: Arun Sanjel, Greg Speegle

Abstract:

Every second, massive amounts of data are generated, and Data Intensive Scalable Computing (DISC) frameworks have evolved into effective tools for analyzing such massive amounts of data. Since the underlying architecture of these distributed computing platforms is often new to users, building a DISC application can often be time-consuming and prone to errors. The automated conversion of a sequential program to a DISC program will consequently significantly improve productivity. However, synthesizing a user’s intended program from an input specification is complex, with several important applications, such as distributed program synthesizing and code refactoring. Existing works such as Tyro and Casper rely entirely on deductive synthesis techniques or similar program synthesis approaches. Our approach is to develop a data-driven synthesis technique to identify sequential components and translate them to equivalent distributed operations. We emphasize using reinforcement learning and unit testing as feedback mechanisms to achieve our objectives.

Keywords: program synthesis, distributed computing, reinforcement learning, unit testing, DISC

Procedia PDF Downloads 106
2230 Capture Zone of a Well Field in an Aquifer Bounded by Two Parallel Streams

Authors: S. Nagheli, N. Samani, D. A. Barry

Abstract:

In this paper, the velocity potential and stream function of capture zone for a well field in an aquifer bounded by two parallel streams with or without a uniform regional flow of any directions are presented. The well field includes any number of extraction or injection wells or a combination of both types with any pumping rates. To delineate the capture envelope, the potential and streamlines equations are derived by conformal mapping method. This method can help us to release constrains of other methods. The equations can be applied as useful tools to design in-situ groundwater remediation systems, to evaluate the surface–subsurface water interaction and to manage the water resources.

Keywords: complex potential, conformal mapping, image well theory, Laplace’s equation, superposition principle

Procedia PDF Downloads 432
2229 Rest API Based System-level Test Automation for Mobile Applications

Authors: Jisoo Song

Abstract:

Today’s mobile applications are communicating with servers more and more in order to access external services or information. Also, server-side code changes are more frequent than client-side code changes in a mobile application. The frequent changes lead to an increase in testing cost increase. To reduce costs, UI based test automation can be one of the solutions. It is a common automation technique in system-level testing. However, it can be unsuitable for mobile applications. When you automate tests based on UI elements for mobile applications, there are some limitations such as the overhead of script maintenance or the difficulty of finding invisible defects that UI elements cannot represent. To overcome these limitations, we present a new automation technique based on Rest API. You can automate system-level tests through test scripts that you write. These scripts call a series of Rest API in a user’s action sequence. This technique does not require testers to know the internal implementation details, only input and expected output of Rest API. You can easily modify test cases by modifying Rest API input values and also find problems that might not be evident from the UI level by validating output values. For example, when an application receives price information from a payment server and user cannot see it at UI level, Rest API based scripts can check whether price information is correct or not. More than 10 mobile applications at our company are being tested automatically based on Rest API scripts whenever application source code, mostly server source code, is built. We are finding defects right away by setting a script as a build job in CI server. The build job starts when application code builds are completed. This presentation will also include field cases from our company.

Keywords: case studies at SK Planet, introduction of rest API based test automation, limitations of UI based test automation

Procedia PDF Downloads 448
2228 Time-Dependent Behaviour of Reinforced Concrete Beams under Sustained and Repeated Loading

Authors: Sultan Daud, John P. Forth, Nikolaos Nikitas

Abstract:

The current study aims to highlight the loading characteristics impact on the time evolution (focusing particularly on long term effects) of the deformation of realized reinforced concrete beams. Namely the tension stiffening code provisions (i.e. within Eurocode 2) are reviewed with a clear intention to reassess their operational value and predicting capacity. In what follows the experimental programme adopted along with some preliminary findings and numerical modelling attempts are presented. For a range of long slender reinforced concrete simply supported beams (4200 mm) constant static sustained and repeated cyclic loadings were applied mapping the time evolution of deformation. All experiments were carried out at the Heavy Structures Lab of the University of Leeds. During tests the mid-span deflection, creep coefficient and shrinkage strains were monitored for duration of 90 days. The obtained results are set against the values predicted by Eurocode 2 and the tools within an FE commercial package (i.e. Midas FEA) to yield that existing knowledge and practise is at times over-conservative.

Keywords: Eurocode2, midas fea, repeated, sustained loading.

Procedia PDF Downloads 347
2227 Pallet Tracking and Cost Optimization of the Flow of Goods in Logistics Operations by Serial Shipping Container Code

Authors: Dominika Crnjac Milic, Martina Martinovic, Vladimir Simovic

Abstract:

The case study method in this paper shows the implementation of Information Technology (IT) and the Serial Shipping Container Code (SSCC) in a Croatian company that deals with logistics operations and provides logistics services in the cold chain segment. This company is aware of the sensitivity of the goods entrusted to them by the user of the service, as well as of the importance of speed and accuracy in providing logistics services. To that end, it has implemented and used the latest IT to ensure the highest standard of high-quality logistics services to its customers. Looking for efficiency and optimization of supply chain management, while maintaining a high level of quality of the products that are sold, today's users of outsourced logistics services are open to the implementation of new IT products that ultimately deliver savings. By analysing the positive results and the difficulties that arise when using this technology, we aim to provide an insight into the potential of this approach of the logistics service provider.

Keywords: logistics operations, serial shipping container code, information technology, cost optimization

Procedia PDF Downloads 360
2226 Model Driven Architecture Methodologies: A Review

Authors: Arslan Murtaza

Abstract:

Model Driven Architecture (MDA) is technique presented by OMG (Object Management Group) for software development in which different models are proposed and converted them into code. The main plan is to identify task by using PIM (Platform Independent Model) and transform it into PSM (Platform Specific Model) and then converted into code. In this review paper describes some challenges and issues that are faced in MDA, type and transformation of models (e.g. CIM, PIM and PSM), and evaluation of MDA-based methodologies.

Keywords: OMG, model driven rrchitecture (MDA), computation independent model (CIM), platform independent model (PIM), platform specific model(PSM), MDA-based methodologies

Procedia PDF Downloads 458