Search results for: automatic repair
479 A Generative Pretrained Transformer-Based Question-Answer Chatbot and Phantom-Less Quantitative Computed Tomography Bone Mineral Density Measurement System for Osteoporosis
Authors: Mian Huang, Chi Ma, Junyu Lin, William Lu
Abstract:
Introduction: Bone health attracts more attention recently and an intelligent question and answer (QA) chatbot for osteoporosis is helpful for science popularization. With Generative Pretrained Transformer (GPT) technology developing, we build an osteoporosis corpus dataset and then fine-tune LLaMA, a famous open-source GPT foundation large language model(LLM), on our self-constructed osteoporosis corpus. Evaluated by clinical orthopedic experts, our fine-tuned model outperforms vanilla LLaMA on osteoporosis QA task in Chinese. Three-dimensional quantitative computed tomography (QCT) measured bone mineral density (BMD) is considered as more accurate than DXA for BMD measurement in recent years. We develop an automatic Phantom-less QCT(PL-QCT) that is more efficient for BMD measurement since no need of an external phantom for calibration. Combined with LLM on osteoporosis, our PL-QCT provides efficient and accurate BMD measurement for our chatbot users. Material and Methods: We build an osteoporosis corpus containing about 30,000 Chinese literatures whose titles are related to osteoporosis. The whole process is done automatically, including crawling literatures in .pdf format, localizing text/figure/table region by layout segmentation algorithm and recognizing text by OCR algorithm. We train our model by continuous pre-training with Low-rank Adaptation (LoRA, rank=10) technology to adapt LLaMA-7B model to osteoporosis domain, whose basic principle is to mask the next word in the text and make the model predict that word. The loss function is defined as cross-entropy between the predicted and ground-truth word. Experiment is implemented on single NVIDIA A800 GPU for 15 days. Our automatic PL-QCT BMD measurement adopt AI-associated region-of-interest (ROI) generation algorithm for localizing vertebrae-parallel cylinder in cancellous bone. Due to no phantom for BMD calibration, we calculate ROI BMD by CT-BMD of personal muscle and fat. Results & Discussion: Clinical orthopaedic experts are invited to design 5 osteoporosis questions in Chinese, evaluating performance of vanilla LLaMA and our fine-tuned model. Our model outperforms LLaMA on over 80% of these questions, understanding ‘Expert Consensus on Osteoporosis’, ‘QCT for osteoporosis diagnosis’ and ‘Effect of age on osteoporosis’. Detailed results are shown in appendix. Future work may be done by training a larger LLM on the whole orthopaedics with more high-quality domain data, or a multi-modal GPT combining and understanding X-ray and medical text for orthopaedic computer-aided-diagnosis. However, GPT model gives unexpected outputs sometimes, such as repetitive text or seemingly normal but wrong answer (called ‘hallucination’). Even though GPT give correct answers, it cannot be considered as valid clinical diagnoses instead of clinical doctors. The PL-QCT BMD system provided by Bone’s QCT(Bone’s Technology(Shenzhen) Limited) achieves 0.1448mg/cm2(spine) and 0.0002 mg/cm2(hip) mean absolute error(MAE) and linear correlation coefficient R2=0.9970(spine) and R2=0.9991(hip)(compared to QCT-Pro(Mindways)) on 155 patients in three-center clinical trial in Guangzhou, China. Conclusion: This study builds a Chinese osteoporosis corpus and develops a fine-tuned and domain-adapted LLM as well as a PL-QCT BMD measurement system. Our fine-tuned GPT model shows better capability than LLaMA model on most testing questions on osteoporosis. Combined with our PL-QCT BMD system, we are looking forward to providing science popularization and early morning screening for potential osteoporotic patients.Keywords: GPT, phantom-less QCT, large language model, osteoporosis
Procedia PDF Downloads 71478 Comparison of Breast Surface Doses for Full-Field Digital Mammography and Digital Breast Tomosynthesis Using Breast Phantoms
Authors: Chia-Hui Chen, Chien-Kuo Wang
Abstract:
Background: Full field digital mammography (FFDM) is widely used in diagnosis of breast cancer. Digital breast tomosynthesis (DBT) has recently been introduced into the clinic and is being used for screening for breast cancer in the general population. Hence, the radiation dose delivered to the patients involved in an imaging protocol is of utmost concern. Aim: To compare the surface radiation dose (ESD) of digital breast tomosynthesis (DBT) and full-field digital mammography (FFDM) by using breast phantoms. Method: We analyzed the average entrance surface dose (ESD) of FFDM and DBT by using breast phantoms. Optically Stimulated luminescent Dosimeters (OSLD) were placed in a tissue-equivalent Breast phantom at difference sites of interest. Absorbed dose measurements were obtained after digital breast tomosynthesis (DBT) and full-field digital mammography (FFDM) exposures. Results: An automatic exposure control (AEC) is proposed for surface dose measurement during DBT and FFDM. The mean ESD values for DBT and FFDM were 6.37 mGy and 3.51mGy, respectively. Using of OSLD measured for surface dose during DBT and FFDM. There were 19.87 mGy and 11.36 mGy, respectively. The surface exposure dose of DBT could possibly be increased by two times with FFDM. Conclusion: The radiation dose from DBT was higher than that of FFDM and the difference in dose between AEC and OSLD measurements at phantom surface.Keywords: full-field digital mammography, digital breast tomosynthesis, optically stimulated luminescent dosimeters, surface dose
Procedia PDF Downloads 420477 Semiautomatic Calculation of Ejection Fraction Using Echocardiographic Image Processing
Authors: Diana Pombo, Maria Loaiza, Mauricio Quijano, Alberto Cadena, Juan Pablo Tello
Abstract:
In this paper, we present a semi-automatic tool for calculating ejection fraction from an echocardiographic video signal which is derived from a database in DICOM format, of Clinica de la Costa - Barranquilla. Described in this paper are each of the steps and methods used to find the respective calculation that includes acquisition and formation of the test samples, processing and finally the calculation of the parameters to obtain the ejection fraction. Two imaging segmentation methods were compared following a methodological framework that is similar only in the initial stages of processing (process of filtering and image enhancement) and differ in the end when algorithms are implemented (Active Contour and Region Growing Algorithms). The results were compared with the measurements obtained by two different medical specialists in cardiology who calculated the ejection fraction of the study samples using the traditional method, which consists of drawing the region of interest directly from the computer using echocardiography equipment and a simple equation to calculate the desired value. The results showed that if the quality of video samples are good (i.e., after the pre-processing there is evidence of an improvement in the contrast), the values provided by the tool are substantially close to those reported by physicians; also the correlation between physicians does not vary significantly.Keywords: echocardiography, DICOM, processing, segmentation, EDV, ESV, ejection fraction
Procedia PDF Downloads 426476 Recognizing an Individual, Their Topic of Conversation and Cultural Background from 3D Body Movement
Authors: Gheida J. Shahrour, Martin J. Russell
Abstract:
The 3D body movement signals captured during human-human conversation include clues not only to the content of people’s communication but also to their culture and personality. This paper is concerned with automatic extraction of this information from body movement signals. For the purpose of this research, we collected a novel corpus from 27 subjects, arranged them into groups according to their culture. We arranged each group into pairs and each pair communicated with each other about different topics. A state-of-art recognition system is applied to the problems of person, culture, and topic recognition. We borrowed modeling, classification, and normalization techniques from speech recognition. We used Gaussian Mixture Modeling (GMM) as the main technique for building our three systems, obtaining 77.78%, 55.47%, and 39.06% from the person, culture, and topic recognition systems respectively. In addition, we combined the above GMM systems with Support Vector Machines (SVM) to obtain 85.42%, 62.50%, and 40.63% accuracy for person, culture, and topic recognition respectively. Although direct comparison among these three recognition systems is difficult, it seems that our person recognition system performs best for both GMM and GMM-SVM, suggesting that inter-subject differences (i.e. subject’s personality traits) are a major source of variation. When removing these traits from culture and topic recognition systems using the Nuisance Attribute Projection (NAP) and the Intersession Variability Compensation (ISVC) techniques, we obtained 73.44% and 46.09% accuracy from culture and topic recognition systems respectively.Keywords: person recognition, topic recognition, culture recognition, 3D body movement signals, variability compensation
Procedia PDF Downloads 541475 Photoleap: An AI-Powered Photo Editing App with Advanced Features and User Satisfaction Analysis
Authors: Joud Basyouni, Rama Zagzoog, Mashael Al Faleh, Jana Alireza
Abstract:
AI is changing many fields and speeding up tasks that used to take a long time. It used to take too long to edit photos. However, many AI-powered apps make photo editing, automatic effects, and animations much easier than other manual editing apps with no AI. The mobile app Photoleap edits photos and creates digital art using AI. Editing photos with text prompts is also becoming a standard these days with the help of apps like Photoleap. Now, users can change backgrounds, add animations, turn text into images, and create scenes with AI. This project report discusses the photo editing app's history and popularity. Photoleap resembles Photoshop, Canva, Photos, and Pixlr. The report includes survey questions to assess Photoleap user satisfaction. The report describes Photoleap's features and functions with screenshots. Photoleap uses AI well. Charts and graphs show Photoleap user ratings and comments from the survey. This project found that most Photoleap users liked how well it worked, was made, and was easy to use. People liked changing photos and adding backgrounds. Users can create stunning photo animations. A few users dislike the app's animations, AI art, and photo effects. The project report discusses the app's pros and cons and offers improvements.Keywords: artificial intelligence, photoleap, images, background, photo editing
Procedia PDF Downloads 61474 Development of an Automatic Monitoring System Based on the Open Architecture Concept
Authors: Andrii Biloshchytskyi, Serik Omirbayev, Alexandr Neftissov, Sapar Toxanov, Svitlana Biloshchytska, Adil Faizullin
Abstract:
Kazakhstan has adopted a carbon neutrality strategy until 2060. In accordance with this strategy, it is necessary to introduce various tools to maintain the environmental safety of the environment. The use of IoT, in combination with the characteristics and requirements of Kazakhstan's environmental legislation, makes it possible to develop a modern environmental monitoring system. The article proposes a solution for developing an example of an automated system for the continuous collection of data on the concentration of pollutants in the atmosphere based on an open architecture. The Audino-based device acts as a microcontroller. It should be noted that the transmission of measured values is carried out via an open wireless communication protocol. The architecture of the system, which was used to build a prototype based on sensors, an Arduino microcontroller, and a wireless data transmission module, is presented. The selection of elementary components may change depending on the requirements of the system; the introduction of new units is limited by the number of ports. The openness of solutions allows you to change the configuration depending on the conditions. The advantages of the solutions are openness, low cost, versatility and mobility. However, there is no comparison of the working processes of the proposed solution with traditional ones.Keywords: environmental monitoring, greenhouse gases emissions, environmental pollution, Industry 4.0, IoT, microcontroller, automated monitoring system.
Procedia PDF Downloads 48473 Designing a Model to Increase the Flow of Circular Economy Startups Using a Systemic and Multi-Generational Approach
Authors: Luís Marques, João Rocha, Andreia Fernandes, Maria Moura, Cláudia Caseiro, Filipa Figueiredo, João Nunes
Abstract:
The implementation of circularity strategies other than recycling, such as reducing the amount of raw material, as well as reusing or sharing existing products, remains marginal. The European Commission announced that the transition towards a more circular economy could lead to the net creation of about 700,000 jobs in Europe by 2030, through additional labour demand from recycling plants, repair services and other circular activities. Efforts to create new circular business models in accordance with completely circular processes, as opposed to linear ones, have increased considerably in recent years. In order to create a societal Circular Economy transition model, it is necessary to include innovative solutions, where startups play a key role. Early-stage startups based on new business models according to circular processes often face difficulties in creating enough impact. The StartUp Zero Program designs a model and approach to increase the flow of startups in the Circular Economy field, focusing on a systemic decision analysis and multi-generational approach, considering Multi-Criteria Decision Analysis to support a decision-making tool, which is also supported by the use of a combination of an Analytical Hierarchy Process and Multi-Attribute Value Theory methods. We define principles, criteria and indicators for evaluating startup prerogatives, quantifying the evaluation process in a unique result. Additionally, this entrepreneurship program spanning 16 months involved more than 2400 young people, from ages 14 to 23, in more than 200 interaction activities.Keywords: circular economy, entrepreneurship, startups;, multi-criteria decision analysis
Procedia PDF Downloads 105472 Domain specific Ontology-Based Knowledge Extraction Using R-GNN and Large Language Models
Authors: Andrey Khalov
Abstract:
The rapid proliferation of unstructured data in IT infrastructure management demands innovative approaches for extracting actionable knowledge. This paper presents a framework for ontology-based knowledge extraction that combines relational graph neural networks (R-GNN) with large language models (LLMs). The proposed method leverages the DOLCE framework as the foundational ontology, extending it with concepts from ITSMO for domain-specific applications in IT service management and outsourcing. A key component of this research is the use of transformer-based models, such as DeBERTa-v3-large, for automatic entity and relationship extraction from unstructured texts. Furthermore, the paper explores how transfer learning techniques can be applied to fine-tune large language models (LLaMA) for using to generate synthetic datasets to improve precision in BERT-based entity recognition and ontology alignment. The resulting IT Ontology (ITO) serves as a comprehensive knowledge base that integrates domain-specific insights from ITIL processes, enabling more efficient decision-making. Experimental results demonstrate significant improvements in knowledge extraction and relationship mapping, offering a cutting-edge solution for enhancing cognitive computing in IT service environments.Keywords: ontology mapping, R-GNN, knowledge extraction, large language models, NER, knowlege graph
Procedia PDF Downloads 16471 Night Shift Work as an Oxidative Stressor: A Systematic Review
Authors: Madeline Gibson
Abstract:
Night shift workers make up an essential part of the modern workforce. However, night shift workers have higher incidences of late in life diseases and earlier mortality. Night shift workers are exposed to constant light and experience circadian rhythm disruption. Sleep disruption is thought to increase oxidative stress, defined as an imbalance of excess pro-oxidative factors and reactive oxygen species over anti-oxidative activity. Oxidative stress can damage cells, proteins and DNA and can eventually lead to varied chronic diseases such as cancer, diabetes, cardiovascular disease, Alzheimer’s and dementia. This review aimed to understand whether night shift workers were at greater risk of oxidative stress and to contribute to a consensus on this relationship. Twelve studies published in 2001-2019 examining 2,081 workers were included in the review. Studies compared both the impact of working a single shift and in comparisons between those who regularly work night shifts and only day shifts. All studies had evidence to support this relationship across a range of oxidative stress indicators, including increased DNA damage, reduced DNA repair capacity, increased lipid peroxidation, higher levels of reactive oxygen species, and to a lesser extent, a reduction in antioxidant defense. This research supports the theory that melatonin and the sleep-wake cycle mediate the relationship between shift work and oxidative stress. It is concluded that night shift work increases the risk for oxidative stress and, therefore, future disease. Recommendations are made to promote the long-term health of shift workers considering these findings.Keywords: night shift work, coxidative stress, circadian rhythm, melatonin, disease, circadian rhythm disruption
Procedia PDF Downloads 266470 Survey of Related Field for Artificial Intelligence Window Development
Authors: Young Kwon Yang, Bo Rang Park, Hyo Eun Lee, Tea Won Kim, Eun Ji Choi, Jin Chul Park
Abstract:
To develop an artificial intelligence based automatic ventilation system, recent research trends were analyzed and analyzed. This research method is as follows. In the field of architecture and window technology, the use of artificial intelligence, the existing study of machine learning model and the theoretical review of the literature were carried out. This paper collected journals such as Journal of Energy and Buildings, Journal of Renewable and Sustainable Energy Reviews, and articles published on Web-sites. The following keywords were searched for articles from 2000 to 2016. We searched for the above keywords mainly in the title, keyword, and abstract. As a result, the global artificial intelligence market is expected to grow at a CAGR of 14.0% from USD127bn in 2015 to USD165bn in 2017. Start-up investments in artificial intelligence increased from the US $ 45 million in 2010 to the US $ 310 million in 2015, and the number of investments increased from 6 to 54. Although AI is making efforts to advance to advanced countries, the level of technology is still in its infant stage. Especially in the field of architecture, artificial intelligence (AI) is very rare. Based on the data of this study, it is expected that the application of artificial intelligence and the application of architectural field will be revitalized through the activation of artificial intelligence in the field of architecture and window.Keywords: artificial intelligence, window, fine dust, thermal comfort, ventilation system
Procedia PDF Downloads 275469 A Methodological Approach to Development of Mental Script for Mental Practice of Micro Suturing
Authors: Vaikunthan Rajaratnam
Abstract:
Intro: Motor imagery (MI) and mental practice (MP) can be an alternative to acquire mastery of surgical skills. One component of using this technique is the use of a mental script. The aim of this study was to design and develop a mental script for basic micro suturing training for skill acquisition using a low-fidelity rubber glove model and to describe the detailed methodology for this process. Methods: This study was based on a design and development research framework. The mental script was developed with 5 expert surgeons performing a cognitive walkthrough of the repair of a vertical opening in a rubber glove model using 8/0 nylon. This was followed by a hierarchal task analysis. A draft script was created, and face and content validity assessed with a checking-back process. The final script was validated with the recruitment of 28 participants, assessed using the Mental Imagery Questionnaire (MIQ). Results: The creation of the mental script is detailed in the full text. After assessment by the expert panel, the mental script had good face and content validity. The average overall MIQ score was 5.2 ± 1.1, demonstrating the validity of generating mental imagery from the mental script developed in this study for micro suturing in the rubber glove model. Conclusion: The methodological approach described in this study is based on an instructional design framework to teach surgical skills. This MP model is inexpensive and easily accessible, addressing the challenge of reduced opportunities to practice surgical skills. However, while motor skills are important, other non-technical expertise required by the surgeon is not addressed with this model. Thus, this model should act a surgical training augment, but not replace it.Keywords: mental script, motor imagery, cognitive walkthrough, verbal protocol analysis, hierarchical task analysis
Procedia PDF Downloads 103468 Estimation of Lungs Physiological Motion for Patient Undergoing External Lung Irradiation
Authors: Yousif Mohamed Y. Abdallah
Abstract:
This is an experimental study deals with detection, measurement and analysis of the periodic physiological organ motion during external beam radiotherapy; to improve the accuracy of the radiation field placement, and to reduce the exposure of healthy tissue during radiation treatments. The importance of this study is to detect the maximum path of the mobile structures during radiotherapy delivery, to define the planning target volume (PTV) and irradiated volume during both inspiration and expiration period and to verify the target volume. In addition to its role to highlight the importance of the application of Intense Guided Radiotherapy (IGRT) methods in the field of radiotherapy. The results showed (body contour was equally (3.17 + 0.23 mm), for left lung displacement reading (2.56 + 0.99 mm) and right lung is (2.42 + 0.77 mm) which the radiation oncologist to take suitable countermeasures in case of significant errors. In addition, the use of the image registration technique for automatic position control is predicted potential motion. The motion ranged between 2.13 mm and 12.2 mm (low and high). In conclusion, individualized assessment of tumor mobility can improve the accuracy of target areas definition in patients undergo Sterostatic RT for stage I, II and III lung cancer (NSCLC). Definition of the target volume based on a single CT scan with a margin of 10 mm is clearly inappropriate.Keywords: respiratory motion, external beam radiotherapy, image processing, lung
Procedia PDF Downloads 535467 Conception of a Regulated, Dynamic and Intelligent Sewerage in Ostrevent
Authors: Rabaa Tlili Yaakoubi, Hind Nakouri, Olivier Blanpain
Abstract:
The current tools for real time management of sewer systems are based on two software tools: the software of weather forecast and the software of hydraulic simulation. The use of the first ones is an important cause of imprecision and uncertainty, the use of the second requires temporal important steps of decision because of their need in times of calculation. This way of proceeding fact that the obtained results are generally different from those waited. The major idea of the CARDIO project is to change the basic paradigm by approaching the problem by the "automatic" face rather than by that "hydrology". The objective is to make possible the realization of a large number of simulations at very short times (a few seconds) allowing to take place weather forecasts by using directly the real time meditative pluviometric data. The aim is to reach a system where the decision-making is realized from reliable data and where the correction of the error is permanent. A first model of control laws was realized and tested with different return-period rainfalls. The gains obtained in rejecting volume vary from 40 to 100%. The development of a new algorithm was then used to optimize calculation time and thus to overcome the subsequent combinatorial problem in our first approach. Finally, this new algorithm was tested with 16- year-rainfall series. The obtained gains are 60% of total volume rejected to the natural environment and of 80 % in the number of discharges.Keywords: RTC, paradigm, optimization, automation
Procedia PDF Downloads 284466 Unbalanced Distribution Optimal Power Flow to Minimize Losses with Distributed Photovoltaic Plants
Authors: Malinwo Estone Ayikpa
Abstract:
Electric power systems are likely to operate with minimum losses and voltage meeting international standards. This is made possible generally by control actions provide by automatic voltage regulators, capacitors and transformers with on-load tap changer (OLTC). With the development of photovoltaic (PV) systems technology, their integration on distribution networks has increased over the last years to the extent of replacing the above mentioned techniques. The conventional analysis and simulation tools used for electrical networks are no longer able to take into account control actions necessary for studying distributed PV generation impact. This paper presents an unbalanced optimal power flow (OPF) model that minimizes losses with association of active power generation and reactive power control of single-phase and three-phase PV systems. Reactive power can be generated or absorbed using the available capacity and the adjustable power factor of the inverter. The unbalance OPF is formulated by current balance equations and solved by primal-dual interior point method. Several simulation cases have been carried out varying the size and location of PV systems and the results show a detailed view of the impact of PV distributed generation on distribution systems.Keywords: distribution system, loss, photovoltaic generation, primal-dual interior point method
Procedia PDF Downloads 332465 Applying Laser Scanning and Digital Photogrammetry for Developing an Archaeological Model Structure for Old Castle in Germany
Authors: Bara' Al-Mistarehi
Abstract:
Documentation and assessment of conservation state of an archaeological structure is a significant procedure in any management plan. However, it has always been a challenge to apply this with a low coast and safe methodology. It is also a time-demanding procedure. Therefore, a low cost, efficient methodology for documenting the state of a structure is needed. In the scope of this research, this paper will employ digital photogrammetry and laser scanner to one of highly significant structures in Germany, The Old Castle (German: Altes Schloss). The site is well known for its unique features. However, the castle suffers from serious deterioration threats because of the environmental conditions and the absence of continuous monitoring, maintenance and repair plans. Digital photogrammetry is a generally accepted technique for the collection of 3D representations of the environment. For this reason, this image-based technique has been extensively used to produce high quality 3D models of heritage sites and historical buildings for documentation and presentation purposes. Additionally, terrestrial laser scanners are used, which directly measure 3D surface coordinates based on the run-time of reflected light pulses. These systems feature high data acquisition rates, good accuracy and high spatial data density. Despite the potential of each single approach, in this research work maximum benefit is to be expected by a combination of data from both digital cameras and terrestrial laser scanners. Within the paper, the usage, application and advantages of the technique will be investigated in terms of building high realistic 3D textured model for some parts of the old castle. The model will be used as diagnosing tool of the conservation state of the castle and monitoring mean for future changes.Keywords: Digital photogrammetry, Terrestrial laser scanners, 3D textured model, archaeological structure
Procedia PDF Downloads 178464 Application of the Mobile Phone for Occupational Self-Inspection Program in Small-Scale Industries
Authors: Jia-Sin Li, Ying-Fang Wang, Cheing-Tong Yan
Abstract:
In this study, an integrated approach of Google Spreadsheet and QR code which is free internet resources was used to improve the inspection procedure. The mobile phone Application(App)was also designed to combine with a web page to create an automatic checklist in order to provide a new integrated information of inspection management system. By means of client-server model, the client App is developed for Android mobile OS and the back end is a web server. It can set up App accounts including authorized data and store some checklist documents in the website. The checklist document URL could generate QR code first and then print and paste on the machine. The user can scan the QR code by the app and filled the checklist in the factory. In the meanwhile, the checklist data will send to the server, it not only save the filled data but also executes the related functions and charts. On the other hand, it also enables auditors and supervisors to facilitate the prevention and response to hazards, as well as immediate report data checks. Finally, statistics and professional analysis are performed using inspection records and other relevant data to not only improve the reliability, integrity of inspection operations and equipment loss control, but also increase plant safety and personnel performance. Therefore, it suggested that the traditional paper-based inspection method could be replaced by the APP which promotes the promotion of industrial security and reduces human error.Keywords: checklist, Google spreadsheet, APP, self-inspection
Procedia PDF Downloads 118463 Development of Advanced Linear Calibration Technique for Air Flow Sensing by Using CTA-Based Hot Wire Anemometry
Authors: Ming-Jong Tsai, T. M. Wu, R. C. Chu
Abstract:
The purpose of this study is to develop an Advanced linear calibration Technique for air flow sensing by using CTA-based Hot wire Anemometry. It contains a host PC with Human Machine Interface, a wind tunnel, a wind speed controller, an automatic data acquisition module, and nonlinear calibration model. To improve the fitting error by using single fitting polynomial, this study proposes a Multiple three-order Polynomial Fitting Method (MPFM) for fitting the non-linear output of a CTA-based Hot wire Anemometry. The CTA-based anemometer with built-in fitting parameters is installed in the wind tunnel, and the wind speed is controlled by the PC-based controller. The Hot-Wire anemometer's thermistor resistance change is converted into a voltage signal or temperature differences, and then sent to the PC through a DAQ card. After completion measurements of original signal, the Multiple polynomial mathematical coefficients can be automatically calculated, and then sent into the micro-processor in the Hot-Wire anemometer. Finally, the corrected Hot-Wire anemometer is verified for the linearity, the repeatability, error percentage, and the system outputs quality control reports.Keywords: flow rate sensing, hot wire, constant temperature anemometry (CTA), linear calibration, multiple three-order polynomial fitting method (MPFM), temperature compensation
Procedia PDF Downloads 416462 Autonomous Vehicle Detection and Classification in High Resolution Satellite Imagery
Authors: Ali J. Ghandour, Houssam A. Krayem, Abedelkarim A. Jezzini
Abstract:
High-resolution satellite images and remote sensing can provide global information in a fast way compared to traditional methods of data collection. Under such high resolution, a road is not a thin line anymore. Objects such as cars and trees are easily identifiable. Automatic vehicles enumeration can be considered one of the most important applications in traffic management. In this paper, autonomous vehicle detection and classification approach in highway environment is proposed. This approach consists mainly of three stages: (i) first, a set of preprocessing operations are applied including soil, vegetation, water suppression. (ii) Then, road networks detection and delineation is implemented using built-up area index, followed by several morphological operations. This step plays an important role in increasing the overall detection accuracy since vehicles candidates are objects contained within the road networks only. (iii) Multi-level Otsu segmentation is implemented in the last stage, resulting in vehicle detection and classification, where detected vehicles are classified into cars and trucks. Accuracy assessment analysis is conducted over different study areas to show the great efficiency of the proposed method, especially in highway environment.Keywords: remote sensing, object identification, vehicle and road extraction, vehicle and road features-based classification
Procedia PDF Downloads 231461 Ridership Study for the Proposed Installation of Automatic Guide-way Transit (AGT) System along Sapphire Street in Balanga City, Bataan
Authors: Nelson Andres, Meeko C. Masangcap, John Denver D. Catapang
Abstract:
Balanga City as, the heart of Bataan, is a growing City and is now at its fast pace of development. The growth of commerce in the city results to an increase in commuters who travel back and forth through the city, leading to congestions. Consequently, queuing of vehicles along national roads and even in the highways of the city have become a regular occurrence. This common scenario of commuters flocking the city, private and public vehicles going bumper to bumper, especially during the rush hours, greatly affect the flow of traffic vehicles and is now a burden not only to the commuters but also to the government who is trying to address this dilemma. Seeing these terrible events, the implementation of an elevated Automated Guide-way transit is seen as a possible solution to help in the decongestion of the affected parts of Balanga City.In response to the problem, the researchers identify if it is feasible to have an elevated guide-way transit in the vicinity of Sapphire Street in Balanga City, Bataan. Specifically, the study aims to determine who will be the riders based on the demographic profile, where the trip can be generated and distributed, the time when volume of people usually peaks and the estimated volume of passengers. Statistical analysis is applied to the data gathered to find out if there is an important relationship between the demographic profile of the respondents and their preference of having an elevated railway transit in the City of Balanga.Keywords: ridership, AGT, railway, elevated track
Procedia PDF Downloads 81460 Design an Algorithm for Software Development in CBSE Envrionment Using Feed Forward Neural Network
Authors: Amit Verma, Pardeep Kaur
Abstract:
In software development organizations, Component based Software engineering (CBSE) is emerging paradigm for software development and gained wide acceptance as it often results in increase quality of software product within development time and budget. In component reusability, main challenges are the right component identification from large repositories at right time. The major objective of this work is to provide efficient algorithm for storage and effective retrieval of components using neural network and parameters based on user choice through clustering. This research paper aims to propose an algorithm that provides error free and automatic process (for retrieval of the components) while reuse of the component. In this algorithm, keywords (or components) are extracted from software document, after by applying k mean clustering algorithm. Then weights assigned to those keywords based on their frequency and after assigning weights, ANN predicts whether correct weight is assigned to keywords (or components) or not, otherwise it back propagates in to initial step (re-assign the weights). In last, store those all keywords into repositories for effective retrieval. Proposed algorithm is very effective in the error correction and detection with user base choice while choice of component for reusability for efficient retrieval is there.Keywords: component based development, clustering, back propagation algorithm, keyword based retrieval
Procedia PDF Downloads 378459 Mouse Knockouts for Elucidating the Role of Cysteine-Rich Angiogenic Inducer 61 in Tendon Development and Maintenance
Authors: Josephine Hai, Jie Jiang, Karen M. Lyons
Abstract:
Of the musculoskeletal tissues, tendon is least understood in terms of biological development. The current study examines Cysteine-rich angiogenic inducer 61, or CCN1, a member of the CCN family of secreted matricellular proteins that regulate cell behavior via intercellular signaling. Though CCN1 is notable in limiting fibrosis by inducing senescence in fibroblasts, little is known about its role in normal fibrous tissue, where it may be essential to the development of ECM-rich structures like tendons. We found that CCN1 knockout mice (using limb-specific Prx1-Cre) exhibited clubfoot and waddling gaits, a unique phenotype not described in any other mutant to date. Histological analysis showed that the Achilles and patellar tendons, where we previously found high CCN1 expression in adult reporter mice, were thicker and denser in the Prx1-Cre knockouts than in their wildtype littermates. We then hypothesized that CCN1 is required directly in tendon progenitor cells for normal tendon development and generated tendon-specific CCN1 knockout mice using Scx-Cre. We observed similar Achilles/patellar tendon morphology among the Scx-Cre and Prx1-Cre mutants, indicating that the phenotype is a direct result of CCN1’s loss in tendon. To further study phenotype onset and progression, we will histologically characterize these tendons across different developmental time-points. We will also perform RNA-seq and qPCR to analyze tenocyte gene expression and expect fibrotic marker upregulation in the Scx-Cre mutants if CCN1 is required to maintain a normal tendon phenotype. Thus, our study aims to elucidate the molecular mechanisms underlying tendon formation and maintenance. Understanding tendons at the most basic level invites novel approaches to tendon repair.Keywords: development, matricellular, musculoskeletal, tendon
Procedia PDF Downloads 181458 Direct Composite Veneers as Treatment of Anterior Teeth: Case Report
Authors: Amerah Alsalem
Abstract:
Aim: Laminate veneers are restorations which are envisioned to correct existing abnormalities, esthetic deficiencies, and discolorations. Laminate veneer restorations may be processed in two different ways: direct or indirect. Materials and methods: Direct composite laminate veneers require minimal preparation compared to indirect composite veneers, cost less and are easier to repair, so are useful in young patients. However, composites can have inherent limitations such as shrinkage, limited toughness; color instability and susceptibility to wear that reduce the lifespan of the restoration and cause postoperative complications. Every new material or method introduced to the field of dentistry aims to achieve esthetics and successful dental treatments with minimal invasiveness. Therefore, direct laminate veneer restorations have been developed for advanced esthetic problems of anterior teeth. Tooth discolorations, rotated teeth, coronal fractures, congenital or acquired malformations, diastemas, discolored restorations, palatally positioned teeth, the absence of lateral incisors, abrasions and erosions are the main indications for direct laminate veneer restorations. Result: Direct veneers, as esthetic procedures, have become treatment alternatives for patients with esthetic problems of anterior teeth in recent years. The cost, social and time factors have to be considered. Although ceramic laminate veneer restorations have some advantages like color stability and high resistance against abrasion, they have also some disadvantages, including high cost and long chair time. Moreover, they have some problems such as the necessity of an additional adhesive cement. Conclusion: Although there are still some disadvantages, especially discolorations and fragility, with the development of new composite resins, direct laminate veneer restorations can be a treatment option for patients with esthetic problems of anterior teeth, when applied judiciously with good patient hygiene motivation.Keywords: direct, veneers, composite, anterior
Procedia PDF Downloads 282457 A Machine Learning Framework Based on Biometric Measurements for Automatic Fetal Head Anomalies Diagnosis in Ultrasound Images
Authors: Hanene Sahli, Aymen Mouelhi, Marwa Hajji, Amine Ben Slama, Mounir Sayadi, Farhat Fnaiech, Radhwane Rachdi
Abstract:
Fetal abnormality is still a public health problem of interest to both mother and baby. Head defect is one of the most high-risk fetal deformities. Fetal head categorization is a sensitive task that needs a massive attention from neurological experts. In this sense, biometrical measurements can be extracted by gynecologist doctors and compared with ground truth charts to identify normal or abnormal growth. The fetal head biometric measurements such as Biparietal Diameter (BPD), Occipito-Frontal Diameter (OFD) and Head Circumference (HC) needs to be monitored, and expert should carry out its manual delineations. This work proposes a new approach to automatically compute BPD, OFD and HC based on morphological characteristics extracted from head shape. Hence, the studied data selected at the same Gestational Age (GA) from the fetal Ultrasound images (US) are classified into two categories: Normal and abnormal. The abnormal subjects include hydrocephalus, microcephaly and dolichocephaly anomalies. By the use of a support vector machines (SVM) method, this study achieved high classification for automated detection of anomalies. The proposed method is promising although it doesn't need expert interventions.Keywords: biometric measurements, fetal head malformations, machine learning methods, US images
Procedia PDF Downloads 288456 Synthetic Method of Contextual Knowledge Extraction
Authors: Olga Kononova, Sergey Lyapin
Abstract:
Global information society requirements are transparency and reliability of data, as well as ability to manage information resources independently; particularly to search, to analyze, to evaluate information, thereby obtaining new expertise. Moreover, it is satisfying the society information needs that increases the efficiency of the enterprise management and public administration. The study of structurally organized thematic and semantic contexts of different types, automatically extracted from unstructured data, is one of the important tasks for the application of information technologies in education, science, culture, governance and business. The objectives of this study are the contextual knowledge typologization, selection or creation of effective tools for extracting and analyzing contextual knowledge. Explication of various kinds and forms of the contextual knowledge involves the development and use full-text search information systems. For the implementation purposes, the authors use an e-library 'Humanitariana' services such as the contextual search, different types of queries (paragraph-oriented query, frequency-ranked query), automatic extraction of knowledge from the scientific texts. The multifunctional e-library «Humanitariana» is realized in the Internet-architecture in WWS-configuration (Web-browser / Web-server / SQL-server). Advantage of use 'Humanitariana' is in the possibility of combining the resources of several organizations. Scholars and research groups may work in a local network mode and in distributed IT environments with ability to appeal to resources of any participating organizations servers. Paper discusses some specific cases of the contextual knowledge explication with the use of the e-library services and focuses on possibilities of new types of the contextual knowledge. Experimental research base are science texts about 'e-government' and 'computer games'. An analysis of the subject-themed texts trends allowed to propose the content analysis methodology, that combines a full-text search with automatic construction of 'terminogramma' and expert analysis of the selected contexts. 'Terminogramma' is made out as a table that contains a column with a frequency-ranked list of words (nouns), as well as columns with an indication of the absolute frequency (number) and the relative frequency of occurrence of the word (in %% ppm). The analysis of 'e-government' materials showed, that the state takes a dominant position in the processes of the electronic interaction between the authorities and society in modern Russia. The media credited the main role in these processes to the government, which provided public services through specialized portals. Factor analysis revealed two factors statistically describing the used terms: human interaction (the user) and the state (government, processes organizer); interaction management (public officer, processes performer) and technology (infrastructure). Isolation of these factors will lead to changes in the model of electronic interaction between government and society. In this study, the dominant social problems and the prevalence of different categories of subjects of computer gaming in science papers from 2005 to 2015 were identified. Therefore, there is an evident identification of several types of contextual knowledge: micro context; macro context; dynamic context; thematic collection of queries (interactive contextual knowledge expanding a composition of e-library information resources); multimodal context (functional integration of iconographic and full-text resources through hybrid quasi-semantic algorithm of search). Further studies can be pursued both in terms of expanding the resource base on which they are held, and in terms of the development of appropriate tools.Keywords: contextual knowledge, contextual search, e-library services, frequency-ranked query, paragraph-oriented query, technologies of the contextual knowledge extraction
Procedia PDF Downloads 359455 Investigation of the Ductility Improvement of Replaceable Hinge Member on Different Types of Precast Concrete Frames
Authors: Ali Berk Bozan, Reşat Atalay Oyguç
Abstract:
The demand for precast reinforced concrete (RC) structures is growing, considering their certain benefits, including faster assembly, homogeneous materials, and high-quality labor. The structural integrity of precast reinforced concrete (RC) constructions is influenced by the effectiveness of the joints and connections. This paper contains an analytical study about four types of precast reinforced concrete frames, which vary according to the number of storeys and the number of bays with two different types of moment-resisting beam-to-column connection is investigated under cyclic displacement loading up to 5.6% drift rate by using ABAQUS software. The first connection type is the widely used moment-resisting connection that is defined as a wet connection in the Turkish Seismic Code (TBDY). The second connection type is known as Artificial Controllable Plastic Hinge. The goal of this connection is to defend reinforced concrete components from earthquake-related plastic deformations by keeping them in a specialized connecting section. It will be possible to repair the broken connections after the earthquake. The cyclic behavior of the four types of frames with the mechanical plastic hinge and wet connection was analytically investigated, and then comparisons and suggestions were made on period, ductility, and structural system behavior coefficient. The analytical study shows that the replaceable plastic hinge element provides a significant period increase. Especially in the case of two storeys and two bays, the change in the period was felt the most compared to other frames. The results for ductility show a significant change in the ductility of the frames with replaceable plastic hinges. For the structural system behavior coefficient, a recommendation between 3.90 and 4.52 values was made.Keywords: precast structures, replaceable plastic hinge, beam to column connections, ductility
Procedia PDF Downloads 48454 Preliminary Geophysical Assessment of Soil Contaminants around Wacot Rice Factory Argungu, North-Western Nigeria
Authors: A. I. Augie, Y. Alhassan, U. Z. Magawata
Abstract:
Geophysical investigation was carried out at wacot rice factory Argungu north-western Nigeria, using the 2D electrical resistivity method. The area falls between latitude 12˚44′23ʺN to 12˚44′50ʺN and longitude 4032′18′′E to 4032′39′′E covering a total area of about 1.85 km. Two profiles were carried out with Wenner configuration using resistivity meter (Ohmega). The data obtained from the study area were modeled using RES2DIVN software which gave an automatic interpretation of the apparent resistivity data. The inverse resistivity models of the profiles show the high resistivity values ranging from 208 Ωm to 651 Ωm. These high resistivity values in the overburden were due to dryness and compactness of the strata that lead to consolidation, which is an indication that the area is free from leachate contaminations. However, from the inverse model, there are regions of low resistivity values (1 Ωm to 18 Ωm), these zones were observed and identified as clayey and the most contaminated zones. The regions of low resistivity thereby indicated the leachate plume or the highly leachate concentrated zones due to similar resistivity values in both clayey and leachate. The regions of leachate are mainly from the factory into the surrounding area and its groundwater. The maximum leachate infiltration was found at depths 1 m to 15.9 m (P1) and 6 m to 15.9 m (P2) vertically, as well as distance along the profiles from 67 m to 75 m (P1), 155 m to 180 m (P1), and 115 m to 192 m (P2) laterally.Keywords: contaminant, leachate, soil, groundwater, electrical, resistivity
Procedia PDF Downloads 160453 Long-Term Subcentimeter-Accuracy Landslide Monitoring Using a Cost-Effective Global Navigation Satellite System Rover Network: Case Study
Authors: Vincent Schlageter, Maroua Mestiri, Florian Denzinger, Hugo Raetzo, Michel Demierre
Abstract:
Precise landslide monitoring with differential global navigation satellite system (GNSS) is well known, but technical or economic reasons limit its application by geotechnical companies. This study demonstrates the reliability and the usefulness of Geomon (Infrasurvey Sàrl, Switzerland), a stand-alone and cost-effective rover network. The system permits deploying up to 15 rovers, plus one reference station for differential GNSS. A dedicated radio communication links all the modules to a base station, where an embedded computer automatically provides all the relative positions (L1 phase, open-source RTKLib software) and populates an Internet server. Each measure also contains information from an internal inclinometer, battery level, and position quality indices. Contrary to standard GNSS survey systems, which suffer from a limited number of beacons that must be placed in areas with good GSM signal, Geomon offers greater flexibility and permits a real overview of the whole landslide with good spatial resolution. Each module is powered with solar panels, ensuring autonomous long-term recordings. In this study, we have tested the system on several sites in the Swiss mountains, setting up to 7 rovers per site, for an 18 month-long survey. The aim was to assess the robustness and the accuracy of the system in different environmental conditions. In one case, we ran forced blind tests (vertical movements of a given amplitude) and compared various session parameters (duration from 10 to 90 minutes). Then the other cases were a survey of real landslides sites using fixed optimized parameters. Sub centimetric-accuracy with few outliers was obtained using the best parameters (session duration of 60 minutes, baseline 1 km or less), with the noise level on the horizontal component half that of the vertical one. The performance (percent of aborting solutions, outliers) was reduced with sessions shorter than 30 minutes. The environment also had a strong influence on the percent of aborting solutions (ambiguity search problem), due to multiple reflections or satellites obstructed by trees and mountains. The length of the baseline (distance reference-rover, single baseline processing) reduced the accuracy above 1 km but had no significant effect below this limit. In critical weather conditions, the system’s robustness was limited: snow, avalanche, and frost-covered some rovers, including the antenna and vertically oriented solar panels, leading to data interruption; and strong wind damaged a reference station. The possibility of changing the sessions’ parameters remotely was very useful. In conclusion, the rover network tested provided the foreseen sub-centimetric-accuracy while providing a dense spatial resolution landslide survey. The ease of implementation and the fully automatic long-term survey were timesaving. Performance strongly depends on surrounding conditions, but short pre-measures should allow moving a rover to a better final placement. The system offers a promising hazard mitigation technique. Improvements could include data post-processing for alerts and automatic modification of the duration and numbers of sessions based on battery level and rover displacement velocity.Keywords: GNSS, GSM, landslide, long-term, network, solar, spatial resolution, sub-centimeter.
Procedia PDF Downloads 111452 Short Answer Grading Using Multi-Context Features
Authors: S. Sharan Sundar, Nithish B. Moudhgalya, Nidhi Bhandari, Vineeth Vijayaraghavan
Abstract:
Automatic Short Answer Grading is one of the prime applications of artificial intelligence in education. Several approaches involving the utilization of selective handcrafted features, graphical matching techniques, concept identification and mapping, complex deep frameworks, sentence embeddings, etc. have been explored over the years. However, keeping in mind the real-world application of the task, these solutions present a slight overhead in terms of computations and resources in achieving high performances. In this work, a simple and effective solution making use of elemental features based on statistical, linguistic properties, and word-based similarity measures in conjunction with tree-based classifiers and regressors is proposed. The results for classification tasks show improvements ranging from 1%-30%, while the regression task shows a stark improvement of 35%. The authors attribute these improvements to the addition of multiple similarity scores to provide ensemble of scoring criteria to the models. The authors also believe the work could reinstate that classical natural language processing techniques and simple machine learning models can be used to achieve high results for short answer grading.Keywords: artificial intelligence, intelligent systems, natural language processing, text mining
Procedia PDF Downloads 133451 Application of the Total Least Squares Estimation Method for an Aircraft Aerodynamic Model Identification
Authors: Zaouche Mohamed, Amini Mohamed, Foughali Khaled, Aitkaid Souhila, Bouchiha Nihad Sarah
Abstract:
The aerodynamic coefficients are important in the evaluation of an aircraft performance and stability-control characteristics. These coefficients also can be used in the automatic flight control systems and mathematical model of flight simulator. The study of the aerodynamic aspect of flying systems is a reserved domain and inaccessible for the developers. Doing tests in a wind tunnel to extract aerodynamic forces and moments requires a specific and expensive means. Besides, the glaring lack of published documentation in this field of study makes the aerodynamic coefficients determination complicated. This work is devoted to the identification of an aerodynamic model, by using an aircraft in virtual simulated environment. We deal with the identification of the system, we present an environment framework based on Software In the Loop (SIL) methodology and we use MicrosoftTM Flight Simulator (FS-2004) as the environment for plane simulation. We propose The Total Least Squares Estimation technique (TLSE) to identify the aerodynamic parameters, which are unknown, variable, classified and used in the expression of the piloting law. In this paper, we define each aerodynamic coefficient as the mean of its numerical values. All other variations are considered as modeling uncertainties that will be compensated by the robustness of the piloting control.Keywords: aircraft aerodynamic model, total least squares estimation, piloting the aircraft, robust control, Microsoft Flight Simulator, MQ-1 predator
Procedia PDF Downloads 287450 Diagnosis and Analysis of Automated Liver and Tumor Segmentation on CT
Authors: R. R. Ramsheeja, R. Sreeraj
Abstract:
For view the internal structures of the human body such as liver, brain, kidney etc have a wide range of different modalities for medical images are provided nowadays. Computer Tomography is one of the most significant medical image modalities. In this paper use CT liver images for study the use of automatic computer aided techniques to calculate the volume of the liver tumor. Segmentation method is used for the detection of tumor from the CT scan is proposed. Gaussian filter is used for denoising the liver image and Adaptive Thresholding algorithm is used for segmentation. Multiple Region Of Interest(ROI) based method that may help to characteristic the feature different. It provides a significant impact on classification performance. Due to the characteristic of liver tumor lesion, inherent difficulties appear selective. For a better performance, a novel proposed system is introduced. Multiple ROI based feature selection and classification are performed. In order to obtain of relevant features for Support Vector Machine(SVM) classifier is important for better generalization performance. The proposed system helps to improve the better classification performance, reason in which we can see a significant reduction of features is used. The diagnosis of liver cancer from the computer tomography images is very difficult in nature. Early detection of liver tumor is very helpful to save the human life.Keywords: computed tomography (CT), multiple region of interest(ROI), feature values, segmentation, SVM classification
Procedia PDF Downloads 509