Search results for: accuracy improvement
7214 Artificial Intelligence-Based Detection of Individuals Suffering from Vestibular Disorder
Authors: Dua Hişam, Serhat İkizoğlu
Abstract:
Identifying the problem behind balance disorder is one of the most interesting topics in the medical literature. This study has considerably enhanced the development of artificial intelligence (AI) algorithms applying multiple machine learning (ML) models to sensory data on gait collected from humans to classify between normal people and those suffering from Vestibular System (VS) problems. Although AI is widely utilized as a diagnostic tool in medicine, AI models have not been used to perform feature extraction and identify VS disorders through training on raw data. In this study, three machine learning (ML) models, the Random Forest Classifier (RF), Extreme Gradient Boosting (XGB), and K-Nearest Neighbor (KNN), have been trained to detect VS disorder, and the performance comparison of the algorithms has been made using accuracy, recall, precision, and f1-score. With an accuracy of 95.28 %, Random Forest Classifier (RF) was the most accurate model.Keywords: vestibular disorder, machine learning, random forest classifier, k-nearest neighbor, extreme gradient boosting
Procedia PDF Downloads 707213 Implementation of a Low-Cost Instrumentation for an Open Cycle Wind Tunnel to Evaluate Pressure Coefficient
Authors: Cristian P. Topa, Esteban A. Valencia, Victor H. Hidalgo, Marco A. Martinez
Abstract:
Wind tunnel experiments for aerodynamic profiles display numerous advantages, such as: clean steady laminar flow, controlled environmental conditions, streamlines visualization, and real data acquisition. However, the experiment instrumentation usually is expensive, and hence, each test implies a incremented in design cost. The aim of this work is to select and implement a low-cost static pressure data acquisition system for a NACA 2412 airfoil in an open cycle wind tunnel. This work compares wind tunnel experiment with Computational Fluid Dynamics (CFD) simulation and parametric analysis. The experiment was evaluated at Reynolds of 1.65 e5, with increasing angles from -5° to 15°. The comparison between the approaches show good enough accuracy, between the experiment and CFD, additional parametric analysis results differ widely from the other methods, which complies with the lack of accuracy of the lateral approach due its simplicity.Keywords: wind tunnel, low cost instrumentation, experimental testing, CFD simulation
Procedia PDF Downloads 1827212 Impacts of Community Forest on Forest Resources Management and Livelihood Improvement of Local People in Nepal
Authors: Samipraj Mishra
Abstract:
Despite the successful implementation of community forestry program, a number of pros and cons have been raised on Terai community forestry in the case of lowland locally called Terai region of Nepal, which is climatically belongs to tropical humid and possessed high quality forests in terms of ecology and economy. The study aims to investigate the local pricing strategy of forest products and its impacts on equitable forest benefit sharing, collection of community fund and carrying out livelihood improvement activities. The study was carried out on six community forests revealed that local people have substantially benefited from the community forests. However, being the region is heterogeneous by socio-economic conditions and forest resources have higher economical potential, the decision of low pricing strategy made by the local people have created inequality problems while sharing the forest benefits, and poorly contributed to community fund collection and consequently carrying out limited activities of livelihood improvement. The paper argued that the decision of low pricing strategy of forest products is counter-productive to promote the equitable benefit sharing in the areas of heterogeneous socio-economic conditions with high value forests. The low pricing strategy has been increasing accessibility of better off households at higher rate than poor; as such households always have higher affording capacity. It is also defective to increase the community fund and carry out activities of livelihood improvement effectively. The study concluded that unilateral decentralized forest policy and decision-making autonomy to the local people seems questionable unless their decision-making capacities are enriched sufficiently. Therefore, it is recommended that empowerment of decision-making capacity of local people and their respective institutions together with policy and program formulation are prerequisite for efficient and equitable community forest management and its long-term sustainability.Keywords: community forest, livelihood, socio-economy, pricing system, Nepal
Procedia PDF Downloads 2727211 Mentoring of Health Professionals to Ensure Better Child-Birth and Newborn Care in Bihar, India: An Intervention Study
Authors: Aboli Gore, Aritra Das, Sunil Sonthalia, Tanmay Mahapatra, Sridhar Srikantiah, Hemant Shah
Abstract:
AMANAT is an initiative, taken in collaboration with the Government of Bihar, aimed at improving the Quality of Maternal and Neonatal care services at Bihar’s public health facilities – those offering either the Basic Emergency Obstetric and Neonatal care (BEmONC) or Comprehensive Emergency Obstetric and Neonatal care (CEmONC) services. The effectiveness of this program is evaluated by conducting cross-sectional assessments at the concerned facilities prior to (baseline) and following completion (endline) of intervention. Direct Observation of Delivery (DOD) methodology is employed for carrying out the baseline and endline assessments – through which key obstetric and neonatal care practices among the Health Care Providers (especially the nurses) are assessed quantitatively by specially trained nursing professionals. Assessment of vitals prior to delivery improved during all three phases of BEmONC and all four phases of CEmONC training with statistically significant improvement noted in: i) pulse measurement in BEmONC phase 2 (9% to 68%), 3 (4% to 57%) & 4 (14% to 59%) and CEmONC phase 2 (7% to 72%) and 3 (0% to 64%); ii) blood pressure measurement in BEmONC phase 2 (27% to 84%), 3 (21% to 76%) & 4 (36% to 71%) and CEmONC phase 2 (23% to 76%) and 3 (2% to 70%); iii) fetal heart rate measurement in BEmONC phase 2 (10% to 72%), 3 (11% to 77%) & 4 (13% to 64%) and CEmONC phase 1 (24% to 38%), 2 (14% to 82%) and 3 (1% to 73%); and iv) abdominal examination in BEmONC phase 2 (14% to 59%), 3 (3% to 59%) & 4 (6% to 56%) and CEmONC phase 1 (0% to 24%), 2 (7% to 62%) & 3 (0% to 62%). Regarding infection control, wearing of apron, mask and cap by the delivery conductors improved significantly in all BEmONC phases. Similarly, the practice of handwashing improved in all BEmONC and CEmONC phases. Even on disaggregation, the handwashing showed significant improvement in all phases but CEmONC phase-4. Not only the positive practices related to handwashing improved but also negative practices such as turning off the tap with bare hands declined significantly in the aforementioned phases. Significant decline was also noted in negative maternal care practices such as application of fundal pressure for hastening the delivery process and administration of oxytocin prior to delivery. One of the notable achievement of AMANAT is an improvement in active management of the third stage of labor (AMTSL). The overall AMTSL (including administration of oxytocin or other uterotonics uterotonic in proper dose, route and time along with controlled cord traction and uterine massage) improved in all phases of BEmONC and CEmONC mentoring. Another key area of improvement, across phases, was in proper cutting/clamping of the umbilical cord. AMANAT mentoring also led to improvement in important immediate newborn care practices such as initiation of skin-to-skin care and timely initiation of breastfeeding. The next phase of the mentoring program seeks to institutionalize mentoring across the state that could potentially perpetuate improvement with minimal external intervention.Keywords: capacity building, nurse-mentoring, quality of care, pregnancy, newborn care
Procedia PDF Downloads 1627210 Optimizing Machine Learning Algorithms for Defect Characterization and Elimination in Liquids Manufacturing
Authors: Tolulope Aremu
Abstract:
The key process steps to produce liquid detergent products will introduce potential defects, such as formulation, mixing, filling, and packaging, which might compromise product quality, consumer safety, and operational efficiency. Real-time identification and characterization of such defects are of prime importance for maintaining high standards and reducing waste and costs. Usually, defect detection is performed by human inspection or rule-based systems, which is very time-consuming, inconsistent, and error-prone. The present study overcomes these limitations in dealing with optimization in defect characterization within the process for making liquid detergents using Machine Learning algorithms. Performance testing of various machine learning models was carried out: Support Vector Machine, Decision Trees, Random Forest, and Convolutional Neural Network on defect detection and classification of those defects like wrong viscosity, color deviations, improper filling of a bottle, packaging anomalies. These algorithms have significantly benefited from a variety of optimization techniques, including hyperparameter tuning and ensemble learning, in order to greatly improve detection accuracy while minimizing false positives. Equipped with a rich dataset of defect types and production parameters consisting of more than 100,000 samples, our study further includes information from real-time sensor data, imaging technologies, and historic production records. The results are that optimized machine learning models significantly improve defect detection compared to traditional methods. Take, for instance, the CNNs, which run at 98% and 96% accuracy in detecting packaging anomaly detection and bottle filling inconsistency, respectively, by fine-tuning the model with real-time imaging data, through which there was a reduction in false positives of about 30%. The optimized SVM model on detecting formulation defects gave 94% in viscosity variation detection and color variation. These values of performance metrics correspond to a giant leap in defect detection accuracy compared to the usual 80% level achieved up to now by rule-based systems. Moreover, this optimization with models can hasten defect characterization, allowing for detection time to be below 15 seconds from an average of 3 minutes using manual inspections with real-time processing of data. With this, the reduction in time will be combined with a 25% reduction in production downtime because of proactive defect identification, which can save millions annually in recall and rework costs. Integrating real-time machine learning-driven monitoring drives predictive maintenance and corrective measures for a 20% improvement in overall production efficiency. Therefore, the optimization of machine learning algorithms in defect characterization optimum scalability and efficiency for liquid detergent companies gives improved operational performance to higher levels of product quality. In general, this method could be conducted in several industries within the Fast moving consumer Goods industry, which would lead to an improved quality control process.Keywords: liquid detergent manufacturing, defect detection, machine learning, support vector machines, convolutional neural networks, defect characterization, predictive maintenance, quality control, fast-moving consumer goods
Procedia PDF Downloads 217209 Effects of Listening to Pleasant Thai Classical Music on Increasing Working Memory in Elderly: An Electroencephalogram Study
Authors: Anchana Julsiri, Seree Chadcham
Abstract:
The present study determined the effects of listening to pleasant Thai classical music on increasing working memory in elderly. Thai classical music without lyrics that made participants feel fun and aroused was used in the experiment for 3.19-5.40 minutes. The accuracy scores of Counting Span Task (CST), upper alpha ERD%, and theta ERS% were used to assess working memory of participants both before and after listening to pleasant Thai classical music. The results showed that the accuracy scores of CST and upper alpha ERD% in the frontal area of participants after listening to Thai classical music were significantly higher than before listening to Thai classical music (p < .05). Theta ERS% in the fronto-parietal network of participants after listening to Thai classical music was significantly lower than before listening to Thai classical music (p < .05).Keywords: brain wave, elderly, pleasant Thai classical music, working memory
Procedia PDF Downloads 4597208 Critical Evaluation of Key Performance Indicators in Procurement Management Information System: In Case of Bangladesh
Authors: Qazi Mahdia Ghyas
Abstract:
Electronic Government Procurement (e-GP) has implemented in Bangladesh to ensure the good Governance. e-GP has transformed Bangladesh's procurement process electronically. But, to our best knowledge, there is no study to understand the key features of e-GP in Bangladesh. So, this study tries to identify the features of performance improvement after implementing an e-GP system that will help for further improvements. Data was collected from the PROMIS Overall Report (Central Procurement Technical Unit website) for the financial year from Q1 _July- Sep 2015-16 to Q4 _Apr- Jun 2021-22. This study did component factor analysis on KPIs and found nineteen KPIs that are statistically significant and represent time savings, efficiency, accountability, anti-corruption and compliance key features in procurement activities of e-GP. Based on the analysis, some practical measures have been recommended for better improvement of e-GP. This study has some limitations. Because of having multicollinearity issues, all the 42 KPIs (except 19) did not show a good fit for component factor analysis.Keywords: public procurement, electronic government procurement, KPI, performance evaluation
Procedia PDF Downloads 997207 The Sequential Estimation of the Seismoacoustic Source Energy in C-OTDR Monitoring Systems
Authors: Andrey V. Timofeev, Dmitry V. Egorov
Abstract:
The practical efficient approach is suggested for estimation of the seismoacoustic sources energy in C-OTDR monitoring systems. This approach represents the sequential plan for confidence estimation both the seismoacoustic sources energy, as well the absorption coefficient of the soil. The sequential plan delivers the non-asymptotic guaranteed accuracy of obtained estimates in the form of non-asymptotic confidence regions with prescribed sizes. These confidence regions are valid for a finite sample size when the distributions of the observations are unknown. Thus, suggested estimates are non-asymptotic and nonparametric, and also these estimates guarantee the prescribed estimation accuracy in the form of the prior prescribed size of confidence regions, and prescribed confidence coefficient value.Keywords: nonparametric estimation, sequential confidence estimation, multichannel monitoring systems, C-OTDR-system, non-lineary regression
Procedia PDF Downloads 3577206 [Keynote Speech]: Feature Selection and Predictive Modeling of Housing Data Using Random Forest
Authors: Bharatendra Rai
Abstract:
Predictive data analysis and modeling involving machine learning techniques become challenging in presence of too many explanatory variables or features. Presence of too many features in machine learning is known to not only cause algorithms to slow down, but they can also lead to decrease in model prediction accuracy. This study involves housing dataset with 79 quantitative and qualitative features that describe various aspects people consider while buying a new house. Boruta algorithm that supports feature selection using a wrapper approach build around random forest is used in this study. This feature selection process leads to 49 confirmed features which are then used for developing predictive random forest models. The study also explores five different data partitioning ratios and their impact on model accuracy are captured using coefficient of determination (r-square) and root mean square error (rsme).Keywords: housing data, feature selection, random forest, Boruta algorithm, root mean square error
Procedia PDF Downloads 3247205 Diversity of Microbial Ground Improvements
Authors: V. Ivanov, J. Chu, V. Stabnikov
Abstract:
Low cost, sustainable, and environmentally friendly microbial cements, grouts, polysaccharides and bioplastics are useful in construction and geotechnical engineering. Construction-related biotechnologies are based on activity of different microorganisms: urease-producing, acidogenic, halophilic, alkaliphilic, denitrifying, iron- and sulphate-reducing bacteria, cyanobacteria, algae, microscopic fungi. The bio-related materials and processes can be used for the bioaggregation, soil biogrouting and bioclogging, biocementation, biodesaturation of water-satured soil, bioencapsulation of soft clay, biocoating, and biorepair of the concrete surface. Altogether with the most popular calcium- and urea based biocementation, there are possible and often are more effective such methods of ground improvement as calcium- and magnesium based biocementation, calcium phosphate strengthening of soil, calcium bicarbonate biocementation, and iron- or polysaccharide based bioclogging. The construction-related microbial biotechnologies have a lot of advantages over conventional construction materials and processes.Keywords: ground improvement, biocementation, biogrouting, microorganisms
Procedia PDF Downloads 2297204 Design and Performance of a Large Diameter Shaft in Old Alluvium
Authors: Tamilmani Thiruvengadam, Ramasthanan Arulampalam
Abstract:
This project comprises laying approximately 1.8km of 400mm, 1200mm and 2400mm diameter sewer pipes using pipe jacking machines along Mugliston Park, Buangkok Drive, and Buangkok Link. The works include an estimated 14 circular shafts with depth ranging from 10.0 meters to 29.0 meters. Cast in-situ circular shaft will be used for the temporary shaft excavation. The geology is predominantly Backfill and old alluvium with weak material encountered in between. Where there is a very soft clay, F1 material or weak soil is expected, ground improvement will be carried out outside of the shaft followed by cast in-situ concrete ring wall within the improved soil zone. This paper presents the design methodology, analysis and results of temporary shafts for micro TBM launching and constructing permanent manholes. There is also a comparison of instrumentation readings with the analysis predicted values.Keywords: circular shaft, ground improvement, old alluvium, temporary shaft
Procedia PDF Downloads 2897203 Research on Reservoir Lithology Prediction Based on Residual Neural Network and Squeeze-and- Excitation Neural Network
Authors: Li Kewen, Su Zhaoxin, Wang Xingmou, Zhu Jian Bing
Abstract:
Conventional reservoir prediction methods ar not sufficient to explore the implicit relation between seismic attributes, and thus data utilization is low. In order to improve the predictive classification accuracy of reservoir lithology, this paper proposes a deep learning lithology prediction method based on ResNet (Residual Neural Network) and SENet (Squeeze-and-Excitation Neural Network). The neural network model is built and trained by using seismic attribute data and lithology data of Shengli oilfield, and the nonlinear mapping relationship between seismic attribute and lithology marker is established. The experimental results show that this method can significantly improve the classification effect of reservoir lithology, and the classification accuracy is close to 70%. This study can effectively predict the lithology of undrilled area and provide support for exploration and development.Keywords: convolutional neural network, lithology, prediction of reservoir, seismic attributes
Procedia PDF Downloads 1787202 A Practice Model for Quality Improvement in Concrete Block Mini Plants Based on Merapi Volcanic Sand
Authors: Setya Winarno
Abstract:
Due to abundant Merapi volcanic sand in Yogyakarta City, many local people have utilized it for mass production of concrete blocks through mini plants although their products are low in quality. This paper presents a practice model for quality improvement in this situation in order to supply the current customer interest in good quality of construction material. The method of this research was to investigate a techno economic evaluation through laboratory test and interview. Samples of twenty existing concrete blocks made by local people had only 19.4 kg/cm2 in average compression strength which was lower than the minimum Indonesian standard of 25 kg/cm2. Through repeat testing in laboratory for fulfilling the standard, the concrete mix design of water cement ratio should not be more than 0.64 by weight basis. The proportion of sand as aggregate content should not be more than 9 parts to 1 part by volume of Portland cement. Considering the production cost, the basic price was Rp 1,820 for each concrete block, comparing to Rp 2,000 as a normal competitive market price. At last, the model describes (a) maximum water cement ratio is 0.64, (b) maximum proportion of sand and cement is 1:9, (c) the basic price is about Rp. 1,820.00 and (d) strategies to win the competitive market on mass production of concrete blocks are focus in quality, building relationships with consumer, rapid respond to customer need, continuous innovation by product diversification, promotion in social media, and strict financial management.Keywords: concrete block, good quality, improvement model, diversification
Procedia PDF Downloads 5167201 A Chinese Nested Named Entity Recognition Model Based on Lexical Features
Abstract:
In the field of named entity recognition, most of the research has been conducted around simple entities. However, for nested named entities, which still contain entities within entities, it has been difficult to identify them accurately due to their boundary ambiguity. In this paper, a hierarchical recognition model is constructed based on the grammatical structure and semantic features of Chinese text for boundary calculation based on lexical features. The analysis is carried out at different levels in terms of granularity, semantics, and lexicality, respectively, avoiding repetitive work to reduce computational effort and using the semantic features of words to calculate the boundaries of entities to improve the accuracy of the recognition work. The results of the experiments carried out on web-based microblogging data show that the model achieves an accuracy of 86.33% and an F1 value of 89.27% in recognizing nested named entities, making up for the shortcomings of some previous recognition models and improving the efficiency of recognition of nested named entities.Keywords: coarse-grained, nested named entity, Chinese natural language processing, word embedding, T-SNE dimensionality reduction algorithm
Procedia PDF Downloads 1307200 Speaker Identification by Atomic Decomposition of Learned Features Using Computational Auditory Scene Analysis Principals in Noisy Environments
Authors: Thomas Bryan, Veton Kepuska, Ivica Kostanic
Abstract:
Speaker recognition is performed in high Additive White Gaussian Noise (AWGN) environments using principals of Computational Auditory Scene Analysis (CASA). CASA methods often classify sounds from images in the time-frequency (T-F) plane using spectrograms or cochleargrams as the image. In this paper atomic decomposition implemented by matching pursuit performs a transform from time series speech signals to the T-F plane. The atomic decomposition creates a sparsely populated T-F vector in “weight space” where each populated T-F position contains an amplitude weight. The weight space vector along with the atomic dictionary represents a denoised, compressed version of the original signal. The arraignment or of the atomic indices in the T-F vector are used for classification. Unsupervised feature learning implemented by a sparse autoencoder learns a single dictionary of basis features from a collection of envelope samples from all speakers. The approach is demonstrated using pairs of speakers from the TIMIT data set. Pairs of speakers are selected randomly from a single district. Each speak has 10 sentences. Two are used for training and 8 for testing. Atomic index probabilities are created for each training sentence and also for each test sentence. Classification is performed by finding the lowest Euclidean distance between then probabilities from the training sentences and the test sentences. Training is done at a 30dB Signal-to-Noise Ratio (SNR). Testing is performed at SNR’s of 0 dB, 5 dB, 10 dB and 30dB. The algorithm has a baseline classification accuracy of ~93% averaged over 10 pairs of speakers from the TIMIT data set. The baseline accuracy is attributable to short sequences of training and test data as well as the overall simplicity of the classification algorithm. The accuracy is not affected by AWGN and produces ~93% accuracy at 0dB SNR.Keywords: time-frequency plane, atomic decomposition, envelope sampling, Gabor atoms, matching pursuit, sparse dictionary learning, sparse autoencoder
Procedia PDF Downloads 2907199 E-Government Websites Accessibility for People with Disabilities (PWD): In Depth Evaluation of Kingdom of Bahrain
Authors: Reem AlKabbi, Hayat Ali, Mariam Yasser
Abstract:
Nowadays, eGovernment websites are becoming indispensable for public, business, personal efficiency or even improvement of livelihoods. Using these websites, citizens undertake number of tasks that would otherwise be difficult or impossible. However, many of these websites are not accessible to all people' types including People with Disabilities (PWDs). Through Web Accessibility Guidelines, Web developers can develop Web applications that are accessible to PWDs. This research is to investigate the Accessibility of eGovernment websites in Kingdom of Bahrain. The accessibility was measured using Web Content Accessibility Guidelines (WCAG) and section 508. For the evaluation purpose, some automatic tools were used. Samples of 43 eGovernment websites were selected. The accessibility of the websites was analyzed by using several automatic evaluation tools such as Total Validator and Functional Accessibility Evaluator (FAE). The evaluation process revealed several errors according to the accessibility guidelines. This research provides few recommendations for further improvement of accessibility features of eGovernment websites based on the highlighted issues and key findings reported in this research.Keywords: website accessibility, W3C, PWD, e-government
Procedia PDF Downloads 4517198 Temperature Control Improvement of Membrane Reactor
Authors: Pornsiri Kaewpradit, Chalisa Pourneaw
Abstract:
Temperature control improvement of a membrane reactor with exothermic and reversible esterification reaction is studied in this work. It is well known that a batch membrane reactor requires different control strategies from a continuous one due to the fact that it is operated dynamically. Due to the effect of the operating temperature, the suitable control scheme has to be designed based reliable predictive model to achieve a desired objective. In the study, the optimization framework has been preliminary formulated in order to determine an optimal temperature trajectory for maximizing a desired product. In model predictive control scheme, a set of predictive models have been initially developed corresponding to the possible operating points of the system. The multiple predictive control moves have been further calculated on-line using the developed models corresponding to current operating point. It is obviously seen in the simulation results that the temperature control has been improved compared to the performance obtained by the conventional predictive controller. Further robustness tests have also been investigated in this study.Keywords: model predictive control, batch reactor, temperature control, membrane reactor
Procedia PDF Downloads 4697197 Modelling of Geotechnical Data Using Geographic Information System and MATLAB for Eastern Ahmedabad City, Gujarat
Authors: Rahul Patel
Abstract:
Ahmedabad, a city located in western India, is experiencing rapid growth due to urbanization and industrialization. It is projected to become a metropolitan city in the near future, resulting in various construction activities. Soil testing is necessary before construction can commence, requiring construction companies and contractors to periodically conduct soil testing. The focus of this study is on the process of creating a spatial database that is digitally formatted and integrated with geotechnical data and a Geographic Information System (GIS). Building a comprehensive geotechnical (Geo)-database involves three steps: collecting borehole data from reputable sources, verifying the accuracy and redundancy of the data, and standardizing and organizing the geotechnical information for integration into the database. Once the database is complete, it is integrated with GIS, allowing users to visualize, analyze, and interpret geotechnical information spatially. Using a Topographic to Raster interpolation process in GIS, estimated values are assigned to all locations based on sampled geotechnical data values. The study area was contoured for SPT N-Values, Soil Classification, Φ-Values, and Bearing Capacity (T/m2). Various interpolation techniques were cross-validated to ensure information accuracy. This GIS map enables the calculation of SPT N-Values, Φ-Values, and bearing capacities for different footing widths and various depths. This study highlights the potential of GIS in providing an efficient solution to complex phenomena that would otherwise be tedious to achieve through other means. Not only does GIS offer greater accuracy, but it also generates valuable information that can be used as input for correlation analysis. Furthermore, this system serves as a decision support tool for geotechnical engineers.Keywords: ArcGIS, borehole data, geographic information system, geo-database, interpolation, SPT N-value, soil classification, Φ-Value, bearing capacity
Procedia PDF Downloads 747196 Park Improvements in a High-Density City: Ecological, Social and Economy Concerns
Authors: Yuchen Niu, Liang Zhao, Fangfang Xie, Weiyu Liu
Abstract:
In the past decades, rapid urbanization in China has significantly promoted economic growth and caused a large number of environmental problems. In consideration of land resources shortage, high-density cities will become a common phenomenon in the future. How to improve the living environment under high density is a new challenge. Shenzhen is a typical high-density city, but also the forefront of China's development and reform area. This study selects 9 urban parks with different natural attributes in Shenzhen and explores the relationship of natural, economic, and social conditions within the service scope. Based on correlation analysis and system analysis, the results indicate that improvement of park design and management methods contribute to obtaining higher ecological value and promote economic and social development.Keywords: correlation analysis, high-density city, park improvement, urban green spaces
Procedia PDF Downloads 1307195 Tree Species Classification Using Effective Features of Polarimetric SAR and Hyperspectral Images
Authors: Milad Vahidi, Mahmod R. Sahebi, Mehrnoosh Omati, Reza Mohammadi
Abstract:
Forest management organizations need information to perform their work effectively. Remote sensing is an effective method to acquire information from the Earth. Two datasets of remote sensing images were used to classify forested regions. Firstly, all of extractable features from hyperspectral and PolSAR images were extracted. The optical features were spectral indexes related to the chemical, water contents, structural indexes, effective bands and absorption features. Also, PolSAR features were the original data, target decomposition components, and SAR discriminators features. Secondly, the particle swarm optimization (PSO) and the genetic algorithms (GA) were applied to select optimization features. Furthermore, the support vector machine (SVM) classifier was used to classify the image. The results showed that the combination of PSO and SVM had higher overall accuracy than the other cases. This combination provided overall accuracy about 90.56%. The effective features were the spectral index, the bands in shortwave infrared (SWIR) and the visible ranges and certain PolSAR features.Keywords: hyperspectral, PolSAR, feature selection, SVM
Procedia PDF Downloads 4197194 Simulation Modeling and Analysis of In-Plant Logistics at a Cement Manufacturing Plant in India
Authors: Sachin Kamble, Shradha Gawankar
Abstract:
This paper presents the findings of successful implementation of Business Process Reengineering (BPR) of cement dispatch activities in a cement manufacturing plant located in India. Simulation model was developed for the purpose of identifying and analyzing the areas for improvement. The company was facing a problem of low throughput rate and subsequent forced stoppages of the plant leading to a high production loss of 15000MT per month. It was found from the study that the present systems and procedures related to the in-plant logistics plant required significant changes. The major recommendations included process improvement at the entry gate, reducing the cycle time at the security gate and installation of an additional weigh bridge. This paper demonstrates how BPR can be implemented for improving the in-plant logistics process. Various recommendations helped the plant to increase its throughput by 14%.Keywords: in-plant logistics, cement logistics, simulation modelling, business process re-engineering, supply chain management
Procedia PDF Downloads 3007193 Variable Refrigerant Flow (VRF) Zonal Load Prediction Using a Transfer Learning-Based Framework
Authors: Junyu Chen, Peng Xu
Abstract:
In the context of global efforts to enhance building energy efficiency, accurate thermal load forecasting is crucial for both device sizing and predictive control. Variable Refrigerant Flow (VRF) systems are widely used in buildings around the world, yet VRF zonal load prediction has received limited attention. Due to differences between VRF zones in building-level prediction methods, zone-level load forecasting could significantly enhance accuracy. Given that modern VRF systems generate high-quality data, this paper introduces transfer learning to leverage this data and further improve prediction performance. This framework also addresses the challenge of predicting load for building zones with no historical data, offering greater accuracy and usability compared to pure white-box models. The study first establishes an initial variable set of VRF zonal building loads and generates a foundational white-box database using EnergyPlus. Key variables for VRF zonal loads are identified using methods including SRRC, PRCC, and Random Forest. XGBoost and LSTM are employed to generate pre-trained black-box models based on the white-box database. Finally, real-world data is incorporated into the pre-trained model using transfer learning to enhance its performance in operational buildings. In this paper, zone-level load prediction was integrated with transfer learning, and a framework was proposed to improve the accuracy and applicability of VRF zonal load prediction.Keywords: zonal load prediction, variable refrigerant flow (VRF) system, transfer learning, energyplus
Procedia PDF Downloads 307192 Facial Recognition of University Entrance Exam Candidates using FaceMatch Software in Iran
Authors: Mahshid Arabi
Abstract:
In recent years, remarkable advancements in the fields of artificial intelligence and machine learning have led to the development of facial recognition technologies. These technologies are now employed in a wide range of applications, including security, surveillance, healthcare, and education. In the field of education, the identification of university entrance exam candidates has been one of the fundamental challenges. Traditional methods such as using ID cards and handwritten signatures are not only inefficient and prone to fraud but also susceptible to errors. In this context, utilizing advanced technologies like facial recognition can be an effective and efficient solution to increase the accuracy and reliability of identity verification in entrance exams. This article examines the use of FaceMatch software for recognizing the faces of university entrance exam candidates in Iran. The main objective of this research is to evaluate the efficiency and accuracy of FaceMatch software in identifying university entrance exam candidates to prevent fraud and ensure the authenticity of individuals' identities. Additionally, this research investigates the advantages and challenges of using this technology in Iran's educational systems. This research was conducted using an experimental method and random sampling. In this study, 1000 university entrance exam candidates in Iran were selected as samples. The facial images of these candidates were processed and analyzed using FaceMatch software. The software's accuracy and efficiency were evaluated using various metrics, including accuracy rate, error rate, and processing time. The research results indicated that FaceMatch software could accurately identify candidates with a precision of 98.5%. The software's error rate was less than 1.5%, demonstrating its high efficiency in facial recognition. Additionally, the average processing time for each candidate's image was less than 2 seconds, indicating the software's high efficiency. Statistical evaluation of the results using precise statistical tests, including analysis of variance (ANOVA) and t-test, showed that the observed differences were significant, and the software's accuracy in identity verification is high. The findings of this research suggest that FaceMatch software can be effectively used as a tool for identifying university entrance exam candidates in Iran. This technology not only enhances security and prevents fraud but also simplifies and streamlines the exam administration process. However, challenges such as preserving candidates' privacy and the costs of implementation must also be considered. The use of facial recognition technology with FaceMatch software in Iran's educational systems can be an effective solution for preventing fraud and ensuring the authenticity of university entrance exam candidates' identities. Given the promising results of this research, it is recommended that this technology be more widely implemented and utilized in the country's educational systems.Keywords: facial recognition, FaceMatch software, Iran, university entrance exam
Procedia PDF Downloads 497191 On the Exergy Analysis of the Aluminum Smelter
Authors: Ayoola T. Brimmo, Mohamed I. Hassan
Abstract:
The push to mitigate the aluminum smelting industry’s enormous energy consumption and high emission releases is now even more persistent with the recent climate change happenings. Common approaches to achieve this have been focused on improving energy efficiency in the pot line and cast house sections of the smelter. However, the conventional energy efficiency analyses are based on the first law of thermodynamics, which do not shed proper light on the smelter’s degradation of energy. This just gives a general idea of the furnace’s performance with no reference to locations where improvement is a possibility based on the second law of thermodynamics. In this study, we apply exergy analyses on the pot line and cast house sections of the smelter to identify the locality and causes of energy degradation. The exergy analyses, which are based on a real life smelter conditions, highlight the possible locations for technology improvement in a typical smelter. With this established, methods of minimizing the smelter’s exergy losses are assessed.Keywords: exergy analysis, electrolytic cell, furnace, heat transfer
Procedia PDF Downloads 2897190 Phantom and Clinical Evaluation of Block Sequential Regularized Expectation Maximization Reconstruction Algorithm in Ga-PSMA PET/CT Studies Using Various Relative Difference Penalties and Acquisition Durations
Authors: Fatemeh Sadeghi, Peyman Sheikhzadeh
Abstract:
Introduction: Block Sequential Regularized Expectation Maximization (BSREM) reconstruction algorithm was recently developed to suppress excessive noise by applying a relative difference penalty. The aim of this study was to investigate the effect of various strengths of noise penalization factor in the BSREM algorithm under different acquisition duration and lesion sizes in order to determine an optimum penalty factor by considering both quantitative and qualitative image evaluation parameters in clinical uses. Materials and Methods: The NEMA IQ phantom and 15 clinical whole-body patients with prostate cancer were evaluated. Phantom and patients were injected withGallium-68 Prostate-Specific Membrane Antigen(68 Ga-PSMA)and scanned on a non-time-of-flight Discovery IQ Positron Emission Tomography/Computed Tomography(PET/CT) scanner with BGO crystals. The data were reconstructed using BSREM with a β-value of 100-500 at an interval of 100. These reconstructions were compared to OSEM as a widely used reconstruction algorithm. Following the standard NEMA measurement procedure, background variability (BV), recovery coefficient (RC), contrast recovery (CR) and residual lung error (LE) from phantom data and signal-to-noise ratio (SNR), signal-to-background ratio (SBR) and tumor SUV from clinical data were measured. Qualitative features of clinical images visually were ranked by one nuclear medicine expert. Results: The β-value acts as a noise suppression factor, so BSREM showed a decreasing image noise with an increasing β-value. BSREM, with a β-value of 400 at a decreased acquisition duration (2 min/ bp), made an approximately equal noise level with OSEM at an increased acquisition duration (5 min/ bp). For the β-value of 400 at 2 min/bp duration, SNR increased by 43.7%, and LE decreased by 62%, compared with OSEM at a 5 min/bp duration. In both phantom and clinical data, an increase in the β-value is translated into a decrease in SUV. The lowest level of SUV and noise were reached with the highest β-value (β=500), resulting in the highest SNR and lowest SBR due to the greater noise reduction than SUV reduction at the highest β-value. In compression of BSREM with different β-values, the relative difference in the quantitative parameters was generally larger for smaller lesions. As the β-value decreased from 500 to 100, the increase in CR was 160.2% for the smallest sphere (10mm) and 12.6% for the largest sphere (37mm), and the trend was similar for SNR (-58.4% and -20.5%, respectively). BSREM visually was ranked more than OSEM in all Qualitative features. Conclusions: The BSREM algorithm using more iteration numbers leads to more quantitative accuracy without excessive noise, which translates into higher overall image quality and lesion detectability. This improvement can be used to shorter acquisition time.Keywords: BSREM reconstruction, PET/CT imaging, noise penalization, quantification accuracy
Procedia PDF Downloads 987189 HPPDFIM-HD: Transaction Distortion and Connected Perturbation Approach for Hierarchical Privacy Preserving Distributed Frequent Itemset Mining over Horizontally-Partitioned Dataset
Authors: Fuad Ali Mohammed Al-Yarimi
Abstract:
Many algorithms have been proposed to provide privacy preserving in data mining. These protocols are based on two main approaches named as: the perturbation approach and the Cryptographic approach. The first one is based on perturbation of the valuable information while the second one uses cryptographic techniques. The perturbation approach is much more efficient with reduced accuracy while the cryptographic approach can provide solutions with perfect accuracy. However, the cryptographic approach is a much slower method and requires considerable computation and communication overhead. In this paper, a new scalable protocol is proposed which combines the advantages of the perturbation and distortion along with cryptographic approach to perform privacy preserving in distributed frequent itemset mining on horizontally distributed data. Both the privacy and performance characteristics of the proposed protocol are studied empirically.Keywords: anonymity data, data mining, distributed frequent itemset mining, gaussian perturbation, perturbation approach, privacy preserving data mining
Procedia PDF Downloads 5057188 Recent Developments in the Application of Deep Learning to Stock Market Prediction
Authors: Shraddha Jain Sharma, Ratnalata Gupta
Abstract:
Predicting stock movements in the financial market is both difficult and rewarding. Analysts and academics are increasingly using advanced approaches such as machine learning techniques to anticipate stock price patterns, thanks to the expanding capacity of computing and the recent advent of graphics processing units and tensor processing units. Stock market prediction is a type of time series prediction that is incredibly difficult to do since stock prices are influenced by a variety of financial, socioeconomic, and political factors. Furthermore, even minor mistakes in stock market price forecasts can result in significant losses for companies that employ the findings of stock market price prediction for financial analysis and investment. Soft computing techniques are increasingly being employed for stock market prediction due to their better accuracy than traditional statistical methodologies. The proposed research looks at the need for soft computing techniques in stock market prediction, the numerous soft computing approaches that are important to the field, past work in the area with their prominent features, and the significant problems or issue domain that the area involves. For constructing a predictive model, the major focus is on neural networks and fuzzy logic. The stock market is extremely unpredictable, and it is unquestionably tough to correctly predict based on certain characteristics. This study provides a complete overview of the numerous strategies investigated for high accuracy prediction, with a focus on the most important characteristics.Keywords: stock market prediction, artificial intelligence, artificial neural networks, fuzzy logic, accuracy, deep learning, machine learning, stock price, trading volume
Procedia PDF Downloads 927187 Ultra-High Precision Diamond Turning of Infrared Lenses
Authors: Khaled Abou-El-Hossein
Abstract:
The presentation will address the features of two IR convex lenses that have been manufactured using an ultra-high precision machining centre based on single-point diamond turning. The lenses are made from silicon and germanium with a radius of curvature of 500 mm. Because of the brittle nature of silicon and germanium, machining parameters were selected in such a way that ductile regime was achieved. The cutting speed was 800 rpm while the feed rate and depth cut were 20 mm/min and 20 um, respectively. Although both materials comprise a mono-crystalline microstructure and are quite similar in terms of optical properties, machining of silicon was accompanied with more difficulties in terms of form accuracy compared to germanium machining. The P-V error of the silicon profile was 0.222 um while it was only 0.055 um for the germanium lens. This could be attributed to the accelerated wear that takes place on the tool edge when turning mono-crystalline silicon. Currently, we are using other ranges of the machining parameters in order to determine their optimal range that could yield satisfactory performance in terms of form accuracy when fabricating silicon lenses.Keywords: diamond turning, optical surfaces, precision machining, surface roughness
Procedia PDF Downloads 3177186 A Spatial Approach to Model Mortality Rates
Authors: Yin-Yee Leong, Jack C. Yue, Hsin-Chung Wang
Abstract:
Human longevity has been experiencing its largest increase since the end of World War II, and modeling the mortality rates is therefore often the focus of many studies. Among all mortality models, the Lee–Carter model is the most popular approach since it is fairly easy to use and has good accuracy in predicting mortality rates (e.g., for Japan and the USA). However, empirical studies from several countries have shown that the age parameters of the Lee–Carter model are not constant in time. Many modifications of the Lee–Carter model have been proposed to deal with this problem, including adding an extra cohort effect and adding another period effect. In this study, we propose a spatial modification and use clusters to explain why the age parameters of the Lee–Carter model are not constant. In spatial analysis, clusters are areas with unusually high or low mortality rates than their neighbors, where the “location” of mortality rates is measured by age and time, that is, a 2-dimensional coordinate. We use a popular cluster detection method—Spatial scan statistics, a local statistical test based on the likelihood ratio test to evaluate where there are locations with mortality rates that cannot be described well by the Lee–Carter model. We first use computer simulation to demonstrate that the cluster effect is a possible source causing the problem of the age parameters not being constant. Next, we show that adding the cluster effect can solve the non-constant problem. We also apply the proposed approach to mortality data from Japan, France, the USA, and Taiwan. The empirical results show that our approach has better-fitting results and smaller mean absolute percentage errors than the Lee–Carter model.Keywords: mortality improvement, Lee–Carter model, spatial statistics, cluster detection
Procedia PDF Downloads 1727185 Preparation of Papers - Developing a Leukemia Diagnostic System Based on Hybrid Deep Learning Architectures in Actual Clinical Environments
Authors: Skyler Kim
Abstract:
An early diagnosis of leukemia has always been a challenge to doctors and hematologists. On a worldwide basis, it was reported that there were approximately 350,000 new cases in 2012, and diagnosing leukemia was time-consuming and inefficient because of an endemic shortage of flow cytometry equipment in current clinical practice. As the number of medical diagnosis tools increased and a large volume of high-quality data was produced, there was an urgent need for more advanced data analysis methods. One of these methods was the AI approach. This approach has become a major trend in recent years, and several research groups have been working on developing these diagnostic models. However, designing and implementing a leukemia diagnostic system in real clinical environments based on a deep learning approach with larger sets remains complex. Leukemia is a major hematological malignancy that results in mortality and morbidity throughout different ages. We decided to select acute lymphocytic leukemia to develop our diagnostic system since acute lymphocytic leukemia is the most common type of leukemia, accounting for 74% of all children diagnosed with leukemia. The results from this development work can be applied to all other types of leukemia. To develop our model, the Kaggle dataset was used, which consists of 15135 total images, 8491 of these are images of abnormal cells, and 5398 images are normal. In this paper, we design and implement a leukemia diagnostic system in a real clinical environment based on deep learning approaches with larger sets. The proposed diagnostic system has the function of detecting and classifying leukemia. Different from other AI approaches, we explore hybrid architectures to improve the current performance. First, we developed two independent convolutional neural network models: VGG19 and ResNet50. Then, using both VGG19 and ResNet50, we developed a hybrid deep learning architecture employing transfer learning techniques to extract features from each input image. In our approach, fusing the features from specific abstraction layers can be deemed as auxiliary features and lead to further improvement of the classification accuracy. In this approach, features extracted from the lower levels are combined into higher dimension feature maps to help improve the discriminative capability of intermediate features and also overcome the problem of network gradient vanishing or exploding. By comparing VGG19 and ResNet50 and the proposed hybrid model, we concluded that the hybrid model had a significant advantage in accuracy. The detailed results of each model’s performance and their pros and cons will be presented in the conference.Keywords: acute lymphoblastic leukemia, hybrid model, leukemia diagnostic system, machine learning
Procedia PDF Downloads 187