Search results for: image generation
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 6049

Search results for: image generation

4669 A Mutually Exclusive Task Generation Method Based on Data Augmentation

Authors: Haojie Wang, Xun Li, Rui Yin

Abstract:

In order to solve the memorization overfitting in the model-agnostic meta-learning MAML algorithm, a method of generating mutually exclusive tasks based on data augmentation is proposed. This method generates a mutex task by corresponding one feature of the data to multiple labels so that the generated mutex task is inconsistent with the data distribution in the initial dataset. Because generating mutex tasks for all data will produce a large number of invalid data and, in the worst case, lead to an exponential growth of computation, this paper also proposes a key data extraction method that only extract part of the data to generate the mutex task. The experiments show that the method of generating mutually exclusive tasks can effectively solve the memorization overfitting in the meta-learning MAML algorithm.

Keywords: mutex task generation, data augmentation, meta-learning, text classification.

Procedia PDF Downloads 143
4668 Colored Image Classification Using Quantum Convolutional Neural Networks Approach

Authors: Farina Riaz, Shahab Abdulla, Srinjoy Ganguly, Hajime Suzuki, Ravinesh C. Deo, Susan Hopkins

Abstract:

Recently, quantum machine learning has received significant attention. For various types of data, including text and images, numerous quantum machine learning (QML) models have been created and are being tested. Images are exceedingly complex data components that demand more processing power. Despite being mature, classical machine learning still has difficulties with big data applications. Furthermore, quantum technology has revolutionized how machine learning is thought of, by employing quantum features to address optimization issues. Since quantum hardware is currently extremely noisy, it is not practicable to run machine learning algorithms on it without risking the production of inaccurate results. To discover the advantages of quantum versus classical approaches, this research has concentrated on colored image data. Deep learning classification models are currently being created on Quantum platforms, but they are still in a very early stage. Black and white benchmark image datasets like MNIST and Fashion MINIST have been used in recent research. MNIST and CIFAR-10 were compared for binary classification, but the comparison showed that MNIST performed more accurately than colored CIFAR-10. This research will evaluate the performance of the QML algorithm on the colored benchmark dataset CIFAR-10 to advance QML's real-time applicability. However, deep learning classification models have not been developed to compare colored images like Quantum Convolutional Neural Network (QCNN) to determine how much it is better to classical. Only a few models, such as quantum variational circuits, take colored images. The methodology adopted in this research is a hybrid approach by using penny lane as a simulator. To process the 10 classes of CIFAR-10, the image data has been translated into grey scale and the 28 × 28-pixel image containing 10,000 test and 50,000 training images were used. The objective of this work is to determine how much the quantum approach can outperform a classical approach for a comprehensive dataset of color images. After pre-processing 50,000 images from a classical computer, the QCNN model adopted a hybrid method and encoded the images into a quantum simulator for feature extraction using quantum gate rotations. The measurements were carried out on the classical computer after the rotations were applied. According to the results, we note that the QCNN approach is ~12% more effective than the traditional classical CNN approaches and it is possible that applying data augmentation may increase the accuracy. This study has demonstrated that quantum machine and deep learning models can be relatively superior to the classical machine learning approaches in terms of their processing speed and accuracy when used to perform classification on colored classes.

Keywords: CIFAR-10, quantum convolutional neural networks, quantum deep learning, quantum machine learning

Procedia PDF Downloads 129
4667 Small Target Recognition Based on Trajectory Information

Authors: Saad Alkentar, Abdulkareem Assalem

Abstract:

Recognizing small targets has always posed a significant challenge in image analysis. Over long distances, the image signal-to-noise ratio tends to be low, limiting the amount of useful information available to detection systems. Consequently, visual target recognition becomes an intricate task to tackle. In this study, we introduce a Track Before Detect (TBD) approach that leverages target trajectory information (coordinates) to effectively distinguish between noise and potential targets. By reframing the problem as a multivariate time series classification, we have achieved remarkable results. Specifically, our TBD method achieves an impressive 97% accuracy in separating target signals from noise within a mere half-second time span (consisting of 10 data points). Furthermore, when classifying the identified targets into our predefined categories—airplane, drone, and bird—we achieve an outstanding classification accuracy of 96% over a more extended period of 1.5 seconds (comprising 30 data points).

Keywords: small targets, drones, trajectory information, TBD, multivariate time series

Procedia PDF Downloads 48
4666 Performance Comparison of Tablet Devices and Medical Diagnostic Display Devices Using Digital Object Patterns in PACS Environment

Authors: Yan-Lin Liu, Cheng-Ting Shih, Jay Wu

Abstract:

Tablet devices have been introduced into the medical environment in recent years. The performance of display can be varied based on the use of different hardware specifications and types of display technologies. Therefore, the differences between tablet devices and medical diagnostic LCDs have to be verified to ensure that image quality is not jeopardized for clinical diagnosis in a picture archiving and communication system (PACS). In this study, a set of randomized object test patterns (ROTPs) were developed, which included randomly located spheres in abdominal CT images. Five radiologists were asked to independently review the CT images on different generations of iPads and a diagnostic monochrome medical LCD monitor. Receiver operating characteristic (ROC) analysis was performed by using a five-point rating scale, and the average area under curve (AUC) and average reading time (ART) were calculated. The AUC values for the second generation iPad, iPad mini, iPad Air, and monochrome medical monitor were 0.712, 0.717, 0.725, and 0.740, respectively. The differences between iPads were not significant. The ARTs were 177 min and 127 min for iPad mini and medical LCD monitor, respectively. A significant difference appeared (p = 0.04). The results show that the iPads were slightly inferior to the monochrome medical LCD monitor. However, tablet devices possess advantages in portability and versatility, which can improve the convenience of rapid diagnosis and teleradiology. With advances in display technology, the applicability of tablet devices and mobile devices may be more diversified in PACS.

Keywords: tablet devices, PACS, receiver operating characteristic, LCD monitor

Procedia PDF Downloads 480
4665 Best Timing for Capturing Satellite Thermal Images, Asphalt, and Concrete Objects

Authors: Toufic Abd El-Latif Sadek

Abstract:

The asphalt object represents the asphalted areas like roads, and the concrete object represents the concrete areas like concrete buildings. The efficient extraction of asphalt and concrete objects from one satellite thermal image occurred at a specific time, by preventing the gaps in times which give the close and same brightness values between asphalt and concrete, and among other objects. So that to achieve efficient extraction and then better analysis. Seven sample objects were used un this study, asphalt, concrete, metal, rock, dry soil, vegetation, and water. It has been found that, the best timing for capturing satellite thermal images to extract the two objects asphalt and concrete from one satellite thermal image, saving time and money, occurred at a specific time in different months. A table is deduced shows the optimal timing for capturing satellite thermal images to extract effectively these two objects.

Keywords: asphalt, concrete, satellite thermal images, timing

Procedia PDF Downloads 322
4664 Active Power Control of PEM Fuel Cell System Power Generation Using Adaptive Neuro-Fuzzy Controller

Authors: Khaled Mammar

Abstract:

This paper presents an application of adaptive neuro-fuzzy controller for PEM fuel cell system. The model proposed for control include a fuel cell stack model, reformer model and DC/AC inverter model. Furthermore, a Fuzzy Logic (FLC) and adaptive neuro-fuzzy controllers are used to control the active power of PEM fuel cell system. The controllers modify the hydrogen flow feedback from the terminal load. The validity of the controller is verified when the fuel cell system model is used in conjunction with the ANFIS controller to predict the response of the active power. Simulation results confirmed the high-performance capability of the neuo-fuzzy to control power generation.

Keywords: fuel cell, PEMFC, modeling, simulation, Fuzzy Logic Controller, FLC, adaptive neuro-fuzzy controller, ANFIS

Procedia PDF Downloads 460
4663 New Method to Increase Contrast of Electromicrograph of Rat Tissues Sections

Authors: Lise Paule Labéjof, Raíza Sales Pereira Bizerra, Galileu Barbosa Costa, Thaísa Barros dos Santos

Abstract:

Since the beginning of the microscopy, improving the image quality has always been a concern of its users. Especially for transmission electron microscopy (TEM), the problem is even more important due to the complexity of the sample preparation technique and the many variables that can affect the conservation of structures, proper operation of the equipment used and then the quality of the images obtained. Animal tissues being transparent it is necessary to apply a contrast agent in order to identify the elements of their ultrastructural morphology. Several methods of contrastation of tissues for TEM imaging have already been developed. The most used are the “in block” contrastation and “in situ” contrastation. This report presents an alternative technique of application of contrast agent in vivo, i.e. before sampling. By this new method the electromicrographies of the tissue sections have better contrast compared to that in situ and present no artefact of precipitation of contrast agent. Another advantage is that a small amount of contrast is needed to get a good result given that most of them are expensive and extremely toxic.

Keywords: image quality, microscopy research, staining technique, ultra thin section

Procedia PDF Downloads 433
4662 Development of a Mobile Image-Based Reminder Application to Support Tuberculosis Treatment in Africa

Authors: Haji Ali Haji, Hussein Suleman, Ulrike Rivett

Abstract:

This paper presents the design, development and evaluation of an application prototype developed to support tuberculosis (TB) patients’ treatment adherence. The system makes use of graphics and voice reminders as opposed to text messaging to encourage patients to follow their medication routine. To evaluate the effect of the prototype applications, participants were given mobile phones on which the reminder system was installed. Thirty-eight people, including TB health workers and patients from Zanzibar, Tanzania, participated in the evaluation exercises. The results indicate that the participants found the mobile graphic-based application is useful to support TB treatment. All participants understood and interpreted the intended meaning of every image correctly. The study findings revealed that the use of a mobile visual-based application may have potential benefit to support TB patients (both literate and illiterate) in their treatment processes.

Keywords: ICT4D, mobile technology, tuberculosis, visual-based reminder

Procedia PDF Downloads 430
4661 Deep Learning-Based Liver 3D Slicer for Image-Guided Therapy: Segmentation and Needle Aspiration

Authors: Ahmedou Moulaye Idriss, Tfeil Yahya, Tamas Ungi, Gabor Fichtinger

Abstract:

Image-guided therapy (IGT) plays a crucial role in minimally invasive procedures for liver interventions. Accurate segmentation of the liver and precise needle placement is essential for successful interventions such as needle aspiration. In this study, we propose a deep learning-based liver 3D slicer designed to enhance segmentation accuracy and facilitate needle aspiration procedures. The developed 3D slicer leverages state-of-the-art convolutional neural networks (CNNs) for automatic liver segmentation in medical images. The CNN model is trained on a diverse dataset of liver images obtained from various imaging modalities, including computed tomography (CT) and magnetic resonance imaging (MRI). The trained model demonstrates robust performance in accurately delineating liver boundaries, even in cases with anatomical variations and pathological conditions. Furthermore, the 3D slicer integrates advanced image registration techniques to ensure accurate alignment of preoperative images with real-time interventional imaging. This alignment enhances the precision of needle placement during aspiration procedures, minimizing the risk of complications and improving overall intervention outcomes. To validate the efficacy of the proposed deep learning-based 3D slicer, a comprehensive evaluation is conducted using a dataset of clinical cases. Quantitative metrics, including the Dice similarity coefficient and Hausdorff distance, are employed to assess the accuracy of liver segmentation. Additionally, the performance of the 3D slicer in guiding needle aspiration procedures is evaluated through simulated and clinical interventions. Preliminary results demonstrate the effectiveness of the developed 3D slicer in achieving accurate liver segmentation and guiding needle aspiration procedures with high precision. The integration of deep learning techniques into the IGT workflow shows great promise for enhancing the efficiency and safety of liver interventions, ultimately contributing to improved patient outcomes.

Keywords: deep learning, liver segmentation, 3D slicer, image guided therapy, needle aspiration

Procedia PDF Downloads 48
4660 Photomicrograph-Based Neuropathology Consultation in Tanzania; The Utility of Static-Image Neurotelepathology in Low- And Middle-Income Countries

Authors: Francis Zerd, Brian E. Moore, Atuganile E. Malango, Patrick W. Hosokawa, Kevin O. Lillehei, Laurence Lemery Mchome, D. Ryan Ormond

Abstract:

Introduction: Since neuropathologic diagnosis in the developing world is hampered by limitations in technical infrastructure, trained laboratory personnel, and subspecialty-trained pathologists, the use of telepathology for diagnostic support, second-opinion consultations, and ongoing training holds promise as a means of addressing these challenges. This research aims to assess the utility of static teleneuropathology in improving neuropathologic diagnoses in low- and middle-income countries. Methods: Consecutive neurosurgical biopsy and resection specimens obtained at Muhimbili National Hospital in Tanzania between July 1, 2018, and June 30, 2019, were selected for retrospective, blinded static-image neuropathologic review followed by on-site review by an expert neuropathologist. Results: A total of 75 neuropathologic cases were reviewed. The agreement of static images and on-site glass diagnosis was 71% with strict criteria and 88% with less stringent criteria. This represents an overall improvement in diagnostic accuracy from 36% by general pathologists to 71% by a neuropathologist using static telepathology (or 76% to 88% with less stringent criteria). Conclusions: Telepathology offers a suitable means of providing diagnostic support, second-opinion consultations, and ongoing training to pathologists practicing in resource-limited countries. Moreover, static digital teleneuropathology is an uncomplicated, cost-effective, and reliable way to achieve these goals.

Keywords: neuropathology, resource-limited settings, static image, Tanzania, teleneuropathology

Procedia PDF Downloads 102
4659 Easily Memorable Strong Password Generation and Retrieval

Authors: Shatadru Das, Natarajan Vijayarangan

Abstract:

In this paper, a system and method for generating and recovering an authorization code has been designed and analyzed. The system creates an authorization code by accepting a base-sentence from a user. Based on the characters present in this base-sentence, the system computes a base-sentence matrix. The system also generates a plurality of patterns. The user can either select the pattern from the multiple patterns suggested by the system or can create his/her own pattern. The system then performs multiplications between the base-sentence matrix and the selected pattern matrix at different stages in the path forward, for obtaining a strong authorization code. In case the user forgets the base sentence, the system has a provision to manage and retrieve 'forgotten authorization code'. This is done by fragmenting the base sentence into different matrices and storing the fragmented matrices into a repository after computing matrix multiplication with a security question-answer approach and with a secret key provided by the user.

Keywords: easy authentication, key retrieval, memorable passwords, strong password generation

Procedia PDF Downloads 400
4658 Multimodality in Storefront Windows: The Impact of Verbo-Visual Design on Consumer Behavior

Authors: Angela Bargenda, Erhard Lick, Dhoha Trabelsi

Abstract:

Research in retailing has identified the importance of atmospherics as an essential element in enhancing store image, store patronage intentions, and the overall shopping experience in a retail environment. However, in the area of atmospherics, store window design, which represents an essential component of external store atmospherics, remains a vastly underrepresented phenomenon in extant scholarship. This paper seeks to fill this gap by exploring the relevance of store window design as an atmospheric tool. In particular, empirical evidence of theme-based theatrical store front windows, which put emphasis on the use of verbo-visual design elements, was found in Paris and New York. The purpose of this study was to identify to what extent such multimodal window designs of high-end department stores in metropolitan cities have an impact on store entry decisions and attitudes towards the retailer’s image. As theoretical construct, the linguistic concept of multimodality and Mehrabian’s and Russell’s model in environmental psychology were applied. To answer the research question, two studies were conducted. For Study 1 a case study approach was selected to define three different types of store window designs based on different types of visual-verbal relations. Each of these types of store window design represented a different level of cognitive elaboration required for the decoding process. Study 2 consisted of an on-line survey carried out among more than 300 respondents to examine the influence of these three types of store window design on the consumer behavioral variables mentioned above. The results of this study show that the higher the cognitive elaboration needed to decode the message of the store window, the lower the store entry propensity. In contrast, the higher the cognitive elaboration, the higher the perceived image of the retailer’s image. One important conclusion is that in order to increase consumers’ propensity to enter stores with theme-based theatrical store front windows, retailers need to limit the cognitive elaboration required to decode their verbo-visual window design.

Keywords: consumer behavior, multimodality, store atmospherics, store window design

Procedia PDF Downloads 202
4657 Heating Behavior of Ni-Embedded Thermoplastic Polyurethane Adhesive Film by Induction Heating

Authors: DuckHwan Bae, YongSung Kwon, Min Young Shon, SanTaek Oh, GuNi Kim

Abstract:

The heating behavior of nanometer and micrometer sized Nickel particle-imbedded thermoplastic polyurethane adhesive (TPU) under induction heating is examined in present study. The effects of particle size and content, TPU film thickness on heating behaviors were examined. The correlation between heating behavior and magnetic properties of Nickel particles were also studied. From the results, heat generation increased with increase of Nickel content and film thickness. However, in terms of particle sizes, heat generation of Nickel-imbedded TPU film were in order of 70nm>1µm>20 µm>70 µm and this results can explain by increasing ration of eddy heating to hysteresis heating with increase of particle size.

Keywords: induction heating, thermoplastic polyurethane, nickel, composite, hysteresis loss, eddy current loss, curie temperature

Procedia PDF Downloads 362
4656 A Neural Network Classifier for Identifying Duplicate Image Entries in Real-Estate Databases

Authors: Sergey Ermolin, Olga Ermolin

Abstract:

A Deep Convolution Neural Network with Triplet Loss is used to identify duplicate images in real-estate advertisements in the presence of image artifacts such as watermarking, cropping, hue/brightness adjustment, and others. The effects of batch normalization, spatial dropout, and various convergence methodologies on the resulting detection accuracy are discussed. For comparative Return-on-Investment study (per industry request), end-2-end performance is benchmarked on both Nvidia Titan GPUs and Intel’s Xeon CPUs. A new real-estate dataset from San Francisco Bay Area is used for this work. Sufficient duplicate detection accuracy is achieved to supplement other database-grounded methods of duplicate removal. The implemented method is used in a Proof-of-Concept project in the real-estate industry.

Keywords: visual recognition, convolutional neural networks, triplet loss, spatial batch normalization with dropout, duplicate removal, advertisement technologies, performance benchmarking

Procedia PDF Downloads 338
4655 A Hybrid Watermarking Model Based on Frequency of Occurrence

Authors: Hamza A. A. Al-Sewadi, Adnan H. M. Al-Helali, Samaa A. K. Khamis

Abstract:

Ownership proofs of multimedia such as text, image, audio or video files can be achieved by the burial of watermark is them. It is achieved by introducing modifications into these files that are imperceptible to the human senses but easily recoverable by a computer program. These modifications would be in the time domain or frequency domain or both. This paper presents a procedure for watermarking by mixing amplitude modulation with frequency transformation histogram; namely a specific value is used to modulate the intensity component Y of the YIQ components of the carrier image. This scheme is referred to as histogram embedding technique (HET). Results comparison with those of other techniques such as discrete wavelet transform (DWT), discrete cosine transform (DCT) and singular value decomposition (SVD) have shown an enhance efficiency in terms of ease and performance. It has manifested a good degree of robustness against various environment effects such as resizing, rotation and different kinds of noise. This method would prove very useful technique for copyright protection and ownership judgment.

Keywords: authentication, copyright protection, information hiding, ownership, watermarking

Procedia PDF Downloads 565
4654 Object Tracking in Motion Blurred Images with Adaptive Mean Shift and Wavelet Feature

Authors: Iman Iraei, Mina Sharifi

Abstract:

A method for object tracking in motion blurred images is proposed in this article. This paper shows that object tracking could be improved with this approach. We use mean shift algorithm to track different objects as a main tracker. But, the problem is that mean shift could not track the selected object accurately in blurred scenes. So, for better tracking result, and increasing the accuracy of tracking, wavelet transform is used. We use a feature named as blur extent, which could help us to get better results in tracking. For calculating of this feature, we should use Harr wavelet. We can look at this matter from two different angles which lead to determine whether an image is blurred or not and to what extent an image is blur. In fact, this feature left an impact on the covariance matrix of mean shift algorithm and cause to better performance of tracking. This method has been concentrated mostly on motion blur parameter. transform. The results reveal the ability of our method in order to reach more accurately tracking.

Keywords: mean shift, object tracking, blur extent, wavelet transform, motion blur

Procedia PDF Downloads 211
4653 Utilizing the Principal Component Analysis on Multispectral Aerial Imagery for Identification of Underlying Structures

Authors: Marcos Bosques-Perez, Walter Izquierdo, Harold Martin, Liangdon Deng, Josue Rodriguez, Thony Yan, Mercedes Cabrerizo, Armando Barreto, Naphtali Rishe, Malek Adjouadi

Abstract:

Aerial imagery is a powerful tool when it comes to analyzing temporal changes in ecosystems and extracting valuable information from the observed scene. It allows us to identify and assess various elements such as objects, structures, textures, waterways, and shadows. To extract meaningful information, multispectral cameras capture data across different wavelength bands of the electromagnetic spectrum. In this study, the collected multispectral aerial images were subjected to principal component analysis (PCA) to identify independent and uncorrelated components or features that extend beyond the visible spectrum captured in standard RGB images. The results demonstrate that these principal components contain unique characteristics specific to certain wavebands, enabling effective object identification and image segmentation.

Keywords: big data, image processing, multispectral, principal component analysis

Procedia PDF Downloads 178
4652 Rainfall Estimation Using Himawari-8 Meteorological Satellite Imagery in Central Taiwan

Authors: Chiang Wei, Hui-Chung Yeh, Yen-Chang Chen

Abstract:

The objective of this study is to estimate the rainfall using the new generation Himawari-8 meteorological satellite with multi-band, high-bit format, and high spatiotemporal resolution, ground rainfall data at the Chen-Yu-Lan watershed of Joushuei River Basin (443.6 square kilometers) in Central Taiwan. Accurate and fine-scale rainfall information is essential for rugged terrain with high local variation for early warning of flood, landslide, and debris flow disasters. 10-minute and 2 km pixel-based rainfall of Typhoon Megi of 2016 and meiyu on June 1-4 of 2017 were tested to demonstrate the new generation Himawari-8 meteorological satellite can capture rainfall variation in the rugged mountainous area both at fine-scale and watershed scale. The results provide the valuable rainfall information for early warning of future disasters.

Keywords: estimation, Himawari-8, rainfall, satellite imagery

Procedia PDF Downloads 194
4651 An Accurate Computer-Aided Diagnosis: CAD System for Diagnosis of Aortic Enlargement by Using Convolutional Neural Networks

Authors: Mahdi Bazarganigilani

Abstract:

Aortic enlargement, also known as an aortic aneurysm, can occur when the walls of the aorta become weak. This disease can become deadly if overlooked and undiagnosed. In this paper, a computer-aided diagnosis (CAD) system was introduced to accurately diagnose aortic enlargement from chest x-ray images. An enhanced convolutional neural network (CNN) was employed and then trained by transfer learning by using three different main areas from the original images. The areas included the left lung, heart, and right lung. The accuracy of the system was then evaluated on 1001 samples by using 4-fold cross-validation. A promising accuracy of 90% was achieved in terms of the F-measure indicator. The results showed using different areas from the original image in the training phase of CNN could increase the accuracy of predictions. This encouraged the author to evaluate this method on a larger dataset and even on different CAD systems for further enhancement of this methodology.

Keywords: computer-aided diagnosis systems, aortic enlargement, chest X-ray, image processing, convolutional neural networks

Procedia PDF Downloads 162
4650 Combustion Improvements by C4/C5 Bio-Alcohol Isomer Blended Fuels Combined with Supercharging and EGR in a Diesel Engine

Authors: Yasufumi Yoshimoto, Enkhjargal Tserenochir, Eiji Kinoshita, Takeshi Otaka

Abstract:

Next generation bio-alcohols produced from non-food based sources like cellulosic biomass are promising renewable energy sources. The present study investigates engine performance, combustion characteristics, and emissions of a small single cylinder direct injection diesel engine fueled by four kinds of next generation bio-alcohol isomer and diesel fuel blends with a constant blending ratio of 3:7 (mass). The tested bio-alcohol isomers here are n-butanol and iso-butanol (C4 alcohol), and n-pentanol and iso-pentanol (C5 alcohol). To obtain simultaneous reductions in NOx and smoke emissions, the experiments employed supercharging combined with EGR (Exhaust Gas Recirculation). The boost pressures were fixed at two conditions, 100 kPa (naturally aspirated operation) and 120 kPa (supercharged operation) provided with a roots blower type supercharger. The EGR rates were varied from 0 to 25% using a cooled EGR technique. The results showed that both with and without supercharging, all the bio-alcohol blended diesel fuels improved the trade-off relation between NOx and smoke emissions at all EGR rates while maintaining good engine performance, when compared with diesel fuel operation. It was also found that regardless of boost pressure and EGR rate, the ignition delays of the tested bio-alcohol isomer blends are in the order of iso-butanol > n-butanol > iso-pentanol > n-pentanol. Overall, it was concluded that, except for the changes in the ignition delays the influence of bio-alcohol isomer blends on the engine performance, combustion characteristics, and emissions are relatively small.

Keywords: alternative fuel, butanol, diesel engine, EGR (Exhaust Gas Recirculation), next generation bio-alcohol isomer blended fuel, pentanol, supercharging

Procedia PDF Downloads 169
4649 An Analysis of the Relations between Aggregates’ Shape and Mechanical Properties throughout the Railway Ballast Service Life

Authors: Daianne Fernandes Diogenes

Abstract:

Railway ballast aggregates’ shape properties and size distribution can be directly affected by several factors, such as traffic, fouling, and maintenance processes, which cause breakage and wearing, leading to the fine particles’ accumulation through the ballast layer. This research aims to analyze the influence of traffic, tamping process, and sleepers’ stiffness on aggregates' shape and mechanical properties, by using traditional and digital image processing (DIP) techniques and cyclic tests, like resilient modulus (RM) and permanent deformation (PD). Aggregates were collected in different phases of the railway service life: (i) right after the crushing process; (ii) after construction, for the aggregates positioned below the sleepers and (iii) after 5 years of operation. An increase in the percentage of cubic particles was observed for the materials (ii) and (iii), providing a better interlocking, increasing stiffness and reducing axial deformation after 5 years of service, when compared to the initial conditions.

Keywords: digital image processing, mechanical behavior, railway ballast, shape properties

Procedia PDF Downloads 122
4648 Distribution System Planning with Distributed Generation and Capacitor Placements

Authors: Nattachote Rugthaicharoencheep

Abstract:

This paper presents a feeder reconfiguration problem in distribution systems. The objective is to minimize the system power loss and to improve bus voltage profile. The optimization problem is subjected to system constraints consisting of load-point voltage limits, radial configuration format, no load-point interruption, and feeder capability limits. A method based on genetic algorithm, a search algorithm based on the mechanics of natural selection and natural genetics, is proposed to determine the optimal pattern of configuration. The developed methodology is demonstrated by a 33-bus radial distribution system with distributed generations and feeder capacitors. The study results show that the optimal on/off patterns of the switches can be identified to give the minimum power loss while respecting all the constraints.

Keywords: network reconfiguration, distributed generation capacitor placement, loss reduction, genetic algorithm

Procedia PDF Downloads 177
4647 Enhancing Large Language Models' Data Analysis Capability with Planning-and-Execution and Code Generation Agents: A Use Case for Southeast Asia Real Estate Market Analytics

Authors: Kien Vu, Jien Min Soh, Mohamed Jahangir Abubacker, Piyawut Pattamanon, Soojin Lee, Suvro Banerjee

Abstract:

Recent advances in Generative Artificial Intelligence (GenAI), in particular Large Language Models (LLMs) have shown promise to disrupt multiple industries at scale. However, LLMs also present unique challenges, notably, these so-called "hallucination" which is the generation of outputs that are not grounded in the input data that hinders its adoption into production. Common practice to mitigate hallucination problem is utilizing Retrieval Agmented Generation (RAG) system to ground LLMs'response to ground truth. RAG converts the grounding documents into embeddings, retrieve the relevant parts with vector similarity between user's query and documents, then generates a response that is not only based on its pre-trained knowledge but also on the specific information from the retrieved documents. However, the RAG system is not suitable for tabular data and subsequent data analysis tasks due to multiple reasons such as information loss, data format, and retrieval mechanism. In this study, we have explored a novel methodology that combines planning-and-execution and code generation agents to enhance LLMs' data analysis capabilities. The approach enables LLMs to autonomously dissect a complex analytical task into simpler sub-tasks and requirements, then convert them into executable segments of code. In the final step, it generates the complete response from output of the executed code. When deployed beta version on DataSense, the property insight tool of PropertyGuru, the approach yielded promising results, as it was able to provide market insights and data visualization needs with high accuracy and extensive coverage by abstracting the complexities for real-estate agents and developers from non-programming background. In essence, the methodology not only refines the analytical process but also serves as a strategic tool for real estate professionals, aiding in market understanding and enhancement without the need for programming skills. The implication extends beyond immediate analytics, paving the way for a new era in the real estate industry characterized by efficiency and advanced data utilization.

Keywords: large language model, reasoning, planning and execution, code generation, natural language processing, prompt engineering, data analysis, real estate, data sense, PropertyGuru

Procedia PDF Downloads 87
4646 Brand Building in Higher Education: A Grounded Theory Investigation of the Impact of the ‘Positive-Visualization-Course in Brand Identity’ upon Freshmen Student's Perception

Authors: Maria Kountouridou, Dino Domic

Abstract:

Within an increasingly competitive and dynamic environment, the higher education sector is becoming more commodified, with the concept of branding to become exceedingly imperative and an inextricable ingredient for the university’s success. Branding in higher education has proven to be an effective strategy that managed to receive considerable attention in the recent few years, and a growing number of articles have begun to appear in the literature. However, a clear void in the literature confirms that the concept of students’ perceptions towards the university’s brand image has not been researched extensively. An investigation on this central concept is of paramount importance since it will facilitate the development of an inductively generated theoretical model concerning branding in higher education. This research focuses on examining the impact of the ‘positive-visualization-course in brand identity’ upon the perception of freshmen students towards a university’s brand image. A grounded theory methodology has been selected, consisting of semi-structured interviews. Forty-two students have participated in the research, among which twenty-five women and seventeen men. The identification of the sample emerged through the use of the snowball sampling technique. The participants were divided into two groups (experimental and control group) after the researcher had taken into consideration the factor ‘program of study’, to eliminate any possible interaction between the participants of each group. An experiment was carried out where a ‘positive-visualization-course in brand identity’ was conducted among the participants of the experimental group, while the participants of the control group have not been exposed to the course. For the purpose of this research, the term ‘positive-visualization-course in brand identity’ refers to a course where brand history, past achievements/recognitions/awards, its values, and its mission are presented. Prior to the course implementation, face-to-face semi-structured interviews were carried out among the participants of both groups, with the aim of examining the freshmen students’ perceptions towards the university’s brand image. One week after the course implementation, the researcher carried out semi-structured interviews with the participants of the experimental group only in order to identify whether students’ perceptions had been affected after the course completion. Four months after the course completion, semi-structured interviews were carried out among the participants of both groups. Eight months after the course completion, semi-structured interviews were conducted with the aim of identifying the freshmen students’ updated perceptions. Data has been analyzed using substantive coding (open and selective coding), theoretical coding, field memos, and constant comparative analysis. The findings strongly suggest that the ‘positive-visualization-course in brand identity’ can positively affect freshmen students’ perceptions towards a university’s brand image. Additionally, other factors conduce to the formation of perception throughout the months. This study contributes and expands upon the existing literature by presenting an inductively generated theoretical model to guide future research in the links between ‘positive-visualization-course in brand identity’ and the perception of freshmen students towards a university’s brand image.

Keywords: brand image, brand name, branding, higher education marketing, perception

Procedia PDF Downloads 178
4645 Frequency of Occurrence Hybrid Watermarking Scheme

Authors: Hamza A. Ali, Adnan H. M. Al-Helali

Abstract:

Generally, a watermark is information that identifies the ownership of multimedia (text, image, audio or video files). It is achieved by introducing modifications into these files that are imperceptible to the human senses but easily recoverable by a computer program. These modifications are done according to a secret key in a descriptive model that would be either in the time domain or frequency domain or both. This paper presents a procedure for watermarking by mixing amplitude modulation with frequency transformation histogram; namely a specific value is used to modulate the intensity component Y of the YIQ components of the carrier image. This scheme is referred to as histogram embedding technique (HET). Results comparison with those of other techniques such as discrete wavelet transform (DWT), discrete cosine transform (DCT) and singular value decomposition (SVD) have shown an enhance efficiency in terms of ease and performance. It has manifested a good degree of robustness against various environment effects such as resizing, rotation and different kinds of noise. This method would prove very useful technique for copyright protection and ownership judgment.

Keywords: watermarking, ownership, copyright protection, steganography, information hiding, authentication

Procedia PDF Downloads 368
4644 [Keynote Talk]: New Generations and Employment: An Exploratory Study about Tensions between the Psycho-Social Characteristics of the Generation Z and Expectations and Actions of Organizational Structures Related with Employment (CABA, 2016)

Authors: Esteban Maioli

Abstract:

Generational studies have an important research tradition in social and human sciences. On the one hand, the speed of social change in the context of globalization imposes the need to research the transformations are identified both the subjectivity of the agents involved and its inclusion in the institutional matrix, specifically employment. Generation Z, (generally considered as the population group whose birth occurs after 1995) have unique psycho-social characteristics. Gen Z is characterized by a different set of values, beliefs, attitudes and ambitions that impact in their concrete action in organizational structures. On the other hand, managers often have to deal with generational differences in the workplace. Organizations have members who belong to different generations; they had never before faced the challenge of having such a diverse group of members. The members of each historical generation are characterized by a different set of values, beliefs, attitudes and ambitions that are manifest in their concrete action in organizational structures. Gen Z it’s the only one who can fully be considered "global," while its members were born in the consolidated context of globalization. Some salient features of the Generation Z can be summarized as follows. They’re the first fully born into a digital world. Social networks and technology are integrated into their lives. They are concerned about the challenges of the modern world (poverty, inequality, climate change, among others). They are self-expressive, more liberal and open to change. They often bore easily, with short attention spans. They do not like routine tasks. They want to achieve a good life-work balance, and they are interested in a flexible work environment, as opposed to traditional work schedule. They are critical thinkers, who come with innovative and creative ideas to help. Research design considered methodological triangulation. Data was collected with two techniques: a self-administered survey with multiple choice questions and attitudinal scales applied over a non-probabilistic sample by reasoned decision. According to the multi-method strategy, also it was conducted in-depth interviews. Organizations constantly face new challenges. One of the biggest ones is to learn to manage a multi-generational scope of work. While Gen Z has not yet been fully incorporated (expected to do so in five years or so), many organizations have already begun to implement a series of changes in its recruitment and development. The main obstacle to retaining young talent is the gap between the expectations of iGen applicants and what companies offer. Members of the iGen expect not only a good salary and job stability but also a clear career plan. Generation Z needs to have immediate feedback on their tasks. However, many organizations have yet to improve both motivation and monitoring practices. It is essential for companies to take a review of organizational practices anchored in the culture of the organization.

Keywords: employment, expectations, generation Z, organizational culture, organizations, psycho-social characteristics

Procedia PDF Downloads 201
4643 Spaces of Interpretation: Personal Space

Authors: Yehuda Roth

Abstract:

In quantum theory, a system’s time evolution is predictable unless an observer performs measurement, as the measurement process can randomize the system. This randomness appears when the measuring device does not accurately describe the measured item, i.e., when the states characterizing the measuring device appear as a superposition of those being measured. When such a mismatch occurs, the measured data randomly collapse into a single eigenstate of the measuring device. This scenario resembles the interpretation process in which the observer does not experience an objective reality but interprets it based on preliminary descriptions initially ingrained into his/her mind. This distinction is the motivation for the present study in which the collapse scenario is regarded as part of the interpretation process of the observer. By adopting the formalism of the quantum theory, we present a complete mathematical approach that describes the interpretation process. We demonstrate this process by applying the proposed interpretation formalism to the ambiguous image "My wife and mother-in-law" to identify whether a woman in the picture is young or old.

Keywords: quantum-like interpretation, ambiguous image, determination, quantum-like collapse, classified representation

Procedia PDF Downloads 104
4642 Modeling and Power Control of DFIG Used in Wind Energy System

Authors: Nadia Ben Si Ali, Nadia Benalia, Nora Zerzouri

Abstract:

Wind energy generation has attracted great interests in recent years. Doubly Fed Induction Generator (DFIG) for wind turbines are largely deployed because variable-speed wind turbines have many advantages over fixed-speed generation such as increased energy capture, operation at maximum power point, improved efficiency, and power quality. This paper presents the operation and vector control of a Doubly-fed Induction Generator (DFIG) system where the stator is connected directly to a stiff grid and the rotor is connected to the grid through bidirectional back-to-back AC-DC-AC converter. The basic operational characteristics, mathematical model of the aerodynamic system and vector control technique which is used to obtain decoupled control of powers are investigated using the software Mathlab/Simulink.

Keywords: wind turbine, Doubly Fed Induction Generator, wind speed controller, power system stability

Procedia PDF Downloads 379
4641 Development of a Data Security Model Using Steganography

Authors: Terungwa Simon Yange, Agana Moses A.

Abstract:

This paper studied steganography and designed a simplistic approach to a steganographic tool for hiding information in image files with the view of addressing the security challenges with data by hiding data from unauthorized users to improve its security. The Structured Systems Analysis and Design Method (SSADM) was used in this work. The system was developed using Java Development Kit (JDK) 1.7.0_10 and MySQL Server as its backend. The system was tested with some hypothetical health records which proved the possibility of protecting data from unauthorized users by making it secret so that its existence cannot be easily recognized by fraudulent users. It further strengthens the confidentiality of patient records kept by medical practitioners in the health setting. In conclusion, this work was able to produce a user friendly steganography software that is very fast to install and easy to operate to ensure privacy and secrecy of sensitive data. It also produced an exact copy of the original image and the one carrying the secret message when compared with each.

Keywords: steganography, cryptography, encryption, decryption, secrecy

Procedia PDF Downloads 266
4640 Simulation of Wind Solar Hybrid Power Generation for Pumping Station

Authors: Masoud Taghavi, Gholamreza Salehi, Ali Lohrasbi Nichkoohi

Abstract:

Despite the growing use of renewable energies in different fields of application of this technology in the field of water supply has been less attention. Photovoltaic and wind hybrid system is that new topics in renewable energy, including photovoltaic arrays, wind turbines, a set of batteries as a storage system and a diesel generator as a backup system is. In this investigation, first climate data including average wind speed and solar radiation at any time during the year, data collection and analysis are performed in the energy. The wind turbines in four models, photovoltaic panels at the 6 position of relative power, batteries and diesel generator capacity in seven states in the two models are combined hours of operation with renewables, diesel generator and battery bank check and a hybrid system of solar power generation-wind, which is optimized conditions, are presented.

Keywords: renewable energy, wind and solar energy, hybrid systems, cloning station

Procedia PDF Downloads 399