Search results for: machine readable format
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 3349

Search results for: machine readable format

979 Narcissism in the Life of Howard Hughes: A Psychobiographical Exploration

Authors: Alida Sandison, Louise A. Stroud

Abstract:

Narcissism is a personality configuration which has both normal and pathological personality expressions. Narcissism is highly complex, and is linked to a broad field of research. There are both dimensional and categorical conceptualisations of narcissism, and a variety of theoretical formulations that have been put forward to understand the narcissistic personality configuration. Currently, Kernberg’s Object Relations theory is well supported for this purpose. The complexity and particular defense mechanisms at play in the narcissistic personality make it a difficult personality configuration worth further research. Psychobiography as a methodology allows for the exploration of the lived life, and is thus a useful methodology to surmount these inherent challenges. Narcissism has been a focus of academic interest for a long time, and although there is a lot of research done in this area, to the researchers' knowledge, narcissistic dynamics have never been explored within a psychobiographical format. Thus, the primary aim of the research was to explore and describe narcissism in the life of Howard Hughes, with the objective of gaining further insight into narcissism through the use of this unconventional research approach. Hughes was chosen as subject for the study as he is renowned as an eccentric billionaire who had his revolutionary effect on the world, but was concurrently disturbed within his personal pathologies. Hughes was dynamic in three different sectors, namely motion pictures, aviation and gambling. He became more and more reclusive as he entered into middle age. From his early fifties he was agoraphobic, and the social network of connectivity that could reasonably be expected from someone in the top of their field was notably distorted. Due to his strong narcissistic personality configuration, and the interpersonal difficulties he experienced, Hughes represents an ideal figure to explore narcissism. The study used a single case study design, and purposive sampling to select Hughes. Qualitative data was sampled, using secondary data sources. Given that Hughes was a famous figure, there is a plethora of information on his life, which is primarily autobiographical. This includes books written about his life, and archival material in the form of newspaper articles, interviews and movies. Gathered data were triangulated to avoid the effect of author bias, and increase the credibility of the data used. It was collected using Yin’s guidelines for data collection. Data was analysed using Miles and Huberman strategy of data analysis, which consists of three steps, namely, data reduction, data display, and conclusion drawing and verification. Patterns which emerged in the data highlighted the defense mechanisms used by Hughes, in particular that of splitting and projection, in defending his sense of self. These defense mechanisms help us to understand the high levels of entitlement and paranoia experienced by Hughes. Findings provide further insight into his sense of isolation and difference, and the consequent difficulty he experienced in maintaining connections with others. Findings furthermore confirm the effectiveness of Kernberg’s theory in understanding narcissism observing an individual life.

Keywords: Howard Hughes, narcissism, narcissistic defenses, object relations

Procedia PDF Downloads 359
978 Mapping Iron Content in the Brain with Magnetic Resonance Imaging and Machine Learning

Authors: Gabrielle Robertson, Matthew Downs, Joseph Dagher

Abstract:

Iron deposition in the brain has been linked with a host of neurological disorders such as Alzheimer’s, Parkinson’s, and Multiple Sclerosis. While some treatment options exist, there are no objective measurement tools that allow for the monitoring of iron levels in the brain in vivo. An emerging Magnetic Resonance Imaging (MRI) method has been recently proposed to deduce iron concentration through quantitative measurement of magnetic susceptibility. This is a multi-step process that involves repeated modeling of physical processes via approximate numerical solutions. For example, the last two steps of this Quantitative Susceptibility Mapping (QSM) method involve I) mapping magnetic field into magnetic susceptibility and II) mapping magnetic susceptibility into iron concentration. Process I involves solving an ill-posed inverse problem by using regularization via injection of prior belief. The end result from Process II highly depends on the model used to describe the molecular content of each voxel (type of iron, water fraction, etc.) Due to these factors, the accuracy and repeatability of QSM have been an active area of research in the MRI and medical imaging community. This work aims to estimate iron concentration in the brain via a single step. A synthetic numerical model of the human head was created by automatically and manually segmenting the human head on a high-resolution grid (640x640x640, 0.4mm³) yielding detailed structures such as microvasculature and subcortical regions as well as bone, soft tissue, Cerebral Spinal Fluid, sinuses, arteries, and eyes. Each segmented region was then assigned tissue properties such as relaxation rates, proton density, electromagnetic tissue properties and iron concentration. These tissue property values were randomly selected from a Probability Distribution Function derived from a thorough literature review. In addition to having unique tissue property values, different synthetic head realizations also possess unique structural geometry created by morphing the boundary regions of different areas within normal physical constraints. This model of the human brain is then used to create synthetic MRI measurements. This is repeated thousands of times, for different head shapes, volume, tissue properties and noise realizations. Collectively, this constitutes a training-set that is similar to in vivo data, but larger than datasets available from clinical measurements. This 3D convolutional U-Net neural network architecture was used to train data-driven Deep Learning models to solve for iron concentrations from raw MRI measurements. The performance was then tested on both synthetic data not used in training as well as real in vivo data. Results showed that the model trained on synthetic MRI measurements is able to directly learn iron concentrations in areas of interest more effectively than other existing QSM reconstruction methods. For comparison, models trained on random geometric shapes (as proposed in the Deep QSM method) are less effective than models trained on realistic synthetic head models. Such an accurate method for the quantitative measurement of iron deposits in the brain would be of important value in clinical studies aiming to understand the role of iron in neurological disease.

Keywords: magnetic resonance imaging, MRI, iron deposition, machine learning, quantitative susceptibility mapping

Procedia PDF Downloads 138
977 Comparative Study of Impact Strength and Fracture Morphological of Nano-CaCO3 and Nanoclay Reinforced HDPE Nanocomposites

Authors: Harun Sepet, Necmettin Tarakcioglu

Abstract:

The present study investigated the impact strength and fracture mechanism of nano-CaCO3 and nanoclay reinforced HDPE nanocomposites by using Charpy impact test. The nano-CaCO3 and nanoclay reinforced HDPE granules were prepared by the melt blending method using a compounder system, which consists of industrial banbury mixer, single screw extruder and granule cutting in industrial-scale. The nano-CaCO3 and nanoclay reinforced HDPE granules were molded using an injection-molding machine as plates, and then impact samples were cut by using punching die from the nanocomposite plates. As a result of impact experiments, nano-CaCO3 and nanoclay reinforced HDPE nanocomposites were determined to have lower impact energy level than neat HDPE. Also, the impact strength of HDPE further decreased by addition nanoclay compared to nano-CaCO3. The occurred fracture areas with the impact were detected by SEM examination. It is understood that fracture surface morphology changes when nano-CaCO3 and nanoclay ratio increases. The fracture surface changes were examined to determine the fracture mechanism of nano-CaCO3 and nanoclay reinforced HDPE nanocomposites.

Keywords: charpy, HDPE, industrial scale nano-CaCO3, nanoclay, nanocomposite

Procedia PDF Downloads 412
976 Development of a Highly Flexible, Sensitive and Stretchable Polymer Nanocomposite for Strain Sensing

Authors: Shaghayegh Shajari, Mehdi Mahmoodi, Mahmood Rajabian, Uttandaraman Sundararaj, Les J. Sudak

Abstract:

Although several strain sensors based on carbon nanotubes (CNTs) have been reported, the stretchability and sensitivity of these sensors have remained as a challenge. Highly stretchable and sensitive strain sensors are in great demand for human motion monitoring and human-machine interface. This paper reports the fabrication and characterization of a new type of strain sensors based on a stretchable fluoropolymer / CNT nanocomposite system made via melt-mixing technique. Electrical and mechanical characterizations were obtained. The results showed that this nanocomposite sensor has high stretchability up to 280% of strain at an optimum level of filler concentration. The piezoresistive properties and the strain sensing mechanism of the strain sensor were investigated using Electrochemical Impedance Spectroscopy (EIS). High sensitivity was obtained (gauge factor as large as 12000 under 120% applied strain) in particular at the concentrations above the percolation threshold. Due to the tunneling effect, a non- linear piezoresistivity was observed at high concentrations of CNT loading. The nanocomposites with good conductivity and lightweight could be a promising candidate for strain sensing applications.

Keywords: carbon nanotubes, fluoropolymer, piezoresistive, strain sensor

Procedia PDF Downloads 296
975 Feature Extraction and Impact Analysis for Solid Mechanics Using Supervised Finite Element Analysis

Authors: Edward Schwalb, Matthias Dehmer, Michael Schlenkrich, Farzaneh Taslimi, Ketron Mitchell-Wynne, Horen Kuecuekyan

Abstract:

We present a generalized feature extraction approach for supporting Machine Learning (ML) algorithms which perform tasks similar to Finite-Element Analysis (FEA). We report results for estimating the Head Injury Categorization (HIC) of vehicle engine compartments across various impact scenarios. Our experiments demonstrate that models learned using features derived with a simple discretization approach provide a reasonable approximation of a full simulation. We observe that Decision Trees could be as effective as Neural Networks for the HIC task. The simplicity and performance of the learned Decision Trees could offer a trade-off of a multiple order of magnitude increase in speed and cost improvement over full simulation for a reasonable approximation. When used as a complement to full simulation, the approach enables rapid approximate feedback to engineering teams before submission for full analysis. The approach produces mesh independent features and is further agnostic of the assembly structure.

Keywords: mechanical design validation, FEA, supervised decision tree, convolutional neural network.

Procedia PDF Downloads 141
974 Deformation Behavior of Virgin and Polypropylene Modified Bituminous Mixture

Authors: Noor Zainab Habib, Ibrahim Kamaruddin, Madzlan Napiah

Abstract:

This paper present a part of research conducted to investigate the creep behavior of bituminous concrete mixture prepared with well graded using the dynamic creep test. The samples were prepared from unmodified control mix and Polypropylene modified bituminous mix. Unmodified or control mix was prepared with 80/100 grade bitumen while polypropylene modified mix was prepared using polypropylene PP polymer as modifier, blended with 80/100 Pen bitumen. The concentration of polymer in the blend was kept at 1%, 2%, and 3% by weight of bitumen content. For Dynamic Creep Test, Marshall Specimen were prepared at optimum bitumen content and then tested using IPC Global Universal Testing Machine (UTM), in order to investigate the creep stiffness of both modified and control mix. From the results obtained it was found that 1% and 2% PP modified bituminous mix offer better results in comparison to control and 3% PP modified mix samples. The results verify all the findings of empirical and viscosity test results which indicates that polymer modification induces stiffening effect in the binder. Enhanced viscous component of the binder was considered responsible for this change which eventually enhances the mechanical strength of the modified bituminous mixes.

Keywords: polymer modified bitumen, stiffness, creep, viscosity

Procedia PDF Downloads 419
973 Human Action Recognition Using Wavelets of Derived Beta Distributions

Authors: Neziha Jaouedi, Noureddine Boujnah, Mohamed Salim Bouhlel

Abstract:

In the framework of human machine interaction systems enhancement, we focus throw this paper on human behavior analysis and action recognition. Human behavior is characterized by actions and reactions duality (movements, psychological modification, verbal and emotional expression). It’s worth noting that many information is hidden behind gesture, sudden motion points trajectories and speeds, many research works reconstructed an information retrieval issues. In our work we will focus on motion extraction, tracking and action recognition using wavelet network approaches. Our contribution uses an analysis of human subtraction by Gaussian Mixture Model (GMM) and body movement through trajectory models of motion constructed from kalman filter. These models allow to remove the noise using the extraction of the main motion features and constitute a stable base to identify the evolutions of human activity. Each modality is used to recognize a human action using wavelets of derived beta distributions approach. The proposed approach has been validated successfully on a subset of KTH and UCF sports database.

Keywords: feautures extraction, human action classifier, wavelet neural network, beta wavelet

Procedia PDF Downloads 411
972 Hyperspectral Image Classification Using Tree Search Algorithm

Authors: Shreya Pare, Parvin Akhter

Abstract:

Remotely sensing image classification becomes a very challenging task owing to the high dimensionality of hyperspectral images. The pixel-wise classification methods fail to take the spatial structure information of an image. Therefore, to improve the performance of classification, spatial information can be integrated into the classification process. In this paper, the multilevel thresholding algorithm based on a modified fuzzy entropy function is used to perform the segmentation of hyperspectral images. The fuzzy parameters of the MFE function have been optimized by using a new meta-heuristic algorithm based on the Tree-Search algorithm. The segmented image is classified by a large distribution machine (LDM) classifier. Experimental results are shown on a hyperspectral image dataset. The experimental outputs indicate that the proposed technique (MFE-TSA-LDM) achieves much higher classification accuracy for hyperspectral images when compared to state-of-art classification techniques. The proposed algorithm provides accurate segmentation and classification maps, thus becoming more suitable for image classification with large spatial structures.

Keywords: classification, hyperspectral images, large distribution margin, modified fuzzy entropy function, multilevel thresholding, tree search algorithm, hyperspectral image classification using tree search algorithm

Procedia PDF Downloads 180
971 System and Method for Providing Web-Based Remote Application Service

Authors: Shuen-Tai Wang, Yu-Ching Lin, Hsi-Ya Chang

Abstract:

With the development of virtualization technologies, a new type of service named cloud computing service is produced. Cloud users usually encounter the problem of how to use the virtualized platform easily over the web without requiring the plug-in or installation of special software. The object of this paper is to develop a system and a method enabling process interfacing within an automation scenario for accessing remote application by using the web browser. To meet this challenge, we have devised a web-based interface that system has allowed to shift the GUI application from the traditional local environment to the cloud platform, which is stored on the remote virtual machine. We designed the sketch of web interface following the cloud virtualization concept that sought to enable communication and collaboration among users. We describe the design requirements of remote application technology and present implementation details of the web application and its associated components. We conclude that this effort has the potential to provide an elastic and resilience environment for several application services. Users no longer have to burden the system maintenances and reduce the overall cost of software licenses and hardware. Moreover, this remote application service represents the next step to the mobile workplace, and it lets user to use the remote application virtually from anywhere.

Keywords: virtualization technology, virtualized platform, web interface, remote application

Procedia PDF Downloads 289
970 Effect of Coffee Grounds on Physical and Heating Value Properties of Sugarcane Bagasse Pellets

Authors: K. Rattawan, W. Intagun, W. Kanoksilapatham

Abstract:

Objective of this research is to study effect of coffee grounds on physical and heating value properties of sugarcane bagasse pellets. The coffee grounds were tested as an additive for pelletizing process of bagasse pellets. Pelletizing was performed using a Flat–die pellet mill machine. Moisture content of raw materials was controlled at 10-13%. Die temperature range during the process was 75-80 oC. Physical characteristics (bulk density and durability) of the bagasse pellet and pellets with 1-5% coffee ground were determined following the standard assigned by the Pellet Fuel Institute (PFI). The results revealed increasing values of 648±3.4, 659 ± 3.1, 679 ± 3.3 and 685 ± 3.1 kg/m3 (for pellet bulk density); and 98.7 ± 0.11, 99.2 ± 0.26, 99.3 ± 0.19 and 99.4 ± 0.07% (for pellet durability), respectively. In addition, the heating values of the coffee ground supplemented pellets (15.9 ± 1.16, 17.0 ± 1.23 and 18.8 ± 1.34 MJ/kg) were improved comparing to the non-supplemented control (14.9 ± 1.14 MJ/kg), respectively. The results indicated that both the bulk density and durability values of the bagasse pellets were increased with the increasing proportion of the coffee ground additive.

Keywords: bagasse, coffee grounds, pelletizing, heating value, sugar cane bagasse

Procedia PDF Downloads 168
969 The Effect of Artificial Intelligence on Accounting and Finance

Authors: Evrime Fawzy Ishak Gadelsayed

Abstract:

This paper presents resource intake accounting as an inventive manner to cope with control accounting, which concentrates on administrators as the crucial customers of the information and offers satisfactory statistics of conventional control accounting. This machine underscores that the association's asset motivates prices; as a consequence, in costing frameworks, the emphasis ought to be on assets and their usage. Resource consumption accounting consolidates two costing methodologies, action-based totally and the German cost accounting approach called GPK. This methodology, however, is a danger to managers when making the management accounting undertaking operational. The motive for this article is to clarify the concept of resource intake accounting, its elements and highlights and use of this approach in associations. Inside the first area, we present useful resource consumption accounting, the basis, reasons for its improvement, and the issues that are faced beyond costing frameworks. At that point, we deliver the requirements and presumptions of this approach; ultimately, we depict the execution of this approach in associations and its preferences over other costing techniques.

Keywords: financial statement fraud, forensic accounting, fraud prevention and detection, auditing, audit expectation gap, corporate governance resource consumption accounting, management accounting, action based method, German cost accounting method

Procedia PDF Downloads 18
968 Intelligent System of the Grinding Robot for Spiral Welded Pipe

Authors: Getachew Demeissie Ayalew, Yongtao Sun, Yang Yang

Abstract:

The spiral welded pipe manufacturing industry requires strict production standards for automated grinders for welding seams. However, traditional grinding machines in this sector are insufficient due to a lack of quality control protocols and inconsistent performance. This research aims to improve the quality of spiral welded pipes by developing intelligent automated abrasive belt grinding equipment. The system has equipped with six degrees of freedom (6 DOF) KUKA KR360 industrial robots, enabling concurrent grinding operations on both internal and external welds. The grinding robot control system is designed with a PLC, and a human-machine interface (HMI) system is employed for operations. The system includes an electric speed controller, data connection card, DC driver, analog amplifier, and HMI for input data. This control system enables the grinding of spiral welded pipe. It ensures consistent production quality and cost-effectiveness by reducing the product life cycle and minimizing risks in the working environment.

Keywords: Intelligent Systems, Spiral Welded Pipe, Grinding, Industrial Robot, End-Effector, PLC Controller System, 3D Laser Sensor, HMI.

Procedia PDF Downloads 297
967 Experimental Investigation of Damaged Reinforced Concrete Beams Repaired with Carbon Fibre Reinforced Polymer (CFRP) Strip under Impact Loading

Authors: M. Al-Farttoosi, M. Y. Rafiq, J. Summerscales, C. Williams

Abstract:

Many buildings and bridges are damaged due to impact loading, explosions, terrorist attacks and wars. Most of the damaged structures members such as beams, columns and slabs are not totally failed and it can be repaired. Nowadays, carbon fibre reinforced polymer CFRP has been wildly used in strengthening and retrofitting the structures members. CFRP can rector the load carrying capacity of the damaged structures members to make them serviceable. An experimental investigation was conducted to investigate the impact behaviour of the damaged beams repaired with CFRP. The tested beams had different degrees of damage and near surface mounted technique NSM was used to install the CFRP. A heavy drop weight impact test machine was used to conduct the experimental work. The study investigated the impact strength, stiffness, cracks and deflection of the CFRP repaired beams. The results show that CFRP significantly increased the impact resistance of the damaged beams. CFRP increased the damaged beams stiffness and reduced the deflection. The results showed that the NSM technique is more effective in repairing beams and preventing the debonding of the CFRP.

Keywords: damaged, concrete, impact, repaired

Procedia PDF Downloads 345
966 Blend of Polyamide 6 with Polybutylene Terephthalate Compatibilized with Epoxidized Natural Rubber (ENR-25) and N Butyl Acrylate Glycidyl Methacrylate Ethylene (EBa-GMA)

Authors: Ramita Vongrat, Pornsri Sapsrithong, Manit Nithitanakul

Abstract:

In this work, blends of polyamide 6 (PA6) and polybutylene terephthalate (PBT) were successfully prepared. The effect of epoxidized natural rubber (ENR-25) and n butyl acrylate glycidyl methacrylate ethylene (EBa-GMA) as a compatibilizer on properties of PA6/PBT blends was also investigated by varying amount of ENR-50 and EBa-GMA, i.e., 0, 0.1, 0.5, 5 and 10 phr. All blends were prepared and shaped by using twin-screw extruder at 230 °C and injection molding machine, respectively. All test specimens were characterized by phase morphology, impact strength, tensile, flexural properties, and hardness. The results exhibited that phase morphology of PA6/PBT blend without compatibilizer was incompatible. This could be attributed to poor interfacial adhesion between the two polymers. SEM micrographs showed that the addition of ENR-25 and EBa-GMA improved the compatibility of PA6/PBT blends. With the addition of ENR-50 as a compatibilizer, the uniformity and the maximum reduction of dispersed phase size were observed. Additionally, the results indicate that, as the amount of ENR-25 increased, and EBa-GMA increased, the mechanical properties, including stress at the peak, tensile modulus, and izod impact strength, were also improved.

Keywords: EBa-GMA, epoxidized natural rubber-25, polyamide 6, polybutylene terephthalate

Procedia PDF Downloads 170
965 Synthetic Aperture Radar Remote Sensing Classification Using the Bag of Visual Words Model to Land Cover Studies

Authors: Reza Mohammadi, Mahmod R. Sahebi, Mehrnoosh Omati, Milad Vahidi

Abstract:

Classification of high resolution polarimetric Synthetic Aperture Radar (PolSAR) images plays an important role in land cover and land use management. Recently, classification algorithms based on Bag of Visual Words (BOVW) model have attracted significant interest among scholars and researchers in and out of the field of remote sensing. In this paper, BOVW model with pixel based low-level features has been implemented to classify a subset of San Francisco bay PolSAR image, acquired by RADARSAR 2 in C-band. We have used segment-based decision-making strategy and compared the result with the result of traditional Support Vector Machine (SVM) classifier. 90.95% overall accuracy of the classification with the proposed algorithm has shown that the proposed algorithm is comparable with the state-of-the-art methods. In addition to increase in the classification accuracy, the proposed method has decreased undesirable speckle effect of SAR images.

Keywords: Bag of Visual Words (BOVW), classification, feature extraction, land cover management, Polarimetric Synthetic Aperture Radar (PolSAR)

Procedia PDF Downloads 213
964 Application of Rapid Prototyping to Create Additive Prototype Using Computer System

Authors: Meftah O. Bashir, Fatma A. Karkory

Abstract:

Rapid prototyping is a new group of manufacturing processes, which allows fabrication of physical of any complexity using a layer by layer deposition technique directly from a computer system. The rapid prototyping process greatly reduces the time and cost necessary to bring a new product to market. The prototypes made by these systems are used in a range of industrial application including design evaluation, verification, testing, and as patterns for casting processes. These processes employ a variety of materials and mechanisms to build up the layers to build the part. The present work was to build a FDM prototyping machine that could control the X-Y motion and material deposition, to generate two-dimensional and three-dimensional complex shapes. This study focused on the deposition of wax material. This work was to find out the properties of the wax materials used in this work in order to enable better control of the FDM process. This study will look at the integration of a computer controlled electro-mechanical system with the traditional FDM additive prototyping process. The characteristics of the wax were also analysed in order to optimize the model production process. These included wax phase change temperature, wax viscosity and wax droplet shape during processing.

Keywords: rapid prototyping, wax, manufacturing processes, shape

Procedia PDF Downloads 466
963 Mechanism of Sinkhole Development on Water-Bearing Soft Ground Tunneling

Authors: H. J. Kim, K. H. Kim, N. H. Park, K. T. Nam, Y. H. Jung, T. H. Kim, J. H. Shin

Abstract:

Underground excavations in an urban area can cause various geotechnical problems such as ground loss and lowering of groundwater level. When the ground loss becomes uncontrollably large, sinkholes can be developed to the ground surface. A sinkhole is commonly known as the natural phenomenon associated with lime rock areas. However, sinkholes in urban areas due to pressurized sewers and/or tunneling are also frequently reported. In this study, mechanism of a sinkhole developed at the site ‘A’ where a tunneling work underwent is investigated. The sinkhole occurred in the sand strata with the high level of groundwater when excavating a tunnel of which diameter is 3.6 m. The sinkhole was progressed in two steps. The first step began with the local failure around the tunnel face followed by tons of groundwater inflow, and the second step was triggered by the TBM (Tunnel Boring Machine) chamber opening which led to the progressive general failure. The possibility of the sinkhole was evaluated by using Limit Equilibrium Method (LEM), and critical height was evaluated by the empirical stability chart. It is found that the lowering of the face pressure and inflow of groundwater into the tunnel face turned to be the main reason for the sinkhole.

Keywords: limit equilibrium method, sinkhole, stability chart, tunneling

Procedia PDF Downloads 254
962 Ensemble of Deep CNN Architecture for Classifying the Source and Quality of Teff Cereal

Authors: Belayneh Matebie, Michael Melese

Abstract:

The study focuses on addressing the challenges in classifying and ensuring the quality of Eragrostis Teff, a small and round grain that is the smallest cereal grain. Employing a traditional classification method is challenging because of its small size and the similarity of its environmental characteristics. To overcome this, this study employs a machine learning approach to develop a source and quality classification system for Teff cereal. Data is collected from various production areas in the Amhara regions, considering two types of cereal (high and low quality) across eight classes. A total of 5,920 images are collected, with 740 images for each class. Image enhancement techniques, including scaling, data augmentation, histogram equalization, and noise removal, are applied to preprocess the data. Convolutional Neural Network (CNN) is then used to extract relevant features and reduce dimensionality. The dataset is split into 80% for training and 20% for testing. Different classifiers, including FVGG16, FINCV3, QSCTC, EMQSCTC, SVM, and RF, are employed for classification, achieving accuracy rates ranging from 86.91% to 97.72%. The ensemble of FVGG16, FINCV3, and QSCTC using the Max-Voting approach outperforms individual algorithms.

Keywords: Teff, ensemble learning, max-voting, CNN, SVM, RF

Procedia PDF Downloads 57
961 Deep Learning and Accurate Performance Measure Processes for Cyber Attack Detection among Web Logs

Authors: Noureddine Mohtaram, Jeremy Patrix, Jerome Verny

Abstract:

As an enormous number of online services have been developed into web applications, security problems based on web applications are becoming more serious now. Most intrusion detection systems rely on each request to find the cyber-attack rather than on user behavior, and these systems can only protect web applications against known vulnerabilities rather than certain zero-day attacks. In order to detect new attacks, we analyze the HTTP protocols of web servers to divide them into two categories: normal attacks and malicious attacks. On the other hand, the quality of the results obtained by deep learning (DL) in various areas of big data has given an important motivation to apply it to cybersecurity. Deep learning for attack detection in cybersecurity has the potential to be a robust tool from small transformations to new attacks due to its capability to extract more high-level features. This research aims to take a new approach, deep learning to cybersecurity, to classify these two categories to eliminate attacks and protect web servers of the defense sector which encounters different web traffic compared to other sectors (such as e-commerce, web app, etc.). The result shows that by using a machine learning method, a higher accuracy rate, and a lower false alarm detection rate can be achieved.

Keywords: anomaly detection, HTTP protocol, logs, cyber attack, deep learning

Procedia PDF Downloads 213
960 Application of Data Mining Techniques for Tourism Knowledge Discovery

Authors: Teklu Urgessa, Wookjae Maeng, Joong Seek Lee

Abstract:

Application of five implementations of three data mining classification techniques was experimented for extracting important insights from tourism data. The aim was to find out the best performing algorithm among the compared ones for tourism knowledge discovery. Knowledge discovery process from data was used as a process model. 10-fold cross validation method is used for testing purpose. Various data preprocessing activities were performed to get the final dataset for model building. Classification models of the selected algorithms were built with different scenarios on the preprocessed dataset. The outperformed algorithm tourism dataset was Random Forest (76%) before applying information gain based attribute selection and J48 (C4.5) (75%) after selection of top relevant attributes to the class (target) attribute. In terms of time for model building, attribute selection improves the efficiency of all algorithms. Artificial Neural Network (multilayer perceptron) showed the highest improvement (90%). The rules extracted from the decision tree model are presented, which showed intricate, non-trivial knowledge/insight that would otherwise not be discovered by simple statistical analysis with mediocre accuracy of the machine using classification algorithms.

Keywords: classification algorithms, data mining, knowledge discovery, tourism

Procedia PDF Downloads 295
959 A Cloud-Based Federated Identity Management in Europe

Authors: Jesus Carretero, Mario Vasile, Guillermo Izquierdo, Javier Garcia-Blas

Abstract:

Currently, there is a so called ‘identity crisis’ in cybersecurity caused by the substantial security, privacy and usability shortcomings encountered in existing systems for identity management. Federated Identity Management (FIM) could be solution for this crisis, as it is a method that facilitates management of identity processes and policies among collaborating entities without enforcing a global consistency, that is difficult to achieve when there are ID legacy systems. To cope with this problem, the Connecting Europe Facility (CEF) initiative proposed in 2014 a federated solution in anticipation of the adoption of the Regulation (EU) N°910/2014, the so-called eIDAS Regulation. At present, a network of eIDAS Nodes is being deployed at European level to allow that every citizen recognized by a member state is to be recognized within the trust network at European level, enabling the consumption of services in other member states that, until now were not allowed, or whose concession was tedious. This is a very ambitious approach, since it tends to enable cross-border authentication of Member States citizens without the need to unify the authentication method (eID Scheme) of the member state in question. However, this federation is currently managed by member states and it is initially applied only to citizens and public organizations. The goal of this paper is to present the results of a European Project, named eID@Cloud, that focuses on the integration of eID in 5 cloud platforms belonging to authentication service providers of different EU Member States to act as Service Providers (SP) for private entities. We propose an initiative based on a private eID Scheme both for natural and legal persons. The methodology followed in the eID@Cloud project is that each Identity Provider (IdP) is subscribed to an eIDAS Node Connector, requesting for authentication, that is subscribed to an eIDAS Node Proxy Service, issuing authentication assertions. To cope with high loads, load balancing is supported in the eIDAS Node. The eID@Cloud project is still going on, but we already have some important outcomes. First, we have deployed the federation identity nodes and tested it from the security and performance point of view. The pilot prototype has shown the feasibility of deploying this kind of systems, ensuring good performance due to the replication of the eIDAS nodes and the load balance mechanism. Second, our solution avoids the propagation of identity data out of the native domain of the user or entity being identified, which avoids problems well known in cybersecurity due to network interception, man in the middle attack, etc. Last, but not least, this system allows to connect any country or collectivity easily, providing incremental development of the network and avoiding difficult political negotiations to agree on a single authentication format (which would be a major stopper).

Keywords: cybersecurity, identity federation, trust, user authentication

Procedia PDF Downloads 167
958 Fuzzy Population-Based Meta-Heuristic Approaches for Attribute Reduction in Rough Set Theory

Authors: Mafarja Majdi, Salwani Abdullah, Najmeh S. Jaddi

Abstract:

One of the global combinatorial optimization problems in machine learning is feature selection. It concerned with removing the irrelevant, noisy, and redundant data, along with keeping the original meaning of the original data. Attribute reduction in rough set theory is an important feature selection method. Since attribute reduction is an NP-hard problem, it is necessary to investigate fast and effective approximate algorithms. In this paper, we proposed two feature selection mechanisms based on memetic algorithms (MAs) which combine the genetic algorithm with a fuzzy record to record travel algorithm and a fuzzy controlled great deluge algorithm to identify a good balance between local search and genetic search. In order to verify the proposed approaches, numerical experiments are carried out on thirteen datasets. The results show that the MAs approaches are efficient in solving attribute reduction problems when compared with other meta-heuristic approaches.

Keywords: rough set theory, attribute reduction, fuzzy logic, memetic algorithms, record to record algorithm, great deluge algorithm

Procedia PDF Downloads 456
957 Study of Heat Exchangers in Small Modular Reactors

Authors: Harish Aryal, Roger Hague, Daniel Sotelo, Felipe Astete Salinas

Abstract:

This paper presents a comparative study of different coolants, materials, and temperatures that can affect the effectiveness of heat exchangers that are used in small modular reactors. The corrugated plate heat exchangers were chosen out of different plate options for testing purposes because of their ease of access and better performance than other existing heat exchangers in recent years. SolidWorks enables us to see various results between water coolants and helium coolants acting upon different types of conducting metals, which were selected from different fluids that ultimately satisfied accessibility requirements and were compatible with the software. Though not every element, material, fluid, or method was used in the testing phase, their purpose is to help further research that is to come since the innovation of nuclear power is the future. The tests that were performed are to help better understand the constant necessities that are seen in heat exchangers and through every adjustment see what the breaking points or improvements in the machine are. Depending on consumers and researchers, the results may give further feedback as to show why different types of materials and fluids would be preferred and why it is necessary to keep failures to improve future research.

Keywords: heat exchangers, Solidworks, coolants, small modular reactors, nuclear power, nanofluids, Nusselt number, friction factor, Reynolds number

Procedia PDF Downloads 76
956 One-Step Time Series Predictions with Recurrent Neural Networks

Authors: Vaidehi Iyer, Konstantin Borozdin

Abstract:

Time series prediction problems have many important practical applications, but are notoriously difficult for statistical modeling. Recently, machine learning methods have been attracted significant interest as a practical tool applied to a variety of problems, even though developments in this field tend to be semi-empirical. This paper explores application of Long Short Term Memory based Recurrent Neural Networks to the one-step prediction of time series for both trend and stochastic components. Two types of data are analyzed - daily stock prices, that are often considered to be a typical example of a random walk, - and weather patterns dominated by seasonal variations. Results from both analyses are compared, and reinforced learning framework is used to select more efficient between Recurrent Neural Networks and more traditional auto regression methods. It is shown that both methods are able to follow long-term trends and seasonal variations closely, but have difficulties with reproducing day-to-day variability. Future research directions and potential real world applications are briefly discussed.

Keywords: long short term memory, prediction methods, recurrent neural networks, reinforcement learning

Procedia PDF Downloads 233
955 Comparison of Different k-NN Models for Speed Prediction in an Urban Traffic Network

Authors: Seyoung Kim, Jeongmin Kim, Kwang Ryel Ryu

Abstract:

A database that records average traffic speeds measured at five-minute intervals for all the links in the traffic network of a metropolitan city. While learning from this data the models that can predict future traffic speed would be beneficial for the applications such as the car navigation system, building predictive models for every link becomes a nontrivial job if the number of links in a given network is huge. An advantage of adopting k-nearest neighbor (k-NN) as predictive models is that it does not require any explicit model building. Instead, k-NN takes a long time to make a prediction because it needs to search for the k-nearest neighbors in the database at prediction time. In this paper, we investigate how much we can speed up k-NN in making traffic speed predictions by reducing the amount of data to be searched for without a significant sacrifice of prediction accuracy. The rationale behind this is that we had a better look at only the recent data because the traffic patterns not only repeat daily or weekly but also change over time. In our experiments, we build several different k-NN models employing different sets of features which are the current and past traffic speeds of the target link and the neighbor links in its up/down-stream. The performances of these models are compared by measuring the average prediction accuracy and the average time taken to make a prediction using various amounts of data.

Keywords: big data, k-NN, machine learning, traffic speed prediction

Procedia PDF Downloads 364
954 Determining the Width and Depths of Cut in Milling on the Basis of a Multi-Dexel Model

Authors: Jens Friedrich, Matthias A. Gebele, Armin Lechler, Alexander Verl

Abstract:

Chatter vibrations and process instabilities are the most important factors limiting the productivity of the milling process. Chatter can leads to damage of the tool, the part or the machine tool. Therefore, the estimation and prediction of the process stability is very important. The process stability depends on the spindle speed, the depth of cut and the width of cut. In milling, the process conditions are defined in the NC-program. While the spindle speed is directly coded in the NC-program, the depth and width of cut are unknown. This paper presents a new simulation based approach for the prediction of the depth and width of cut of a milling process. The prediction is based on a material removal simulation with an analytically represented tool shape and a multi-dexel approach for the work piece. The new calculation method allows the direct estimation of the depth and width of cut, which are the influencing parameters of the process stability, instead of the removed volume as existing approaches do. The knowledge can be used to predict the stability of new, unknown parts. Moreover with an additional vibration sensor, the stability lobe diagram of a milling process can be estimated and improved based on the estimated depth and width of cut.

Keywords: dexel, process stability, material removal, milling

Procedia PDF Downloads 525
953 Flo: Period-Tracking App with AI Powered Tools

Authors: Dania Baaboud, Renad Al-zahrani, Mahnoor Khan, Riya Afroz

Abstract:

Flo is a smart period-tracking tool that uses artificial intelligence (AI) to offer individualized reproductive health predictions and insights. Flo makes very accurate predictions about menstrual cycles, ovulation, and fertility windows by evaluating user inputs, including cycle duration, symptoms, and patterns. Its machine learning algorithms are constantly evolving, providing personalized health recommendations, instructional materials, and early identification of possible health abnormalities such as reproductive problems and hormone imbalances. Flo, which was introduced in 2015 and upgraded with AI in 2017, is a revolutionary use of technology in healthcare that empowers people to make knowledgeable decisions regarding their well-being. Despite its advantages, our study included drawbacks, such as limited access to premium services and a small sample size. While highlighting unique characteristics, a comparative comparison with similar applications such as Clue and Glow confirmed Flo's outstanding AI integration for individualized healthcare. All things considered, Flo is a prime example of how AI can be used to tackle intricate biological processes, giving consumers the ability to efficiently control their reproductive health and opening the door for improvements in individualized medical technology.

Keywords: Flo, period-tracking app, period symptoms, women’s health, machinery

Procedia PDF Downloads 7
952 The Development of Monk’s Food Bowl Production on Occupational Health Safety and Environment at Work for the Strength of Rattanakosin Local Wisdom

Authors: Thammarak Srimarut, Witthaya Mekhum

Abstract:

This study analysed and developed a model for monk’s food bowl production on occupational health safety and environment at work for the encouragement of Rattanakosin local wisdom at Banbart Community. The process of blowpipe welding was necessary to produce the bowl which was very dangerous or 93.59% risk. After the employment of new sitting posture, the work risk was lower 48.41% or moderate risk. When considering in details, it was found that: 1) the traditional sitting posture could create work risk at 88.89% while the new sitting posture could create the work risk at 58.86%. 2) About the environmental pollution, with the traditional sitting posture, workers exposed to the polluted fume from welding at 61.11% while with the new sitting posture workers exposed to the polluted fume from welding at 40.47%. 3) On accidental risk, with the traditional sitting posture, workers exposed to the accident from welding at 94.44% while with the new sitting posture workers exposed to the accident from welding at 62.54%.

Keywords: occupational health safety, environment at work, Monk’s food bowl, machine intelligence

Procedia PDF Downloads 440
951 Using Bidirectional Encoder Representations from Transformers to Extract Topic-Independent Sentiment Features for Social Media Bot Detection

Authors: Maryam Heidari, James H. Jones Jr.

Abstract:

Millions of online posts about different topics and products are shared on popular social media platforms. One use of this content is to provide crowd-sourced information about a specific topic, event or product. However, this use raises an important question: what percentage of information available through these services is trustworthy? In particular, might some of this information be generated by a machine, i.e., a bot, instead of a human? Bots can be, and often are, purposely designed to generate enough volume to skew an apparent trend or position on a topic, yet the consumer of such content cannot easily distinguish a bot post from a human post. In this paper, we introduce a model for social media bot detection which uses Bidirectional Encoder Representations from Transformers (Google Bert) for sentiment classification of tweets to identify topic-independent features. Our use of a Natural Language Processing approach to derive topic-independent features for our new bot detection model distinguishes this work from previous bot detection models. We achieve 94\% accuracy classifying the contents of data as generated by a bot or a human, where the most accurate prior work achieved accuracy of 92\%.

Keywords: bot detection, natural language processing, neural network, social media

Procedia PDF Downloads 116
950 Internal Power Recovery in Cryogenic Cooling Plants Part I: Expander Development

Authors: Ambra Giovannelli, Erika Maria Archilei

Abstract:

The amount of the electrical power required by refrigeration systems is relevant worldwide. It is evaluated in the order of 15% of the total electricity production taking refrigeration and air-conditioning into consideration. For this reason, in the last years several energy saving techniques have been proposed to reduce the power demand of such plants. The paper deals with the development of an innovative internal recovery system for cryogenic cooling plants. Such a system consists in a Compressor-Expander Group (CEG) designed on the basis of the automotive turbocharging technology. In particular, the paper is focused on the design of the expander, the critical component of the CEG system. Due to the low volumetric flow entering the expander and the high expansion ratio, a commercial turbocharger expander wheel was strongly modified. It was equipped with a transonic nozzle, designed to have a radially inflow full admission. To verify the performance of such a machine and suggest improvements, two different set of nozzles have been designed and modelled by means of the commercial Ansys-CFX software. steady-state 3D CFD simulations of the second-generation prototype are presented and compared with the initial ones.

Keywords: vapour cCompression systems, energy saving, refrigeration plant, organic fluids, radial turbine

Procedia PDF Downloads 208