Search results for: algorithms and data structure
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 32292

Search results for: algorithms and data structure

31092 The Organizational Structure, Development Features, and Metadiscoursal Elements in the Expository Writing of College Freshman Students

Authors: Lota Largavista

Abstract:

This study entitled, ‘The Organizational Structure, Development Features, and Metadiscoursal Elements in the Expository Writing of Freshman College Writers’ aimed to examine essays written by college students. It seeks to examine the organizational structure and development features of the essays and describe their defining characteristics, the linguistic elements at both macrostructural and microstructural discourse levels and the types of textual and interpersonal metadiscourse markers that are employed in order to negotiate meanings with their prospective readers. The different frameworks used to analyze the essays include Toulmin’s ( 1984) model for argument structure, Olson’s ( 2003) three-part essay structure; Halliday and Matthiesen (2004) in Herriman (2011) notions of thematic structure, Danes (1974) thematic progression or method of development, Halliday’s (2004) concept of grammatical and lexical cohesion ;Hyland’s (2005) metadiscourse strategies; and Chung and Nation’s( 2003) four-step scale for technical vocabulary. This descriptive study analyzes qualitatively and quantitatively how freshman students generally express their written compositions. Coding of units is done to determine what linguistic features are present in the essays. Findings revealed that students’ expository essays observe a three-part structure having all three moves, the Introduction, the Body and the Conclusion. Stance assertion, stance support, and emerging moves/strategies are found to be employed in the essays. Students have more marked themes on the essays and also prefer constant theme progression as their method of development. The analysis of salient linguistic elements reveals frequently used cohesive devices and metadiscoursal strategies. Based on the findings, an instructional learning plan is being proposed. This plan is characterized by a genre approach that focuses on expository and linguistic conventions.

Keywords: metadiscourse, organization, theme progression, structure

Procedia PDF Downloads 244
31091 Half Mode Substrate Integrated Wave Guide of Band Pass Filter Based to Defected Ground Structure Cells

Authors: Damou Mehdi, Nouri Keltoum, Feham Mohammed, Khazini Mohammed, Bouazza Tayb Habibi Chawki

Abstract:

The Half mode SIW filter is treated by two softwares (HFSS (High Frequency Structure Simulator) and CST (Computer Simulation Technology)). The filter HMSIW has a very simple structure and a very compact size. The simulated results by CST are presented and compared with the results simulated by a high-frequency structure simulator. Good agreement between the simulated CST and simulated results by HFSS is observed. By cascading two of them according to design requirement, a X-band bandpass filter is designed and simulated to meet compact size, low insertion loss, good return loss as well as second harmonic suppression. As an example, we designed the proposed HMSIW filter at X band by HFSS. The filter has a pass-band from 7.3 GHz to 9.8 GHz, and its relative operating fraction bandwidth is 29.5 %. There are one transmission zeros are located at 14.4 GHz.

Keywords: substrate integrated waveguide, filter, HMSIW, defected ground structures (DGS), simulation BPF

Procedia PDF Downloads 591
31090 Engineering Topology of Ecological Model for Orientation Impact of Sustainability Urban Environments: The Spatial-Economic Modeling

Authors: Moustafa Osman Mohammed

Abstract:

The modeling of a spatial-economic database is crucial in recitation economic network structure to social development. Sustainability within the spatial-economic model gives attention to green businesses to comply with Earth’s Systems. The natural exchange patterns of ecosystems have consistent and periodic cycles to preserve energy and materials flow in systems ecology. When network topology influences formal and informal communication to function in systems ecology, ecosystems are postulated to valence the basic level of spatial sustainable outcome (i.e., project compatibility success). These referred instrumentalities impact various aspects of the second level of spatial sustainable outcomes (i.e., participant social security satisfaction). The sustainability outcomes are modeling composite structure based on a network analysis model to calculate the prosperity of panel databases for efficiency value, from 2005 to 2025. The database is modeling spatial structure to represent state-of-the-art value-orientation impact and corresponding complexity of sustainability issues (e.g., build a consistent database necessary to approach spatial structure; construct the spatial-economic-ecological model; develop a set of sustainability indicators associated with the model; allow quantification of social, economic and environmental impact; use the value-orientation as a set of important sustainability policy measures), and demonstrate spatial structure reliability. The structure of spatial-ecological model is established for management schemes from the perspective pollutants of multiple sources through the input–output criteria. These criteria evaluate the spillover effect to conduct Monte Carlo simulations and sensitivity analysis in a unique spatial structure. The balance within “equilibrium patterns,” such as collective biosphere features, has a composite index of many distributed feedback flows. The following have a dynamic structure related to physical and chemical properties for gradual prolong to incremental patterns. While these spatial structures argue from ecological modeling of resource savings, static loads are not decisive from an artistic/architectural perspective. The model attempts to unify analytic and analogical spatial structure for the development of urban environments in a relational database setting, using optimization software to integrate spatial structure where the process is based on the engineering topology of systems ecology.

Keywords: ecological modeling, spatial structure, orientation impact, composite index, industrial ecology

Procedia PDF Downloads 71
31089 Regenerated Cellulose Prepared by Using NaOH/Urea

Authors: Lee Chiau Yeng, Norhayani Othman

Abstract:

Regenerated cellulose fiber is fabricated in the NaOH/urea aqueous solution. In this work, cellulose is dissolved in 7 .wt% NaOH/12 .wt% urea in the temperature of -12 °C to prepare regenerated cellulose. Thermal and structure properties of cellulose and regenerated cellulose was compared and investigated by Field Emission Scanning Electron Microscopy (FeSEM), Fourier Transform Infrared Spectroscopy (FTIR), X-Ray Diffraction (XRD), Thermogravimetric analysis (TGA), and Differential Scanning Calorimetry. Results of FeSEM revealed that the regenerated cellulose fibers showed a more circular shape with irregular size due to fiber agglomeration. FTIR showed the difference in between the structure of cellulose and the regenerated cellulose fibers. In this case, regenerated cellulose fibers have a cellulose II crystalline structure with lower degree of crystallinity. Regenerated cellulose exhibited better thermal stability than the cellulose.

Keywords: regenerated cellulose, cellulose, NaOH, urea

Procedia PDF Downloads 435
31088 Examining Motivational Dynamics and L2 Learning Transitions of Air Cadets Between Year One and Year Two: A Retrodictive Qualitative Modelling Approach

Authors: Kanyaporn Sommeechai

Abstract:

Air cadets who aspire to become military pilots upon graduation undergo rigorous training at military academies. As first-year cadets are akin to civilian freshmen, they encounter numerous challenges within the seniority-based military academy system. Imposed routines, such as mandatory morning runs and restrictions on mobile phone usage for two semesters, have the potential to impact their learning process and motivation to study, including second language (L2) acquisition. This study aims to investigate the motivational dynamics and L2 learning transitions experienced by air cadets. To achieve this, a Retrodictive Qualitative Modelling approach will be employed, coupled with the adaptation of the three-barrier structure encompassing institutional factors, situational factors, and dispositional factors. Semi-structured interviews will be conducted to gather rich qualitative data. By analyzing and interpreting the collected data, this research seeks to shed light on the motivational factors that influence air cadets' L2 learning journey. The three-barrier structure will provide a comprehensive framework to identify and understand the institutional, situational, and dispositional factors that may impede or facilitate their motivation and language learning progress. Moreover, the study will explore how these factors interact and shape cadets' motivation and learning experiences. The outcomes of this research will yield fundamental data that can inform strategies and interventions to enhance the motivation and language learning outcomes of air cadets. By better understanding their motivational dynamics and transitions, educators and institutions can create targeted initiatives, tailored pedagogical approaches, and supportive environments that effectively inspire and engage air cadets as L2 learners.

Keywords: second language, education, motivational dynamics, learning transitions

Procedia PDF Downloads 75
31087 AI-Enabled Smart Contracts for Reliable Traceability in the Industry 4.0

Authors: Harris Niavis, Dimitra Politaki

Abstract:

The manufacturing industry was collecting vast amounts of data for monitoring product quality thanks to the advances in the ICT sector and dedicated IoT infrastructure is deployed to track and trace the production line. However, industries have not yet managed to unleash the full potential of these data due to defective data collection methods and untrusted data storage and sharing. Blockchain is gaining increasing ground as a key technology enabler for Industry 4.0 and the smart manufacturing domain, as it enables the secure storage and exchange of data between stakeholders. On the other hand, AI techniques are more and more used to detect anomalies in batch and time-series data that enable the identification of unusual behaviors. The proposed scheme is based on smart contracts to enable automation and transparency in the data exchange, coupled with anomaly detection algorithms to enable reliable data ingestion in the system. Before sensor measurements are fed to the blockchain component and the smart contracts, the anomaly detection mechanism uniquely combines artificial intelligence models to effectively detect unusual values such as outliers and extreme deviations in data coming from them. Specifically, Autoregressive integrated moving average, Long short-term memory (LSTM) and Dense-based autoencoders, as well as Generative adversarial networks (GAN) models, are used to detect both point and collective anomalies. Towards the goal of preserving the privacy of industries' information, the smart contracts employ techniques to ensure that only anonymized pointers to the actual data are stored on the ledger while sensitive information remains off-chain. In the same spirit, blockchain technology guarantees the security of the data storage through strong cryptography as well as the integrity of the data through the decentralization of the network and the execution of the smart contracts by the majority of the blockchain network actors. The blockchain component of the Data Traceability Software is based on the Hyperledger Fabric framework, which lays the ground for the deployment of smart contracts and APIs to expose the functionality to the end-users. The results of this work demonstrate that such a system can increase the quality of the end-products and the trustworthiness of the monitoring process in the smart manufacturing domain. The proposed AI-enabled data traceability software can be employed by industries to accurately trace and verify records about quality through the entire production chain and take advantage of the multitude of monitoring records in their databases.

Keywords: blockchain, data quality, industry4.0, product quality

Procedia PDF Downloads 193
31086 Performance Comparison of Outlier Detection Techniques Based Classification in Wireless Sensor Networks

Authors: Ayadi Aya, Ghorbel Oussama, M. Obeid Abdulfattah, Abid Mohamed

Abstract:

Nowadays, many wireless sensor networks have been distributed in the real world to collect valuable raw sensed data. The challenge is to extract high-level knowledge from this huge amount of data. However, the identification of outliers can lead to the discovery of useful and meaningful knowledge. In the field of wireless sensor networks, an outlier is defined as a measurement that deviates from the normal behavior of sensed data. Many detection techniques of outliers in WSNs have been extensively studied in the past decade and have focused on classic based algorithms. These techniques identify outlier in the real transaction dataset. This survey aims at providing a structured and comprehensive overview of the existing researches on classification based outlier detection techniques as applicable to WSNs. Thus, we have identified key hypotheses, which are used by these approaches to differentiate between normal and outlier behavior. In addition, this paper tries to provide an easier and a succinct understanding of the classification based techniques. Furthermore, we identified the advantages and disadvantages of different classification based techniques and we presented a comparative guide with useful paradigms for promoting outliers detection research in various WSN applications and suggested further opportunities for future research.

Keywords: bayesian networks, classification-based approaches, KPCA, neural networks, one-class SVM, outlier detection, wireless sensor networks

Procedia PDF Downloads 500
31085 Customer Preference in the Textile Market: Fabric-Based Analysis

Authors: Francisca Margarita Ocran

Abstract:

Underwear, and more particularly bras and panties, are defined as intimate clothing. Strictly speaking, they enhance the place of women in the public or private satchel. Therefore, women's lingerie is a complex garment with a high involvement profile, motivating consumers to buy it not only by its functional utility but also by the multisensory experience it provides them. Customer behavior models are generally based on customer data mining, and each model is designed to answer questions at a specific time. Predicting the customer experience is uncertain and difficult. Thus, knowledge of consumers' tastes in lingerie deserves to be treated as an experiential product, where the dimensions of the experience motivating consumers to buy a lingerie product and to remain faithful to it must be analyzed in detail by the manufacturers and retailers to engage and retain consumers, which is why this research aims to identify the variables that push consumers to choose their lingerie product, based on an in-depth analysis of the types of fabrics used to make lingerie. The data used in this study comes from online purchases. Machine learning approach with the use of Python programming language and Pycaret gives us a precision of 86.34%, 85.98%, and 84.55% for the three algorithms to use concerning the preference of a buyer in front of a range of lingerie. Gradient Boosting, random forest, and K Neighbors were used in this study; they are very promising and rich in the classification of preference in the textile industry.

Keywords: consumer behavior, data mining, lingerie, machine learning, preference

Procedia PDF Downloads 94
31084 Quantification of Effects of Structure-Soil-Structure Interactions on Urban Environment under Rayleigh Wave Loading

Authors: Neeraj Kumar, J. P. Narayan

Abstract:

The effects of multiple Structure-Soil-Structure Interactions (SSSI) on the seismic wave-field is generally disregarded by earthquake engineers, particularly the surface waves which cause more damage to buildings. Closely built high rise buildings exchange substantial seismic energy with each other and act as a full-coupled dynamic system. In this paper, SSI effects on the building responses and the free field motion due to a small city consisting 25- homogenous buildings blocks of 10-storey are quantified. The rocking and translational behavior of building under Rayleigh wave loading is studied for different dimensions of the building. The obtained dynamic parameters of buildings revealed a reduction in building roof drift with an increase in number of buildings ahead of the considered building. The strain developed by vertical component of Rayleigh may cause tension in structural components of building. A matching of fundamental frequency of building for the horizontal component of Rayleigh wave with that for vertically incident SV-wave is obtained. Further, the fundamental frequency of building for the vertical vibration is approximately twice to that for horizontal vibration. The city insulation has caused a reduction of amplitude of Rayleigh wave up to 19.3% and 21.6% in the horizontal and vertical components, respectively just outside the city. Further, the insulating effect of city was very large at fundamental frequency of buildings for both the horizontal and vertical components. Therefore, it is recommended to consider the insulating effects of city falling in the path of Rayleigh wave propagation in seismic hazard assessment for an area.

Keywords: structure-soil-structure interactions, Rayleigh wave propagation, finite difference simulation, dynamic response of buildings

Procedia PDF Downloads 220
31083 A Virtual Grid Based Energy Efficient Data Gathering Scheme for Heterogeneous Sensor Networks

Authors: Siddhartha Chauhan, Nitin Kumar Kotania

Abstract:

Traditional Wireless Sensor Networks (WSNs) generally use static sinks to collect data from the sensor nodes via multiple forwarding. Therefore, network suffers with some problems like long message relay time, bottle neck problem which reduces the performance of the network. Many approaches have been proposed to prevent this problem with the help of mobile sink to collect the data from the sensor nodes, but these approaches still suffer from the buffer overflow problem due to limited memory size of sensor nodes. This paper proposes an energy efficient scheme for data gathering which overcomes the buffer overflow problem. The proposed scheme creates virtual grid structure of heterogeneous nodes. Scheme has been designed for sensor nodes having variable sensing rate. Every node finds out its buffer overflow time and on the basis of this cluster heads are elected. A controlled traversing approach is used by the proposed scheme in order to transmit data to sink. The effectiveness of the proposed scheme is verified by simulation.

Keywords: buffer overflow problem, mobile sink, virtual grid, wireless sensor networks

Procedia PDF Downloads 395
31082 Adaption of the Design Thinking Method for Production Planning in the Meat Industry Using Machine Learning Algorithms

Authors: Alica Höpken, Hergen Pargmann

Abstract:

The resource-efficient planning of the complex production planning processes in the meat industry and the reduction of food waste is a permanent challenge. The complexity of the production planning process occurs in every part of the supply chain, from agriculture to the end consumer. It arises from long and uncertain planning phases. Uncertainties such as stochastic yields, fluctuations in demand, and resource variability are part of this process. In the meat industry, waste mainly relates to incorrect storage, technical causes in production, or overproduction. The high amount of food waste along the complex supply chain in the meat industry could not be reduced by simple solutions until now. Therefore, resource-efficient production planning by conventional methods is currently only partially feasible. The realization of intelligent, automated production planning is basically possible through the application of machine learning algorithms, such as those of reinforcement learning. By applying the adapted design thinking method, machine learning methods (especially reinforcement learning algorithms) are used for the complex production planning process in the meat industry. This method represents a concretization to the application area. A resource-efficient production planning process is made available by adapting the design thinking method. In addition, the complex processes can be planned efficiently by using this method, since this standardized approach offers new possibilities in order to challenge the complexity and the high time consumption. It represents a tool to support the efficient production planning in the meat industry. This paper shows an elegant adaption of the design thinking method to apply the reinforcement learning method for a resource-efficient production planning process in the meat industry. Following, the steps that are necessary to introduce machine learning algorithms into the production planning of the food industry are determined. This is achieved based on a case study which is part of the research project ”REIF - Resource Efficient, Economic and Intelligent Food Chain” supported by the German Federal Ministry for Economic Affairs and Climate Action of Germany and the German Aerospace Center. Through this structured approach, significantly better planning results are achieved, which would be too complex or very time consuming using conventional methods.

Keywords: change management, design thinking method, machine learning, meat industry, reinforcement learning, resource-efficient production planning

Procedia PDF Downloads 132
31081 Image Compression on Region of Interest Based on SPIHT Algorithm

Authors: Sudeepti Dayal, Neelesh Gupta

Abstract:

Image abbreviation is utilized for reducing the size of a file without demeaning the quality of the image to an objectionable level. The depletion in file size permits more images to be deposited in a given number of spaces. It also minimizes the time necessary for images to be transferred. Storage of medical images is a most researched area in the current scenario. To store a medical image, there are two parameters on which the image is divided, regions of interest and non-regions of interest. The best way to store an image is to compress it in such a way that no important information is lost. Compression can be done in two ways, namely lossy, and lossless compression. Under that, several compression algorithms are applied. In the paper, two algorithms are used which are, discrete cosine transform, applied to non-region of interest (lossy), and discrete wavelet transform, applied to regions of interest (lossless). The paper introduces SPIHT (set partitioning hierarchical tree) algorithm which is applied onto the wavelet transform to obtain good compression ratio from which an image can be stored efficiently.

Keywords: Compression ratio, DWT, SPIHT, DCT

Procedia PDF Downloads 353
31080 An Application to Predict the Best Study Path for Information Technology Students in Learning Institutes

Authors: L. S. Chathurika

Abstract:

Early prediction of student performance is an important factor to be gained academic excellence. Whatever the study stream in secondary education, students lay the foundation for higher studies during the first year of their degree or diploma program in Sri Lanka. The information technology (IT) field has certain improvements in the education domain by selecting specialization areas to show the talents and skills of students. These specializations can be software engineering, network administration, database administration, multimedia design, etc. After completing the first-year, students attempt to select the best path by considering numerous factors. The purpose of this experiment is to predict the best study path using machine learning algorithms. Five classification algorithms: decision tree, support vector machine, artificial neural network, Naïve Bayes, and logistic regression are selected and tested. The support vector machine obtained the highest accuracy, 82.4%. Then affecting features are recognized to select the best study path.

Keywords: algorithm, classification, evaluation, features, testing, training

Procedia PDF Downloads 122
31079 Recurrent Neural Networks for Complex Survival Models

Authors: Pius Marthin, Nihal Ata Tutkun

Abstract:

Survival analysis has become one of the paramount procedures in the modeling of time-to-event data. When we encounter complex survival problems, the traditional approach remains limited in accounting for the complex correlational structure between the covariates and the outcome due to the strong assumptions that limit the inference and prediction ability of the resulting models. Several studies exist on the deep learning approach to survival modeling; moreover, the application for the case of complex survival problems still needs to be improved. In addition, the existing models need to address the data structure's complexity fully and are subject to noise and redundant information. In this study, we design a deep learning technique (CmpXRnnSurv_AE) that obliterates the limitations imposed by traditional approaches and addresses the above issues to jointly predict the risk-specific probabilities and survival function for recurrent events with competing risks. We introduce the component termed Risks Information Weights (RIW) as an attention mechanism to compute the weighted cumulative incidence function (WCIF) and an external auto-encoder (ExternalAE) as a feature selector to extract complex characteristics among the set of covariates responsible for the cause-specific events. We train our model using synthetic and real data sets and employ the appropriate metrics for complex survival models for evaluation. As benchmarks, we selected both traditional and machine learning models and our model demonstrates better performance across all datasets.

Keywords: cumulative incidence function (CIF), risk information weight (RIW), autoencoders (AE), survival analysis, recurrent events with competing risks, recurrent neural networks (RNN), long short-term memory (LSTM), self-attention, multilayers perceptrons (MLPs)

Procedia PDF Downloads 94
31078 Effect of the Applied Bias on Mini-Band Structures in Dimer Fibonacci InAs/Ga1-XInXAs Superlattices

Authors: Z. Aziz, S. Terkhi, Y. Sefir, R. Djelti, S. Bentata

Abstract:

The effect of a uniform electric field across multi-barrier systems (InAs/InxGa1-xAs) is exhaustively explored by a computational model using exact Airy function formalism and the transfer-matrix technique. In the case of biased DFHBSL structure a strong reduction in transmission properties was observed and the width of the mini-band structure linearly decreases with the increase of the applied bias. This is due to the confinement of the states in the mini-band structure, which becomes increasingly important (Wannier-Stark Effect).

Keywords: dimer fibonacci height barrier superlattices, singular extended state, exact Airy function and transfer matrix formalism, bioinformatics

Procedia PDF Downloads 293
31077 The Hyundai Model: A Self-Sufficient State like Entity Masquerading as a Company

Authors: Nikita Koradia

Abstract:

Hyundai Motor Company, which started off as a small fish in a big sea, paved its way out successfully and established itself as an independent group from the conglomerate. Hyundai, with its officious power across the globe and particularly in South Korea in the automobile industry, has one the most complex yet fascinating governance structure. Being the second largest contributor to the Gross Domestic Product of South Korea after Samsung and having a market share of 51.3% domestically in automobile industry, Hyundai has faced its part of criticism owing to its anti-labor union approach and owing to its internalization of supply chain management. The censure has been coming from across jurisdictions like China, India, Canada, the EU, etc. The paper focuses on the growth of Hyundai and its inward and outward investment structure. The paper questions the ability of Hyundai to become a mini-state in itself by focusing on its governance structure. The paper further elaborates on its compliance and disclosure regime in the field of Corporate social responsibility and explores how far the business structure adopted by Hyundai works in its favor to become one of the leading automobile contenders in the market.

Keywords: compliance regime, disclosure regime, Hyundai motor company, supply-chain management

Procedia PDF Downloads 123
31076 Compact 3-D Co-Planar Waveguide Fed Dual-Port Ultrawideband-Multiple-Input and Multiple-Output Antenna with WLAN Band-Notched Characteristics

Authors: Asim Quddus

Abstract:

A miniaturized three dimensional co-planar waveguide (CPW) two-port MIMO antenna, exhibiting high isolation and WLAN band-notched characteristics is presented in this paper for ultrawideband (UWB) communication applications. The microstrip patch antenna operates as a single UWB antenna element. The proposed design is a cuboid-shaped structure having compact size of 35 x 27 x 45 mm³. Radiating as well as decoupling structure is placed around cuboidal polystyrene sheet. The radiators are 27 mm apart, placed Face-to-Face in vertical direction. Decoupling structure is placed on the side walls of polystyrene. The proposed antenna consists of an oval shaped radiating patch. A rectangular structure with fillet edges is placed on ground plan to enhance the bandwidth. The proposed antenna exhibits a good impedance match (S11 ≤ -10 dB) over frequency band of 2 GHz – 10.6 GHz. A circular slotted structure is employed as a decoupling structure on substrate, and it is placed on the side walls of polystyrene to enhance the isolation between antenna elements. Moreover, to achieve immunity from WLAN band distortion, a modified, inverted crescent shaped slotted structure is etched on radiating patches to achieve band-rejection characteristics at WLAN frequency band 4.8 GHz – 5.2 GHz. The suggested decoupling structure provides isolation better than 15 dB over the desired UWB spectrum. The envelope correlation coefficient (ECC) and gain for the MIMO antenna are analyzed as well. Finite Element Method (FEM) simulations are carried out in Ansys High Frequency Structural Simulator (HFSS) for the proposed design. The antenna is realized on a Rogers RT/duroid 5880 with thickness 1 mm, relative permittivity ɛr = 2.2. The proposed antenna achieves a stable omni-directional radiation patterns as well, while providing rejection at desired WLAN band. The S-parameters as well as MIMO parameters like ECC are analyzed and the results show conclusively that the design is suitable for portable MIMO-UWB applications.

Keywords: 3-D antenna, band-notch, MIMO, UWB

Procedia PDF Downloads 299
31075 Automated Prediction of HIV-associated Cervical Cancer Patients Using Data Mining Techniques for Survival Analysis

Authors: O. J. Akinsola, Yinan Zheng, Rose Anorlu, F. T. Ogunsola, Lifang Hou, Robert Leo-Murphy

Abstract:

Cervical Cancer (CC) is the 2nd most common cancer among women living in low and middle-income countries, with no associated symptoms during formative periods. With the advancement and innovative medical research, there are numerous preventive measures being utilized, but the incidence of cervical cancer cannot be truncated with the application of only screening tests. The mortality associated with this invasive cervical cancer can be nipped in the bud through the important role of early-stage detection. This study research selected an array of different top features selection techniques which was aimed at developing a model that could validly diagnose the risk factors of cervical cancer. A retrospective clinic-based cohort study was conducted on 178 HIV-associated cervical cancer patients in Lagos University teaching Hospital, Nigeria (U54 data repository) in April 2022. The outcome measure was the automated prediction of the HIV-associated cervical cancer cases, while the predictor variables include: demographic information, reproductive history, birth control, sexual history, cervical cancer screening history for invasive cervical cancer. The proposed technique was assessed with R and Python programming software to produce the model by utilizing the classification algorithms for the detection and diagnosis of cervical cancer disease. Four machine learning classification algorithms used are: the machine learning model was split into training and testing dataset into ratio 80:20. The numerical features were also standardized while hyperparameter tuning was carried out on the machine learning to train and test the data. Logistic Regression (LR), Decision Tree (DT), Random Forest (RF), and K-Nearest Neighbor (KNN). Some fitting features were selected for the detection and diagnosis of cervical cancer diseases from selected characteristics in the dataset using the contribution of various selection methods for the classification cervical cancer into healthy or diseased status. The mean age of patients was 49.7±12.1 years, mean age at pregnancy was 23.3±5.5 years, mean age at first sexual experience was 19.4±3.2 years, while the mean BMI was 27.1±5.6 kg/m2. A larger percentage of the patients are Married (62.9%), while most of them have at least two sexual partners (72.5%). Age of patients (OR=1.065, p<0.001**), marital status (OR=0.375, p=0.011**), number of pregnancy live-births (OR=1.317, p=0.007**), and use of birth control pills (OR=0.291, p=0.015**) were found to be significantly associated with HIV-associated cervical cancer. On top ten 10 features (variables) considered in the analysis, RF claims the overall model performance, which include: accuracy of (72.0%), the precision of (84.6%), a recall of (84.6%) and F1-score of (74.0%) while LR has: an accuracy of (74.0%), precision of (70.0%), recall of (70.0%) and F1-score of (70.0%). The RF model identified 10 features predictive of developing cervical cancer. The age of patients was considered as the most important risk factor, followed by the number of pregnancy livebirths, marital status, and use of birth control pills, The study shows that data mining techniques could be used to identify women living with HIV at high risk of developing cervical cancer in Nigeria and other sub-Saharan African countries.

Keywords: associated cervical cancer, data mining, random forest, logistic regression

Procedia PDF Downloads 89
31074 Optimized Electron Diffraction Detection and Data Acquisition in Diffraction Tomography: A Complete Solution by Gatan

Authors: Saleh Gorji, Sahil Gulati, Ana Pakzad

Abstract:

Continuous electron diffraction tomography, also known as microcrystal electron diffraction (MicroED) or three-dimensional electron diffraction (3DED), is a powerful technique, which in combination with cryo-electron microscopy (cryo-ED), can provide atomic-scale 3D information about the crystal structure and composition of different classes of crystalline materials such as proteins, peptides, and small molecules. Unlike the well-established X-ray crystallography method, 3DED does not require large single crystals and can collect accurate electron diffraction data from crystals as small as 50 – 100 nm. This is a critical advantage as growing larger crystals, as required by X-ray crystallography methods, is often very difficult, time-consuming, and expensive. In most cases, specimens studied via 3DED method are electron beam sensitive, which means there is a limitation on the maximum amount of electron dose one can use to collect the required data for a high-resolution structure determination. Therefore, collecting data using a conventional scintillator-based fiber coupled camera brings additional challenges. This is because of the inherent noise introduced during the electron-to-photon conversion in the scintillator and transfer of light via the fibers to the sensor, which results in a poor signal-to-noise ratio and requires a relatively higher and commonly specimen-damaging electron dose rates, especially for protein crystals. As in other cryo-EM techniques, damage to the specimen can be mitigated if a direct detection camera is used which provides a high signal-to-noise ratio at low electron doses. In this work, we have used two classes of such detectors from Gatan, namely the K3® camera (a monolithic active pixel sensor) and Stela™ (that utilizes DECTRIS hybrid-pixel technology), to address this problem. The K3 is an electron counting detector optimized for low-dose applications (like structural biology cryo-EM), and Stela is also a counting electron detector but optimized for diffraction applications with high speed and high dynamic range. Lastly, data collection workflows, including crystal screening, microscope optics setup (for imaging and diffraction), stage height adjustment at each crystal position, and tomogram acquisition, can be one of the other challenges of the 3DED technique. Traditionally this has been all done manually or in a partly automated fashion using open-source software and scripting, requiring long hours on the microscope (extra cost) and extensive user interaction with the system. We have recently introduced Latitude® D in DigitalMicrograph® software, which is compatible with all pre- and post-energy-filter Gatan cameras and enables 3DED data acquisition in an automated and optimized fashion. Higher quality 3DED data enables structure determination with higher confidence, while automated workflows allow these to be completed considerably faster than before. Using multiple examples, this work will demonstrate how to direct detection electron counting cameras enhance 3DED results (3 to better than 1 Angstrom) for protein and small molecule structure determination. We will also show how Latitude D software facilitates collecting such data in an integrated and fully automated user interface.

Keywords: continuous electron diffraction tomography, direct detection, diffraction, Latitude D, Digitalmicrograph, proteins, small molecules

Procedia PDF Downloads 112
31073 Behavior Consistency Analysis for Workflow Nets Based on Branching Processes

Authors: Wang Mimi, Jiang Changjun, Liu Guanjun, Fang Xianwen

Abstract:

Loop structure often appears in the business process modeling, analyzing the consistency of corresponding workflow net models containing loop structure is a problem, the existing behavior consistency methods cannot analyze effectively the process models with the loop structure. In the paper, by analyzing five kinds of behavior relations of transitions, a three-dimensional figure and two-dimensional behavior relation matrix are proposed. Based on this, analysis method of behavior consistency of business process based on Petri net branching processes is proposed. Finally, an example is given out, which shows the method is effective.

Keywords: workflow net, behavior consistency measures, loop, branching process

Procedia PDF Downloads 391
31072 Impact Position Method Based on Distributed Structure Multi-Agent Coordination with JADE

Authors: YU Kaijun, Liang Dong, Zhang Yarong, Jin Zhenzhou, Yang Zhaobao

Abstract:

For the impact monitoring of distributed structures, the traditional positioning methods are based on the time difference, which includes the four-point arc positioning method and the triangulation positioning method. But in the actual operation, these two methods have errors. In this paper, the Multi-Agent Blackboard Coordination Principle is used to combine the two methods. Fusion steps: (1) The four-point arc locating agent calculates the initial point and records it to the Blackboard Module.(2) The triangulation agent gets its initial parameters by accessing the initial point.(3) The triangulation agent constantly accesses the blackboard module to update its initial parameters, and it also logs its calculated point into the blackboard.(4) When the subsequent calculation point and the initial calculation point are within the allowable error, the whole coordination fusion process is finished. This paper presents a Multi-Agent collaboration method whose agent framework is JADE. The JADE platform consists of several agent containers, with the agent running in each container. Because of the perfect management and debugging tools of the JADE, it is very convenient to deal with complex data in a large structure. Finally, based on the data in Jade, the results show that the impact location method based on Multi-Agent coordination fusion can reduce the error of the two methods.

Keywords: impact monitoring, structural health monitoring(SHM), multi-agent system(MAS), black-board coordination, JADE

Procedia PDF Downloads 181
31071 Market Acceptance of a Murabaha-Based Finance Structure within a Social Network of Non-Islamic Small and Medium Enterprise Owners in African Procurement

Authors: Craig M. Allen

Abstract:

Twenty two African entrepreneurs with Small and Medium Enterprises (SMEs) in a single social network centered around a non-Muslim population in a smaller African country, selected an Islamic financing structure, a form of Murabaha, based solely on market rationale. These entrepreneurs had all won procurement contracts from major purchasers of goods within their country and faced difficulty arranging traditional bank financing to support their supply-chain needs. The Murabaha-based structure satisfied their market-driven demand and provided an attractive alternative to the traditional bank-offered lending products. The Murabaha-styled trade-financing structure was not promoted with any religious implications, but solely as a market solution to the existing problems associated with bank-related financing. This indicates the strong market forces that draw SMEs to financing structures that are traditionally considered within the framework of Islamic finance.

Keywords: Africa, entrepreneurs, Islamic finance, market acceptance, Murabaha, SMEs

Procedia PDF Downloads 186
31070 Analysis of the Behavior of the Structure Under Internal Anfo Explosion

Authors: Seung-Min Ko, Seung-Jai Choi, Gun Jung, Jang-Ho Jay Kim

Abstract:

Although extensive explosion-related research has been performed in the past several decades, almost no research has focused on internal blasts. However, internal blast research is needed to understand about the behavior of a containment structure or building under internal blast loading, as in the case of the Chornobyl and Fukushima nuclear accidents. Therefore, the internal blast study concentrated on RC and PSC structures is performed. The test data obtained from reinforced concrete (RC) and prestressed concrete (PSC) tubular structures applied with an internal explosion using ammonium nitrate/fuel oil (ANFO) charge are used to assess their deformation resistance and ultimate failure load based on the structural stiffness change under various charge weight. For the internal blast charge weight, ANFO explosive charge weights of 15.88, 20.41, 22.68 and 24.95 kg were selected for the RC tubular structures, and 22.68, 24.95, 27.22, 29.48, and 31.75 kg were selected for PSC tubular structures, which were detonated at the center of cross section at the mid-span with a standoff distance of 1,000mm to the inner wall surface. Then, the test data were used to predict the internal charge weight required to fail a real scale reinforced concrete containment vessels (RCCV) and prestressed concrete containment vessel (PCCV). Then, the analytical results based on the experimental data were derived using the simple assumptions of the models, and another approach using the stiffness, deformation and explosion weight relationship was used to formulate a general method for analyzing internal blasted tubular structures. A model of the internal explosion of a steel tube was used as an example for validation. The proposed method can be used generically, using factors according to the material characteristics of the target structures. The results of the study are discussed in detail in the paper.

Keywords: internal blast, reinforced concrete, RCCV, PCCV, stiffness, blast safety

Procedia PDF Downloads 83
31069 An Active Subsurface Geological Structure Pattern of Mud Volcano Phenomenon as an Environmental Impact of Petroleum Withdrawal in Sidoarjo, East Java, Indonesia

Authors: M. M. S. Prahastomi, M. Muhajir Saputra, Axel Derian

Abstract:

Lapindo mud (LUSI ) phenomenon which occurred in Sidoarjo 2006 is a national scale of the geological phenomenon. This mudflow forms a mud volcano that spreads by time is in the need of serious treatment. Some further research has been conducted either by the application method of geodesy, geophysics, and subsurface geology, but still remains a mystery to this phenomenon. Sidoarjo Physiographic regions are included in the Kendeng zone flanked by Rembang zones in northern and Solo zones in southern. In this region revealed Kabuh formation, Jombang formation, and alluvium. In general, in the northern part of the area is composed of sedimentary rocks Sidoarjo klastika, epiklastic, pyroclastics, and older alluvium of the Early Pleistocene to Resen. The study was conducted with the literature study of the stratigraphy and regional geology as well as secondary data from observations coupled gravity method (Anomaly Bouger). The aim of the study is to reveal the subsurface geology structure pattern and the changes in mass flow. Gravity anomaly data were obtained from the calculation of the value of gravity and altitude, then processed into gravity anomaly contours which reflect changes in density of each group observed gravity. The gravity data could indicate a bottom surface which deformation occur the stronger or more intense to the south. Deformation in the form of gravity impairment was associated with a decrease in future density which is indicated by the presence of gas, water and gas bursts. Sectional analysis of changes in the measured value of gravity at different times indicates a change in the value of gravity caused by the presence of subsurface subsidence. While the gravity anomaly section describes the fault zone causes the zone to be unstable.

Keywords: mud volcano, Lumpur Sidoarjo, Bouger anomaly, Indonesia

Procedia PDF Downloads 476
31068 Experimental and Computational Investigations of Baffle Position Effects on ‎the Performance of Oil and Water Separator Tanks

Authors: Haitham A. Hussein, Rozi Abdullah‏‎, Md Azlin Md Said ‎

Abstract:

Gravity separator tanks are used to separate oil from water in treatment units. Achieving the best flow ‎uniformity in a separator tank will improve the maximum removal efficiency of oil globules from water. ‎In this study, the effect on hydraulic performance of different baffle structure positions inside a tank ‎was investigated. Experimental data and 2D computation fluid dynamics were used for analysis. In the ‎numerical model, two-phase flow (drift flux model) was used to validate one-phase flow. For ‎laboratory measurements, the velocity fields were measured using an acoustic Doppler velocimeter. The ‎measurements were compared with the result of the computational model. The results of the ‎experimental and computational simulations indicate that the best location of a baffle structure is ‎achieved when the standard deviation of the velocity profile and the volume of the circulation zone ‎inside the tank are minimized.‎

Keywords: gravity separator tanks, CFD, baffle position, two phase flow, ADV, oil droplet

Procedia PDF Downloads 333
31067 A Novel Gateway Location Algorithm for Wireless Mesh Networks

Authors: G. M. Komba

Abstract:

The Internet Gateway (IGW) has extra ability than a simple Mesh Router (MR) and the responsibility to route mostly the all traffic from Mesh Clients (MCs) to the Internet backbone however, IGWs are more expensive. Choosing strategic locations for the Internet Gateways (IGWs) best location in Backbone Wireless Mesh (BWM) precarious to the Wireless Mesh Network (WMN) and the location of IGW can improve a quantity of performance related problem. In this paper, we propose a novel algorithm, namely New Gateway Location Algorithm (NGLA), which aims to achieve four objectives, decreasing the network cost effective, minimizing delay, optimizing the throughput capacity, Different from existing algorithms, the NGLA increasingly recognizes IGWs, allocates mesh routers (MRs) to identify IGWs and promises to find a feasible IGW location and install minimum as possible number of IGWs while regularly conserving the all Quality of Service (QoS) requests. Simulation results showing that the NGLA outperforms other different algorithms by comparing the number of IGWs with a large margin and it placed 40% less IGWs and 80% gain of throughput. Furthermore the NGLA is easy to implement and could be employed for BWM.

Keywords: Wireless Mesh Network, Gateway Location Algorithm, Quality of Service, BWM

Procedia PDF Downloads 373
31066 How to Use Big Data in Logistics Issues

Authors: Mehmet Akif Aslan, Mehmet Simsek, Eyup Sensoy

Abstract:

Big Data stands for today’s cutting-edge technology. As the technology becomes widespread, so does Data. Utilizing massive data sets enable companies to get competitive advantages over their adversaries. Out of many area of Big Data usage, logistics has significance role in both commercial sector and military. This paper lays out what big data is and how it is used in both military and commercial logistics.

Keywords: big data, logistics, operational efficiency, risk management

Procedia PDF Downloads 644
31065 Optimization of Shear Frame Structures Applying Various Forms of Wavelet Transforms

Authors: Seyed Sadegh Naseralavi, Sohrab Nemati, Ehsan Khojastehfar, Sadegh Balaghi

Abstract:

In the present research, various formulations of wavelet transform are applied on acceleration time history of earthquake. The mentioned transforms decompose the strong ground motion into low and high frequency parts. Since the high frequency portion of strong ground motion has a minor effect on dynamic response of structures, the structure is excited by low frequency part. Consequently, the seismic response of structure is predicted consuming one half of computational time, comparing with conventional time history analysis. Towards reducing the computational effort needed in seismic optimization of structure, seismic optimization of a shear frame structure is conducted by applying various forms of mentioned transformation through genetic algorithm.

Keywords: time history analysis, wavelet transform, optimization, earthquake

Procedia PDF Downloads 237
31064 Basic Modal Displacements (BMD) for Optimizing the Buildings Subjected to Earthquakes

Authors: Seyed Sadegh Naseralavi, Mohsen Khatibinia

Abstract:

In structural optimizations through meta-heuristic algorithms, analyses of structures are performed for many times. For this reason, performing the analyses in a time saving way is precious. The importance of the point is more accentuated in time-history analyses which take much time. To this aim, peak picking methods also known as spectrum analyses are generally utilized. However, such methods do not have the required accuracy either done by square root of sum of squares (SRSS) or complete quadratic combination (CQC) rules. The paper presents an efficient technique for evaluating the dynamic responses during the optimization process with high speed and accuracy. In the method, first by using a static equivalent of the earthquake, an initial design is obtained. Then, the displacements in the modal coordinates are achieved. The displacements are herein called basic modal displacements (MBD). For each new design of the structure, the responses can be derived by well scaling each of the MBD along the time and amplitude and superposing them together using the corresponding modal matrices. To illustrate the efficiency of the method, an optimization problems is studied. The results show that the proposed approach is a suitable replacement for the conventional time history and spectrum analyses in such problems.

Keywords: basic modal displacements, earthquake, optimization, spectrum

Procedia PDF Downloads 364
31063 An Examination of the Role of Perceived Leadership Styles on Job Satisfaction among Selected Bank Employees

Authors: Solomon Ojo

Abstract:

The study set out to investigate the role of perceived leadership style on achievement motivation of selected bank employees. The study was a cross-sectional survey. A total of 585 bank workers took part in the study; 283 (48.4%) were males while 302% (51.6%) were females. Mean age of 31.8 yrs (SD = 7.8 yrs) was reported for the participants for the study. Questionnaires were used for data collection. Data was analyzed using both descriptive and inferential statistic. The t- test for independent measures was used to test all the hypotheses, using the statistical package for social sciences version 21.0. The results in the study revealed that bank employees who perceived their leaders as high on consideration style of leadership reported more job satisfaction than bank employees who perceived their leaders as low on consideration style of leadership [t(583) = 16.43, p<.001]; bank employees who perceived their leaders as high in initiating structure style reported more job satisfaction than bank employees who perceived their leaders as low in initiating structure style [t(583)=12.06, p<.01]. The results showed further the influence of perceived leadership styles on all measures of job satisfaction. First, the result showed that bank employees who perceived their leaders as high on consideration style reported more satisfaction with hours worked each day than bank employees who perceived their leaders as low on consideration style [t(583) = 9.23, p<.01]. Second, the results revealed that bank employees who perceived their leaders as high on consideration style reported more satisfaction with flexibility in scheduling than bank employees who perceived their leaders as low on consideration style [t(583) = 8.80, p<.01]. Third, it was shown that bank employees who perceived their leaders as high on consideration style reported more satisfaction with location of work than bank employees who perceived their leaders as low on consideration style [t(583) = 14.17, p<.01] e.t.c. The results were extensively discussed in relation to relevant body of literature.

Keywords: leadership styles, job satisfaction, bank employees, perceived

Procedia PDF Downloads 225