Search results for: computational analysis
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 29143

Search results for: computational analysis

27793 A CDA-Driven Study of World English Series Published by Cengage Heinle

Authors: Mohammad Amin Mozaheb, Jalal Farzaneh Dehkordi, Khojasteh Hosseinzadehpilehvar

Abstract:

English Language Teaching (ELT) is widely promoted across the world. ELT textbooks play pivotal roles in the mentioned process. Since biases of authors have been an issue of continuing interest to analysts over the past few years, the present study seeks to analyze an ELT textbook using Critical Discourse Analysis (CDA). To obtain the goal of the study, the listening section of a book called World English 3 (new edition) has been analyzed in terms of the cultures and countries mentioned in the listening section of the book using content-based analysis. The analysis indicates biases towards certain cultures. Moreover, some countries are shown as rich and powerful countries, while some others have been shown as poor ones without considering the history behind them.

Keywords: ELT, textbooks, critical discourse analysis, World English

Procedia PDF Downloads 231
27792 Analysis of Lead Time Delays in Supply Chain: A Case Study

Authors: Abdel-Aziz M. Mohamed, Nermeen Coutry

Abstract:

Lead time is an important measure of supply chain performance. It impacts both customer satisfactions as well as the total cost of inventory. This paper presents the result of a study on the analysis of the customer order lead-time for a multinational company. In the study, the lead time was divided into three stages: order entry, order fulfillment, and order delivery. A sample of size 2,425 order lines from the company records were considered for this study. The sample data includes information regarding customer orders from the time of order entry until order delivery. Data regarding the lead time of each sage for different orders were also provided. Summary statistics on lead time data reveals that about 30% of the orders were delivered after the scheduled due date. The result of the multiple linear regression analysis technique revealed that component type, logistics parameter, order size and the customer type have significant impact on lead time. Data analysis on the stages of lead time indicates that stage 2 consumes over 50% of the lead time. Pareto analysis was made to study the reasons for the customer order delay in each of the 3 stages. Recommendation was given to resolve the problem.

Keywords: lead time reduction, customer satisfaction, service quality, statistical analysis

Procedia PDF Downloads 731
27791 Real Time Acquisition and Psychoacoustic Analysis of Brain Wave

Authors: Shweta Singh, Dipali Bansal, Rashima Mahajan

Abstract:

Psychoacoustics has become a potential area of research due to the growing interest of both laypersons and medical and mental health professionals. Non-invasive brain computer interface like Electroencephalography (EEG) is widely being used in this field. An attempt has been made in this paper to examine the response of EEG signals to acoustic stimuli further analysing the brain electrical activity. The real time EEG is acquired for 6 participants using a cost effective and portable EMOTIV EEG neuron headset. EEG data analysis is further done using EMOTIV test bench, EDF browser and EEGLAB (MATLAB Tool) application software platforms. Spectral analysis of acquired neural signals (AF3 channel) using these software platforms are clearly indicative of increased brain activity in various bands. The inferences drawn from such an analysis have significant correlation with subject’s subjective reporting of the experiences. The results suggest that the methodology adopted can further be used to assist patients with sleeping and depressive disorders.

Keywords: OM chant, spectral analysis, EDF browser, EEGLAB, EMOTIV, real time acquisition

Procedia PDF Downloads 281
27790 Features of Rail Strength Analysis in Conditions of Increased Force Loading

Authors: G. Guramishvili, M. Moistsrapishvili, L. Andghuladze

Abstract:

In the article are considered the problems arising at increasing of transferring from rolling stock axles on rail loading from 210 KN up to 270 KN and is offered for rail strength analysis definition of rail force loading complex integral characteristic with taking into account all affecting force factors that is characterizing specific operation condition of rail structure and defines the working capability of structure. As result of analysis due mentioned method is obtained that in the conditions of 270 KN loading the rail meets the working assessment criteria of rail and rail structures: Strength, rail track stability, rail links stability and its transverse stability, traffic safety condition that is rather important for post-Soviet countries railways.

Keywords: axial loading, rail force loading, rail structure, rail strength analysis, rail track stability

Procedia PDF Downloads 426
27789 Text Mining of Twitter Data Using a Latent Dirichlet Allocation Topic Model and Sentiment Analysis

Authors: Sidi Yang, Haiyi Zhang

Abstract:

Twitter is a microblogging platform, where millions of users daily share their attitudes, views, and opinions. Using a probabilistic Latent Dirichlet Allocation (LDA) topic model to discern the most popular topics in the Twitter data is an effective way to analyze a large set of tweets to find a set of topics in a computationally efficient manner. Sentiment analysis provides an effective method to show the emotions and sentiments found in each tweet and an efficient way to summarize the results in a manner that is clearly understood. The primary goal of this paper is to explore text mining, extract and analyze useful information from unstructured text using two approaches: LDA topic modelling and sentiment analysis by examining Twitter plain text data in English. These two methods allow people to dig data more effectively and efficiently. LDA topic model and sentiment analysis can also be applied to provide insight views in business and scientific fields.

Keywords: text mining, Twitter, topic model, sentiment analysis

Procedia PDF Downloads 179
27788 Scientific Theoretical Fundamentals of Comparative Analysis

Authors: Khalliyeva Gulnoz Iskandarovna, Mannonova Feruzabonu Sherali Qizi

Abstract:

A scientific field called comparative literature or literary comparative studies compares two or more literary phenomena. One of the most important scientific fields nowadays, when global social, cultural, and literary relations are growing daily, is comparative literature. Any comparative investigation reveals shared and unique characteristics of literary phenomena, which provide the cornerstone for the creation of overarching theoretical principles that apply to all literature. Comparative analysis consists of objects, and they are their constituents. For researchers, it is enough to know this. Comparative analysis, in addition to the above-mentioned actions, also focuses on comparing the components of the objects of analysis with each other. The purpose of this article is to investigate comparative analysis in literature and to identify similarities and differences between comparable objects. Students, teachers, and researchers should be able to describe comparative research techniques and their fundamental ideas when studying this topic. They should also have a basic understanding of comparative literature and their summary.

Keywords: object, natural, social, spiritual, epistemological, logical, methodological, methodological, axiological tasks, stages of comparison, environment, internal features, and typical situations

Procedia PDF Downloads 60
27787 A Study of Cloud Computing Solution for Transportation Big Data Processing

Authors: Ilgin Gökaşar, Saman Ghaffarian

Abstract:

The need for fast processed big data of transportation ridership (eg., smartcard data) and traffic operation (e.g., traffic detectors data) which requires a lot of computational power is incontrovertible in Intelligent Transportation Systems. Nowadays cloud computing is one of the important subjects and popular information technology solution for data processing. It enables users to process enormous measure of data without having their own particular computing power. Thus, it can also be a good selection for transportation big data processing as well. This paper intends to examine how the cloud computing can enhance transportation big data process with contrasting its advantages and disadvantages, and discussing cloud computing features.

Keywords: big data, cloud computing, Intelligent Transportation Systems, ITS, traffic data processing

Procedia PDF Downloads 469
27786 A Syntactic Errors Analysis in the Malaysian ESL Learners' Written Composition

Authors: Annie Gedion, Johan Severinus Tati, Jacinta Caroline Peter

Abstract:

Syntax error analysis studies have a significant role in English language teaching especially in the second language. This study investigates the syntax errors in written composition by 50 multilingual ESL learners in Politeknik Kota Kinabalu Sabah, Malaysia. The subjects speak their own dialect, Malay as their second language and English as their third or foreign language. Data were collected from the written discourse in the form of descriptive essays. The subjects were asked to write in the classroom within 45 minutes. 15 categories of errors were classified into a set of syntactic categories and were analysed based on the five steps of the syntactic analysis procedure. The findings of the study showed that the mother tongue interference, as well as lack of vocabulary and grammar knowledge, were the major sources of syntax errors in the learners’ written composition. Learners should be exposed to the differentiation of Malay and English grammar to avoid interference and effective learning of second language writing.

Keywords: errors analysis, syntactic analysis, English as a second language, ESL writing

Procedia PDF Downloads 283
27785 Element-Independent Implementation for Method of Lagrange Multipliers

Authors: Gil-Eon Jeong, Sung-Kie Youn, K. C. Park

Abstract:

Treatment for the non-matching interface is an important computational issue. To handle this problem, the method of Lagrange multipliers including classical and localized versions are the most popular technique. It essentially imposes the interface compatibility conditions by introducing Lagrange multipliers. However, the numerical system becomes unstable and inefficient due to the Lagrange multipliers. The interface element-independent formulation that does not include the Lagrange multipliers can be obtained by modifying the independent variables mathematically. Through this modification, more efficient and stable system can be achieved while involving equivalent accuracy comparing with the conventional method. A numerical example is conducted to verify the validity of the presented method.

Keywords: element-independent formulation, interface coupling, methods of Lagrange multipliers, non-matching interface

Procedia PDF Downloads 403
27784 Value Chain Based New Business Opportunity

Authors: Seonjae Lee, Sungjoo Lee

Abstract:

Excavation is necessary to remain competitive in the current business environment. The company survived the rapidly changing industry conditions by adapting new business strategy and reducing technology challenges. Traditionally, the two methods are conducted excavations for new businesses. The first method is, qualitative analysis of expert opinion, which is gathered through opportunities and secondly, new technologies are discovered through quantitative data analysis of method patents. The second method increases time and cost. Patent data is restricted for use and the purpose of discovering business opportunities. This study presents the company's characteristics (sector, size, etc.), of new business opportunities in customized form by reviewing the value chain perspective and to contributing to creating new business opportunities in the proposed model. It utilizes the trademark database of the Korean Intellectual Property Office (KIPO) and proprietary company information database of the Korea Enterprise Data (KED). This data is key to discovering new business opportunities with analysis of competitors and advanced business trademarks (Module 1) and trading analysis of competitors found in the KED (Module 2).

Keywords: value chain, trademark, trading analysis, new business opportunity

Procedia PDF Downloads 373
27783 The Application of Extend Spectrum-Based Pushover Analysis for Seismic Evaluation of Reinforced Concrete Wall Structures

Authors: Yang Liu

Abstract:

Reinforced concrete (RC) shear wall structures are one of the most popular and efficient structural forms for medium- and high-rise buildings to resist the action of earthquake loading. Thus, it is of great significance to evaluate the seismic demands of the RC shear walls. In this paper, the application of the extend spectrum-based pushover analysis (ESPA) method on the seismic evaluation of the shear wall structure is presented. The ESPA method includes a nonlinear consecutive pushover analysis procedure and a linear elastic modal response analysis procedure to consider the combination of modes in both elastic and inelastic cases. It is found from the results of case study that the ESPA method can predict the seismic performance of shear wall structures, including internal forces and deformations very well.

Keywords: reinforced concrete shear wall, seismic performance, high mode effect, nonlinear analysis

Procedia PDF Downloads 157
27782 The Visualizer for Real-Time Analysis of Internet Trends

Authors: Radek Malinský, Ivan Jelínek

Abstract:

The current web has become a modern encyclopedia, where people share their thoughts and ideas on various topics around them. Such kind of encyclopedia is very useful for other people who are looking for answers to their questions. However, with the growing popularity of social networking and blogging and ever expanding network services, there has also been a growing diversity of technologies along with different structure of individual websites. It is, therefore, difficult to directly find a relevant answer for a common Internet user. This paper presents a web application for the real-time end-to-end analysis of selected Internet trends; where the trend can be whatever the people post online. The application integrates fully configurable tools for data collection and analysis using selected webometric algorithms, and for its chronological visualization to user. It can be assumed that the application facilitates the users to evaluate the quality of various products that are mentioned online.

Keywords: Trend, visualizer, web analysis, web 2.0.

Procedia PDF Downloads 264
27781 A Parallel Cellular Automaton Model of Tumor Growth for Multicore and GPU Programming

Authors: Manuel I. Capel, Antonio Tomeu, Alberto Salguero

Abstract:

Tumor growth from a transformed cancer-cell up to a clinically apparent mass spans through a range of spatial and temporal magnitudes. Through computer simulations, Cellular Automata (CA) can accurately describe the complexity of the development of tumors. Tumor development prognosis can now be made -without making patients undergo through annoying medical examinations or painful invasive procedures- if we develop appropriate CA-based software tools. In silico testing mainly refers to Computational Biology research studies of application to clinical actions in Medicine. To establish sound computer-based models of cellular behavior, certainly reduces costs and saves precious time with respect to carrying out experiments in vitro at labs or in vivo with living cells and organisms. These aim to produce scientifically relevant results compared to traditional in vitro testing, which is slow, expensive, and does not generally have acceptable reproducibility under the same conditions. For speeding up computer simulations of cellular models, specific literature shows recent proposals based on the CA approach that include advanced techniques, such the clever use of supporting efficient data structures when modeling with deterministic stochastic cellular automata. Multiparadigm and multiscale simulation of tumor dynamics is just beginning to be developed by the concerned research community. The use of stochastic cellular automata (SCA), whose parallel programming implementations are open to yield a high computational performance, are of much interest to be explored up to their computational limits. There have been some approaches based on optimizations to advance in multiparadigm models of tumor growth, which mainly pursuit to improve performance of these models through efficient memory accesses guarantee, or considering the dynamic evolution of the memory space (grids, trees,…) that holds crucial data in simulations. In our opinion, the different optimizations mentioned above are not decisive enough to achieve the high performance computing power that cell-behavior simulation programs actually need. The possibility of using multicore and GPU parallelism as a promising multiplatform and framework to develop new programming techniques to speed-up the computation time of simulations is just starting to be explored in the few last years. This paper presents a model that incorporates parallel processing, identifying the synchronization necessary for speeding up tumor growth simulations implemented in Java and C++ programming environments. The speed up improvement that specific parallel syntactic constructs, such as executors (thread pools) in Java, are studied. The new tumor growth parallel model is proved using implementations with Java and C++ languages on two different platforms: chipset Intel core i-X and a HPC cluster of processors at our university. The parallelization of Polesczuk and Enderling model (normally used by researchers in mathematical oncology) proposed here is analyzed with respect to performance gain. We intend to apply the model and overall parallelization technique presented here to solid tumors of specific affiliation such as prostate, breast, or colon. Our final objective is to set up a multiparadigm model capable of modelling angiogenesis, or the growth inhibition induced by chemotaxis, as well as the effect of therapies based on the presence of cytotoxic/cytostatic drugs.

Keywords: cellular automaton, tumor growth model, simulation, multicore and manycore programming, parallel programming, high performance computing, speed up

Procedia PDF Downloads 244
27780 Chest Trauma and Early Pulmonary Embolism: The Risks

Authors: Vignesh Ratnaraj, Daniel Marascia, Kelly Ruecker

Abstract:

Purpose: Pulmonary embolism (PE) is a major cause of morbidity and mortality in trauma patients. Data suggests PE is occurring earlier in trauma patients, with attention being turned to possible de novo events. Here, we examine the incidence of early PE at a level 1 trauma center and examine the relationship with a chest injury. Method: A retrospective analysis was performed from a prospective trauma registry at a level 1 trauma center. All patients admitted from 1 January 2010 to 30 June 2019 diagnosed with PE following trauma were included. Early PE was considered a diagnosis within 72 hours of admission. The severity of the chest injury was determined by the Abbreviated Injury Score (AIS). Analysis of severe chest injury and incidence of early PE was performed using chi-square analysis. Sub-analysis on the timing of PE and PE location was also performed using chi-square analysis. Results: Chest injury was present in 125 of 184 patients diagnosed with PE. Early PE occurred in 28% (n=35) of patients with a chest injury, including 24.39% (n=10) with a severe chest injury. Neither chest injury nor severe chest injury determined the presence of early PE (p= > 0.05). Sub-analysis showed a trend toward central clots in early PE (37.14%, n=13) compared to late (27.78%, n=25); however, this was not found to be significant (p= > 0.05). Conclusion: PE occurs early in trauma patients, with almost one-third being diagnosed before 72 hours. This analysis does not support the paradigm that chest injury, nor severe chest injury, results in statistically significant higher rates of early PE. Interestingly, a trend toward early central PE was noted in those suffering chest trauma.

Keywords: trauma, PE, chest injury, anticoagulation

Procedia PDF Downloads 102
27779 Numerical Analysis of Laminar Reflux Condensation from Gas-Vapour Mixtures in Vertical Parallel Plate Channels

Authors: Foad Hassaninejadafarahani, Scott Ormiston

Abstract:

Reflux condensation occurs in a vertical channels and tubes when there is an upward core flow of vapor (or gas-vapor mixture) and a downward flow of the liquid film. The understanding of this condensation configuration is crucial in the design of reflux condensers, distillation columns, and in loss-of-coolant safety analyses in nuclear power plant steam generators. The unique feature of this flow is the upward flow of the vapor-gas mixture (or pure vapor) that retards the liquid flow via shear at the liquid-mixture interface. The present model solves the full, elliptic governing equations in both the film and the gas-vapor core flow. The computational mesh is non-orthogonal and adapts dynamically the phase interface, thus produces sharp and accurate interface. Shear forces and heat and mass transfer at the interface are accounted for fundamentally. This modeling is a big step ahead of current capabilities by removing the limitations of previous reflux condensation models which inherently cannot account for the detailed local balances of shear, mass, and heat transfer at the interface. Discretisation has been done based on a finite volume method and a co-located variable storage scheme. An in-house computer code was developed to implement the numerical solution scheme. Detailed results are presented for laminar reflux condensation from steam-air mixtures flowing in vertical parallel plate channels. The results include velocity and pressure profiles, as well as axial variations of film thickness, Nusselt number and interface gas mass fraction.

Keywords: Reflux, Condensation, CFD-Two Phase, Nusselt number

Procedia PDF Downloads 364
27778 Opposed Piston Engine Crankshaft Strength Calculation Using Finite Element Method

Authors: Konrad Pietrykowski, Michał Gęca, Michał Bialy

Abstract:

The paper presents the results of the crankshaft strength simulation. The crankshaft was taken from the opposed piston engine. Calculations were made using finite element method (FEM) in Abaqus software. This program allows to perform strength tests of individual machine parts as well as their assemblies. The crankshaft that was used in the calculations will be used in the two-stroke aviation research aircraft engine. The assumptions for the calculations were obtained from the AVL Boost software, from one-dimensional engine cycle model and from the multibody model using the method developed in the MSC Adams software. The research engine will be equipped with 3 combustion chambers and two crankshafts. In order to shorten the calculation time, only one crankcase analysis was performed. The cut of the shaft has been selected with the greatest forces resulting from the engine operation. Calculations were made for two cases. For maximum piston force when maximum bending load occurs and for the maximum torque. Cast iron material was adopted. For this material, Poisson's number, density, and Young's modulus were determined. The computational grid contained of 1,977,473 Tet elements. This type of elements was chosen because of the complex design of the crankshaft. Results are presented in the form of stress distributions maps and displacements on the surface and inside the geometry of the shaft. The results show the places of tension stresses, however, no stresses are exceeded at any place. The shaft can thus be applied to the engine in its present form. Acknowledgement: This work has been realized in the cooperation with The Construction Office of WSK 'PZL-KALISZ’ S.A. and is part of Grant Agreement No. POIR.01.02.00-00-0002/15 financed by the Polish National Centre for Research and Development.

Keywords: aircraft diesel engine, crankshaft, finite element method, two-stroke engine

Procedia PDF Downloads 181
27777 Flow Characterization in Complex Terrain for Aviation Safety

Authors: Adil Rasheed, Mandar Tabib

Abstract:

The paper describes the ability of a high-resolution Computational Fluid Dynamics model to predict terrain-induced turbulence and wind shear close to the ground. Various sensitivity studies to choose the optimal simulation setup for modeling the flow characteristics in a complex terrain are presented. The capabilities of the model are demonstrated by applying it to the Sandnessjøen Airport, Stokka in Norway, an airport that is located in a mountainous area. The model is able to forecast turbulence in real time and trigger an alert when atmospheric conditions might result in high wind shear and turbulence.

Keywords: aviation safety, terrain-induced turbulence, atmospheric flow, alert system

Procedia PDF Downloads 416
27776 Brand Placement Strategies in Turkey: The Case of “Yalan Dünya”

Authors: Burçe Boyraz

Abstract:

This study examines appearances of brand placement as an alternative communication strategy in television series by focusing on Yalan Dünya which is one of the most popular television series in Turkey. Consequently, this study has a descriptive research design and quantitative content analysis method is used in order to analyze frequency and time data of brand placement appearances in first 3 seasons of Yalan Dünya with 16 episodes. Analysis of brand placement practices in Yalan Dünya is dealt in three categories: episode-based analysis, season-based analysis and comparative analysis. At the end, brand placement practices in Yalan Dünya are evaluated in terms of type, form, duration and legal arrangements. As a result of this study, it is seen that brand placement plays a determinant role in Yalan Dünya content. Also, current legal arrangements make brand placement closer to other traditional communication strategies instead of differing brand placement from them distinctly.

Keywords: advertising, alternative communication strategy, brand placement, Yalan Dünya

Procedia PDF Downloads 248
27775 An Unbiased Profiling of Immune Repertoire via Sequencing and Analyzing T-Cell Receptor Genes

Authors: Yi-Lin Chen, Sheng-Jou Hung, Tsunglin Liu

Abstract:

Adaptive immune system recognizes a wide range of antigens via expressing a large number of structurally distinct T cell and B cell receptor genes. The distinct receptor genes arise from complex rearrangements called V(D)J recombination, and constitute the immune repertoire. A common method of profiling immune repertoire is via amplifying recombined receptor genes using multiple primers and high-throughput sequencing. This multiplex-PCR approach is efficient; however, the resulting repertoire can be distorted because of primer bias. To eliminate primer bias, 5’ RACE is an alternative amplification approach. However, the application of RACE approach is limited by its low efficiency (i.e., the majority of data are non-regular receptor sequences, e.g., containing intronic segments) and lack of the convenient tool for analysis. We propose a computational tool that can correctly identify non-regular receptor sequences in RACE data via aligning receptor sequences against the whole gene instead of only the exon regions as done in all other tools. Using our tool, the remaining regular data allow for an accurate profiling of immune repertoire. In addition, a RACE approach is improved to yield a higher fraction of regular T-cell receptor sequences. Finally, we quantify the degree of primer bias of a multiplex-PCR approach via comparing it to the RACE approach. The results reveal significant differences in frequency of VJ combination by the two approaches. Together, we provide a new experimental and computation pipeline for an unbiased profiling of immune repertoire. As immune repertoire profiling has many applications, e.g., tracing bacterial and viral infection, detection of T cell lymphoma and minimal residual disease, monitoring cancer immunotherapy, etc., our work should benefit scientists who are interested in the applications.

Keywords: immune repertoire, T-cell receptor, 5' RACE, high-throughput sequencing, sequence alignment

Procedia PDF Downloads 194
27774 Effectiveness of Column Geometry in High-Rise Buildings

Authors: Man Singh Meena

Abstract:

Structural engineers are facing different kind of challenges due to innovative & bold ideas of architects who are trying to design every structure with uniqueness. In RCC frame structures different geometry of columns can be used in design and rectangular columns can be placed with different type orientation. The analysis is design of structures can also be carried out by different type of software available i.e., STAAD Pro, ETABS and TEKLA. In recent times high-rise building modeling & analysis is done by ETABS due to its certain features which are superior to other software. The case study in this paper mainly emphasizes on structural behavior of high rise building for different column shape configurations like Circular, Square, Rectangular and Rectangular with 90-degree Rotation and rectangular shape plan. In all these column shapes the areas of columns are kept same to study the effect on design of concrete area is same. Modelling of 20-storeys R.C.C. framed building is done on the ETABS software for analysis. Post analysis of the structure, maximum bending moments, shear forces and maximum longitudinal reinforcement are computed and compared for three different story structures to identify the effectiveness of geometry of column.

Keywords: high-rise building, column geometry, building modelling, ETABS analysis, building design, structural analysis, structural optimization

Procedia PDF Downloads 81
27773 Hierarchical Checkpoint Protocol in Data Grids

Authors: Rahma Souli-Jbali, Minyar Sassi Hidri, Rahma Ben Ayed

Abstract:

Grid of computing nodes has emerged as a representative means of connecting distributed computers or resources scattered all over the world for the purpose of computing and distributed storage. Since fault tolerance becomes complex due to the availability of resources in decentralized grid environment, it can be used in connection with replication in data grids. The objective of our work is to present fault tolerance in data grids with data replication-driven model based on clustering. The performance of the protocol is evaluated with Omnet++ simulator. The computational results show the efficiency of our protocol in terms of recovery time and the number of process in rollbacks.

Keywords: data grids, fault tolerance, clustering, chandy-lamport

Procedia PDF Downloads 341
27772 Dynamics Analyses of Swing Structure Subject to Rotational Forces

Authors: Buntheng Chhorn, WooYoung Jung

Abstract:

Large-scale swing has been used in entertainment and performance, especially in circus, for a very long time. To increase the safety of this type of structure, a thorough analysis for displacement and bearing stress was performed for an extreme condition where a full cycle swing occurs. Different masses, ranging from 40 kg to 220 kg, and velocities were applied on the swing. Then, based on the solution of differential dynamics equation, swing velocity response to harmonic force was obtained. Moreover, the resistance capacity was estimated based on ACI steel structure design guide. Subsequently, numerical analysis was performed in ABAQUS to obtain the stress on each frame of the swing. Finally, the analysis shows that the expansion of swing structure frame section was required for mass bigger than 150kg.

Keywords: swing structure, displacement, bearing stress, dynamic loads response, finite element analysis

Procedia PDF Downloads 378
27771 Solving the Pseudo-Geometric Traveling Salesman Problem with the “Union Husk” Algorithm

Authors: Boris Melnikov, Ye Zhang, Dmitrii Chaikovskii

Abstract:

This study explores the pseudo-geometric version of the extensively researched Traveling Salesman Problem (TSP), proposing a novel generalization of existing algorithms which are traditionally confined to the geometric version. By adapting the "onion husk" method and introducing auxiliary algorithms, this research fills a notable gap in the existing literature. Through computational experiments using randomly generated data, several metrics were analyzed to validate the proposed approach's efficacy. Preliminary results align with expected outcomes, indicating a promising advancement in TSP solutions.

Keywords: optimization problems, traveling salesman problem, heuristic algorithms, “onion husk” algorithm, pseudo-geometric version

Procedia PDF Downloads 207
27770 Concentration of Droplets in a Transient Gas Flow

Authors: Timur S. Zaripov, Artur K. Gilfanov, Sergei S. Sazhin, Steven M. Begg, Morgan R. Heikal

Abstract:

The calculation of the concentration of inertial droplets in complex flows is encountered in the modelling of numerous engineering and environmental phenomena; for example, fuel droplets in internal combustion engines and airborne pollutant particles. The results of recent research, focused on the development of methods for calculating concentration and their implementation in the commercial CFD code, ANSYS Fluent, is presented here. The study is motivated by the investigation of the mixture preparation processes in internal combustion engines with direct injection of fuel sprays. Two methods are used in our analysis; the Fully Lagrangian method (also known as the Osiptsov method) and the Eulerian approach. The Osiptsov method predicts droplet concentrations along path lines by solving the equations for the components of the Jacobian of the Eulerian-Lagrangian transformation. This method significantly decreases the computational requirements as it does not require counting of large numbers of tracked droplets as in the case of the conventional Lagrangian approach. In the Eulerian approach the average droplet velocity is expressed as a function of the carrier phase velocity as an expansion over the droplet response time and transport equation can be solved in the Eulerian form. The advantage of the method is that droplet velocity can be found without solving additional partial differential equations for the droplet velocity field. The predictions from the two approaches were compared in the analysis of the problem of a dilute gas-droplet flow around an infinitely long, circular cylinder. The concentrations of inertial droplets, with Stokes numbers of 0.05, 0.1, 0.2, in steady-state and transient laminar flow conditions, were determined at various Reynolds numbers. In the steady-state case, flows with Reynolds numbers of 1, 10, and 100 were investigated. It has been shown that the results predicted using both methods are almost identical at small Reynolds and Stokes numbers. For larger values of these numbers (Stokes — 0.1, 0.2; Reynolds — 10, 100) the Eulerian approach predicted a wider spread in concentration in the perturbations caused by the cylinder that can be attributed to the averaged droplet velocity field. The transient droplet flow case was investigated for a Reynolds number of 200. Both methods predicted a high droplet concentration in the zones of high strain rate and low concentrations in zones of high vorticity. The maxima of droplet concentration predicted by the Osiptsov method was up to two orders of magnitude greater than that predicted by the Eulerian method; a significant variation for an approach widely used in engineering applications. Based on the results of these comparisons, the Osiptsov method has resulted in a more precise description of the local properties of the inertial droplet flow. The method has been applied to the analysis of the results of experimental observations of a liquid gasoline spray at representative fuel injection pressure conditions. The preliminary results show good qualitative agreement between the predictions of the model and experimental data.

Keywords: internal combustion engines, Eulerian approach, fully Lagrangian approach, gasoline fuel sprays, droplets and particle concentrations

Procedia PDF Downloads 257
27769 Development and Characterization of Biodegradable Films Based on Biopolymer Extracted From Natural Sources

Authors: Dalila Hammiche, Lisa Klaai, Sonia Imzi, Amar Boukerrou

Abstract:

The fight against plastic pollution implies the development of polymers as alternatives to synthetic polymers. Starch is a natural polymer that can easily be plasticized by means of additives. The objective of this work is to develop and characterize biodegradable biofilms based on starch, plasticized by glycerol (20 and 30%). The elaboration of the biofilms was carried out by the casting method under simple conditions. The samples were characterized by infrared spectroscopy analysis with Fourier transform (FTIR), thermogravimetric analysis, and biodegradability test. Infrared spectral analysis showed that the 30% and 20% glycerol films have the same chemical structure and no functional group changes occurred. Thermogravimetric analysis showed that a 30% glycerol film has higher thermal stability than a 20% glycerol film. Biodegradability test showed that the lower the percentage of glycerol, the more easily the biofilm degrades.

Keywords: starch, natural sources, FTIR, thermogravimetric analysis, biodegradability test

Procedia PDF Downloads 102
27768 The Analysis on Leadership Skills in UK Automobile Manufacturing Enterprises

Authors: Yanting Cao

Abstract:

The UK has strong economic growth, which attracts other countries to invest there through globalization. This research process will be based on quantitative and qualitative descriptive analysis using interviews. The secondary analysis will involve a case study approach to understand the important aspects of leadership skills. The research outcomes will be identifying the strength and weaknesses of the leadership skills of UK automobile manufacturing enterprises and suggest the best practices adopted by the respective countries for better results.

Keywords: engineering management, leadership, Industrial project management, Project managers, automobile manufacturing

Procedia PDF Downloads 186
27767 Analysis of Thermal Comfort in Educational Buildings Using Computer Simulation: A Case Study in Federal University of Parana, Brazil

Authors: Ana Julia C. Kfouri

Abstract:

A prerequisite of any building design is to provide security to the users, taking the climate and its physical and physical-geometrical variables into account. It is also important to highlight the relevance of the right material elements, which arise between the person and the agent, and must provide improved thermal comfort conditions and low environmental impact. Furthermore, technology is constantly advancing, as well as computational simulations for projects, and they should be used to develop sustainable building and to provide higher quality of life for its users. In relation to comfort, the more satisfied the building users are, the better their intellectual performance will be. Based on that, the study of thermal comfort in educational buildings is of relative relevance, since the thermal characteristics in these environments are of vital importance to all users. Moreover, educational buildings are large constructions and when they are poorly planned and executed they have negative impacts to the surrounding environment, as well as to the user satisfaction, throughout its whole life cycle. In this line of thought, to evaluate university classroom conditions, it was accomplished a detailed case study on the thermal comfort situation at Federal University of Parana (UFPR). The main goal of the study is to perform a thermal analysis in three classrooms at UFPR, in order to address the subjective and physical variables that influence thermal comfort inside the classroom. For the assessment of the subjective components, a questionnaire was applied in order to evaluate the reference for the local thermal conditions. Regarding the physical variables, it was carried out on-site measurements, which consist of performing measurements of air temperature and air humidity, both inside and outside the building, as well as meteorological variables, such as wind speed and direction, solar radiation and rainfall, collected from a weather station. Then, a computer simulation based on results from the EnergyPlus software to reproduce air temperature and air humidity values of the three classrooms studied was conducted. The EnergyPlus outputs were analyzed and compared with the on-site measurement results to be possible to come out with a conclusion related to the local thermal conditions. The methodological approach included in the study allowed a distinct perspective in an educational building to better understand the classroom thermal performance, as well as the reason of such behavior. Finally, the study induces a reflection about the importance of thermal comfort for educational buildings and propose thermal alternatives for future projects, as well as a discussion about the significant impact of using computer simulation on engineering solutions, in order to improve the thermal performance of UFPR’s buildings.

Keywords: computer simulation, educational buildings, EnergyPlus, humidity, temperature, thermal comfort

Procedia PDF Downloads 388
27766 Effects of Video Games and Online Chat on Mathematics Performance in High School: An Approach of Multivariate Data Analysis

Authors: Lina Wu, Wenyi Lu, Ye Li

Abstract:

Regarding heavy video game players for boys and super online chat lovers for girls as a symbolic phrase in the current adolescent culture, this project of data analysis verifies the displacement effect on deteriorating mathematics performance. To evaluate correlation or regression coefficients between a factor of playing video games or chatting online and mathematics performance compared with other factors, we use multivariate analysis technique and take gender difference into account. We find the most important reason for the negative sign of the displacement effect on mathematics performance due to students’ poor academic background. Statistical analysis methods in this project could be applied to study internet users’ academic performance from the high school education to the college education.

Keywords: correlation coefficients, displacement effect, multivariate analysis technique, regression coefficients

Procedia PDF Downloads 365
27765 Application of Wavelet Based Approximation for the Solution of Partial Integro-Differential Equation Arising from Viscoelasticity

Authors: Somveer Singh, Vineet Kumar Singh

Abstract:

This work contributes a numerical method based on Legendre wavelet approximation for the treatment of partial integro-differential equation (PIDE). Operational matrices of Legendre wavelets reduce the solution of PIDE into the system of algebraic equations. Some useful results concerning the computational order of convergence and error estimates associated to the suggested scheme are presented. Illustrative examples are provided to show the effectiveness and accuracy of proposed numerical method.

Keywords: legendre wavelets, operational matrices, partial integro-differential equation, viscoelasticity

Procedia PDF Downloads 448
27764 A Hebbian Neural Network Model of the Stroop Effect

Authors: Vadim Kulikov

Abstract:

The classical Stroop effect is the phenomenon that it takes more time to name the ink color of a printed word if the word denotes a conflicting color than if it denotes the same color. Over the last 80 years, there have been many variations of the experiment revealing various mechanisms behind semantic, attentional, behavioral and perceptual processing. The Stroop task is known to exhibit asymmetry. Reading the words out loud is hardly dependent on the ink color, but naming the ink color is significantly influenced by the incongruent words. This asymmetry is reversed, if instead of naming the color, one has to point at a corresponding color patch. Another debated aspects are the notions of automaticity and how much of the effect is due to semantic and how much due to response stage interference. Is automaticity a continuous or an all-or-none phenomenon? There are many models and theories in the literature tackling these questions which will be discussed in the presentation. None of them, however, seems to capture all the findings at once. A computational model is proposed which is based on the philosophical idea developed by the author that the mind operates as a collection of different information processing modalities such as different sensory and descriptive modalities, which produce emergent phenomena through mutual interaction and coherence. This is the framework theory where ‘framework’ attempts to generalize the concepts of modality, perspective and ‘point of view’. The architecture of this computational model consists of blocks of neurons, each block corresponding to one framework. In the simplest case there are four: visual color processing, text reading, speech production and attention selection modalities. In experiments where button pressing or pointing is required, a corresponding block is added. In the beginning, the weights of the neural connections are mostly set to zero. The network is trained using Hebbian learning to establish connections (corresponding to ‘coherence’ in framework theory) between these different modalities. The amount of data fed into the network is supposed to mimic the amount of practice a human encounters, in particular it is assumed that converting written text into spoken words is a more practiced skill than converting visually perceived colors to spoken color-names. After the training, the network performs the Stroop task. The RT’s are measured in a canonical way, as these are continuous time recurrent neural networks (CTRNN). The above-described aspects of the Stroop phenomenon along with many others are replicated. The model is similar to some existing connectionist models but as will be discussed in the presentation, has many advantages: it predicts more data, the architecture is simpler and biologically more plausible.

Keywords: connectionism, Hebbian learning, artificial neural networks, philosophy of mind, Stroop

Procedia PDF Downloads 267