Search results for: tempo or speed
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 2922

Search results for: tempo or speed

792 Indian Business-Papers in Industrial Revolution 4.0: A Paradigm Shift

Authors: Disha Batra

Abstract:

The Industrial Revolution 4.0 is quite different, and a paradigm shift is underway in the media industry. With the advent of automated journalism and social media platforms, newspaper organizations have changed the way news was gathered and reported. The emergence of the fourth industrial revolution in the early 21st century has made the newspapers to adapt the changing technologies to remain relevant. This paper investigates the content of Indian business-papers in the era of the fourth industrial revolution and how these organizations have emerged in the time of convergence. The study is the content analyses of the top three Indian business dailies as per IRS (Indian Readership Survey) 2017 over a decade. The parametric analysis of the different parameters (source of information, use of illustrations, advertisements, layout, and framing, etc.) have been done in order to come across with the distinct adaptations and modifications by these dailies. The paper significantly dwells upon the thematic analysis of these newspapers in order to explore and find out the coverage given to various sub-themes of EBF (economic, business, and financial) journalism. Further, this study reveals the effect of high-speed algorithm-based trading, the aftermath of the fourth industrial revolution on the creative and investigative aspect of delivering financial stories by these respective newspapers. The study indicates a change heading towards an ongoing paradigm shift in the business newspaper industry with an adequate change in the source of information gathering along with the subtle increase in the coverage of financial news stories over the time.

Keywords: business-papers, business news, financial news, industrial revolution 4.0.

Procedia PDF Downloads 106
791 Vehicle Gearbox Fault Diagnosis Based on Cepstrum Analysis

Authors: Mohamed El Morsy, Gabriela Achtenová

Abstract:

Research on damage of gears and gear pairs using vibration signals remains very attractive, because vibration signals from a gear pair are complex in nature and not easy to interpret. Predicting gear pair defects by analyzing changes in vibration signal of gears pairs in operation is a very reliable method. Therefore, a suitable vibration signal processing technique is necessary to extract defect information generally obscured by the noise from dynamic factors of other gear pairs. This article presents the value of cepstrum analysis in vehicle gearbox fault diagnosis. Cepstrum represents the overall power content of a whole family of harmonics and sidebands when more than one family of sidebands is present at the same time. The concept for the measurement and analysis involved in using the technique are briefly outlined. Cepstrum analysis is used for detection of an artificial pitting defect in a vehicle gearbox loaded with different speeds and torques. The test stand is equipped with three dynamometers; the input dynamometer serves as the internal combustion engine, the output dynamometers introduce the load on the flanges of the output joint shafts. The pitting defect is manufactured on the tooth side of a gear of the fifth speed on the secondary shaft. Also, a method for fault diagnosis of gear faults is presented based on order cepstrum. The procedure is illustrated with the experimental vibration data of the vehicle gearbox. The results show the effectiveness of cepstrum analysis in detection and diagnosis of the gear condition.

Keywords: cepstrum analysis, fault diagnosis, gearbox, vibration signals

Procedia PDF Downloads 370
790 Isolated Iterating Fractal Independently Corresponds with Light and Foundational Quantum Problems

Authors: Blair D. Macdonald

Abstract:

After nearly one hundred years of its origin, foundational quantum mechanics remains one of the greatest unexplained mysteries in physicists today. Within this time, chaos theory and its geometry, the fractal, has developed. In this paper, the propagation behaviour with an iteration of a simple fractal, the Koch Snowflake, was described and analysed. From an arbitrary observation point within the fractal set, the fractal propagates forward by oscillation—the focus of this study and retrospectively behind by exponential growth from a point beginning. It propagates a potentially infinite exponential oscillating sinusoidal wave of discrete triangle bits sharing many characteristics of light and quantum entities. The model's wave speed is potentially constant, offering insights into the perception and a direction of time where, to an observer, when travelling at the frontier of propagation, time may slow to a stop. In isolation, the fractal is a superposition of component bits where position and scale present a problem of location. In reality, this problem is experienced within fractal landscapes or fields where 'position' is only 'known' by the addition of information or markers. The quantum' measurement problem', 'uncertainty principle,' 'entanglement,' and the classical-quantum interface are addressed; these are a problem of scale invariance associated with isolated fractality. Dual forward and retrospective perspectives of the fractal model offer the opportunity for unification between quantum mechanics and cosmological mathematics, observations, and conjectures. Quantum and cosmological problems may be different aspects of the one fractal geometry.

Keywords: measurement problem, observer, entanglement, unification

Procedia PDF Downloads 82
789 Arithmetic Operations Based on Double Base Number Systems

Authors: K. Sanjayani, C. Saraswathy, S. Sreenivasan, S. Sudhahar, D. Suganya, K. S. Neelukumari, N. Vijayarangan

Abstract:

Double Base Number System (DBNS) is an imminent system of representing a number using two bases namely 2 and 3, which has its application in Elliptic Curve Cryptography (ECC) and Digital Signature Algorithm (DSA).The previous binary method representation included only base 2. DBNS uses an approximation algorithm namely, Greedy Algorithm. By using this algorithm, the number of digits required to represent a larger number is less when compared to the standard binary method that uses base 2 algorithms. Hence, the computational speed is increased and time being reduced. The standard binary method uses binary digits 0 and 1 to represent a number whereas the DBNS method uses binary digit 1 alone to represent any number (canonical form). The greedy algorithm uses two ways to represent the number, one is by using only the positive summands and the other is by using both positive and negative summands. In this paper, arithmetic operations are used for elliptic curve cryptography. Elliptic curve discrete logarithm problem is the foundation for most of the day to day elliptic curve cryptography. This appears to be a momentous hard slog compared to digital logarithm problem. In elliptic curve digital signature algorithm, the key generation requires 160 bit of data by usage of standard binary representation. Whereas, the number of bits required generating the key can be reduced with the help of double base number representation. In this paper, a new technique is proposed to generate key during encryption and extraction of key in decryption.

Keywords: cryptography, double base number system, elliptic curve cryptography, elliptic curve digital signature algorithm

Procedia PDF Downloads 391
788 Development and Characterization of a Film Based on Hydroxypropyl Methyl Cellulose Incorporated by a Phenolic Extract of Fennel and Reinforced by Magnesium Oxide: In Vivo - in Vitro

Authors: Mazouzi Nourdjihane, K. Boutemak, A. Haddad, Y. Chegreouche

Abstract:

In the last decades, biodegradable polymers have been considered as one of the most popular options for the delivery of drugs and various conventional doses. The film forming system (FFS) can be used in topical, transdermal, ophthalmic, oral and gastric applications. Recently this system has focused on improving drug delivery, which can promote drug release. In this context, the aim of this study is to create polymeric film-forming systems for the stomach and to evaluate and test their gastroprotective effects, comparing the effects of changes in composition on film characteristics. It uses a plant-derived polyphenol extract extracted from fennel to demonstrate anti-inflammatory activity in the film. The films are made from hydroxypropyl methylcellulose polymer and different types of plastic, glycerol and polyethylene glycol. The ffs properties show that MgO-glycerol-reinforced hydroxypropylmethylcellulose (HPMC-MgO-Gly) is better than that based on MgO-PEG-reinforced hydroxypropylmethylcellulose (HPMC-MgO-PEG). It is durable, has a faster drying time and allows for maximum recovery. Water vapor strength and blowing speed and other additions show another advantage of HPMC-MgO-Gly compared to HPMC-MgO-PEG, indicating good adhesion between the support (top) and film production. In this study, the gastroprotective effect of fennel phenol extract was found, showing that this plant material has a gastroprotective effect on ulcers and that the film can absorb the active substance.

Keywords: film formin system, hydroxypropyl methylcellulose, magnesium oxide, in vivo

Procedia PDF Downloads 59
787 Atmospheric Dispersion Modeling for a Hypothetical Accidental Release from the 3 MW TRIGA Research Reactor of Bangladesh

Authors: G. R. Khan, Sadia Mahjabin, A. S. Mollah, M. R. Mawla

Abstract:

Atmospheric dispersion modeling is significant for any nuclear facilities in the country to predict the impact of radiological doses on environment as well as human health. That is why to ensure safety of workers and population at plant site; Atmospheric dispersion modeling and radiation dose calculations were carried out for a hypothetical accidental release of airborne radionuclide from the 3 MW TRIGA research reactor of Savar, Bangladesh. It is designed with reactor core which consists of 100 fuel elements(1.82245 cm in diameter and 38.1 cm in length), arranged in an annular corefor steady-state and square wave power level of 3 MW (thermal) and for pulsing with maximum power level of 860MWth.The fuel is in the form of a uniform mixture of 20% uranium and 80% zirconium hydride. Total effective doses (TEDs) to the public at various downwind distances were evaluated with a health physics computer code “HotSpot” developed by Lawrence Livermore National Laboratory, USA. The doses were estimated at different Pasquill stability classes (categories A-F) with site-specific averaged meteorological conditions. The meteorological data, such as, average wind speed, frequency distribution of wind direction, etc. have also been analyzed based on the data collected near the reactor site. The results of effective doses obtained remain within the recommended maximum effective dose.

Keywords: accidental release, dispersion modeling, total effective dose, TRIGA

Procedia PDF Downloads 127
786 Assessment of Personal Level Exposures to Particulate Matter among Children in Rural Preliminary Schools as an Indoor Air Pollution Monitoring

Authors: Seyedtaghi Mirmohammadi, J. Yazdani, S. M. Asadi, M. Rokni, A. Toosi

Abstract:

There are many indoor air quality studies with an emphasis on indoor particulate matters (PM2.5) monitoring. Whereas, there is a lake of data about indoor PM2.5 concentrations in rural area schools (especially in classrooms), since preliminary children are assumed to be more defenseless to health hazards and spend a large part of their time in classrooms. The objective of this study was indoor PM2.5 concentration quality assessment. Fifteen preliminary schools by time-series sampling were selected to evaluate the indoor air quality in the rural district of Sari city, Iran. Data on indoor air climate parameters (temperature, relative humidity and wind speed) were measured by a hygrometer and thermometer. Particulate matters (PM2.5) were collected and assessed by Real Time Dust Monitor, (MicroDust Pro, Casella, UK). The mean indoor PM2.5 concentration in the studied classrooms was 135µg/m3 in average. The multiple linear regression revealed that a correlation between PM2.5 concentration and relative humidity, distance from city center and classroom size. Classroom size yields reasonable negative relationship, the PM2.5 concentration was ranged from 65 to 540μg/m3 and statistically significant at 0.05 level and the relative humidity was ranged from 70 to 85% and dry bulb temperature ranged from 28 to 29°C were statistically significant at 0.035 and 0.05 level, respectively. A statistical predictive model was obtained from multiple regressions modeling for PM2.5 and indoor psychrometric parameters.

Keywords: particulate matters, classrooms, regression, concentration, humidity

Procedia PDF Downloads 303
785 Beyond the Jingoism of “Infodemic” in the Use of Language: Prospects for a Better Nigeria

Authors: Anacletus Ogbunkwu

Abstract:

It is very disheartening that fake news or inaccurate information spread like wide fire and even with greater speed than fact based news/information. The peak of this anomaly is manifest in information management on the Corona virus pandemic, political/leadership based information, ethnic bigotry, unwarranted panics, false alarms, religious fanaticism, and business moguls in their advertorials, comedies, etc. This ugly situation has left Nigeria and her citizens with emotional trauma, unguided agitations, incessant tribal wars, lost of life and property, widened disunity among Nigerian ethnic and religious groups, amplified insecurity, aided election violence, etc. Unfortunately, among the major driving factors to this misinformation and conspiracy are the official/government and private news agencies, gossip, comedians, and social media handles such as; facebook, twitter, whatsapp, instagram, and online news agencies, etc. Thus this paper examines the impact of misinformation here referred to as infodemic. Also, it studies the epistemic effect of misinformation on the citizens of Nigeria in order to find ways of abating this anomaly for a better society. The methods of exposition and hermeneutics will be used in order to gain in-depth study of the details of infodemic in Nigeria and to offer philosophical analysis/interpretation of data as gathered, respectively. This paper concludes that misinformation or fake news has a perilous effect of epistemic mistrust to Nigeria and her citizens; hence infodemic is a cog in the wheel of National progress.

Keywords: nigeria, infodemic, language, media, news, progress

Procedia PDF Downloads 109
784 An Application of Integrated Multi-Objective Particles Swarm Optimization and Genetic Algorithm Metaheuristic through Fuzzy Logic for Optimization of Vehicle Routing Problems in Sugar Industry

Authors: Mukhtiar Singh, Sumeet Nagar

Abstract:

Vehicle routing problem (VRP) is a combinatorial optimization and nonlinear programming problem aiming to optimize decisions regarding given set of routes for a fleet of vehicles in order to provide cost-effective and efficient delivery of both services and goods to the intended customers. This paper proposes the application of integrated particle swarm optimization (PSO) and genetic optimization algorithm (GA) to address the Vehicle routing problem in sugarcane industry in India. Suger industry is very prominent agro-based industry in India due to its impacts on rural livelihood and estimated to be employing around 5 lakhs workers directly in sugar mills. Due to various inadequacies, inefficiencies and inappropriateness associated with the current vehicle routing model it costs huge money loss to the industry which needs to be addressed in proper context. The proposed algorithm utilizes the crossover operation that originally appears in genetic algorithm (GA) to improve its flexibility and manipulation more readily and avoid being trapped in local optimum, and simultaneously for improving the convergence speed of the algorithm, level set theory is also added to it. We employ the hybrid approach to an example of VRP and compare its result with those generated by PSO, GA, and parallel PSO algorithms. The experimental comparison results indicate that the performance of hybrid algorithm is superior to others, and it will become an effective approach for solving discrete combinatory problems.

Keywords: fuzzy logic, genetic algorithm, particle swarm optimization, vehicle routing problem

Procedia PDF Downloads 388
783 The Utilization of Salicylic Acid of the Extract from Avocado Skin as an Inhibitor of Ethylene Production to Keep the Quality of Banana in Storage

Authors: Adira Nofeadri Ryofi, Alvin Andrianus, Anna Khairunnisa, Anugrah Cahyo Widodo, Arbhyando Tri Putrananda, Arsy Imanda N. Raswati, Gita Rahmaningsih, Ina Agustina

Abstract:

The consumption level of fresh bananas from 2005 until 2010, increased from 8.2 to 10 kg/capita/year. The commercial scale of banana generally harvested when it still green to make the banana avoid physical damage, chemical, and disease after harvest and ripe fruit. That first metabolism activity can be used as a synthesis reaction. Ripening fruit was influenced by ethylene hormone that synthesized in fruit which is experiencing ripe and including hormone in the ripening fruit process in klimaterik phase. This ethylene hormone is affected by the respiration level that would speed up the restructuring of carbohydrates inside the fruit, so the weighting of fruit will be decreased. Compared to other klimaterik fruit, banana is a fruit that has a medium ethylene production rate and the rate of respiration is low. The salicylic acid can regulate the result number of the growth process or the development of fruits and plants. Salicylic acid serves to hinder biosynthesis ethylene and delay senses. The research aims to understand the influence of salicylic acid concentration that derived from the waste of avocado skin in inhibition process to ethylene production that the maturation can be controlled, so it can keep the quality of banana for storage. It is also to increase the potential value of the waste of avocado skin that were still used in industrial cosmetics.

Keywords: ethylene hormone, extract avocado skin, inhibitor, salicylic acid

Procedia PDF Downloads 227
782 Correlation Between Ore Mineralogy and the Dissolution Behavior of K-Feldspar

Authors: Adrian Keith Caamino, Sina Shakibania, Lena Sunqvist-Öqvist, Jan Rosenkranz, Yousef Ghorbani

Abstract:

Feldspar minerals are one of the main components of the earth’s crust. They are tectosilicate, meaning that they mainly contain aluminum and silicon. Besides aluminum and silicon, they contain either potassium, sodium, or calcium. Accordingly, feldspar minerals are categorized into three main groups: K-feldspar, Na-feldspar, and Ca-feldspar. In recent years, the trend to use K-feldspar has grown tremendously, considering its potential to produce potash and alumina. However, the feldspar minerals, in general, are difficult to decompose for the dissolution of their metallic components. Several methods, including intensive milling, leaching under elevated pressure and temperature, thermal pretreatment, and the use of corrosive leaching reagents, have been proposed to improve its low dissolving efficiency. In this study, as part of the POTASSIAL EU project, to overcome the low dissolution efficiency of the K-feldspar components, mechanical activation using intensive milling followed by leaching using hydrochloric acid (HCl) was practiced. Grinding operational parameters, namely time, rotational speed, and ball-to-sample weight ratio, were studied using the Taguchi optimization method. Then, the mineralogy of the grinded samples was analyzed using a scanning electron microscope (SEM) equipped with automated quantitative mineralogy. After grinding, the prepared samples were subjected to HCl leaching. In the end, the dissolution efficiency of the main elements and impurities of different samples were correlated to the mineralogical characterization results. K-feldspar component dissolution is correlated with ore mineralogy, which provides insight into how to best optimize leaching conditions for selective dissolution. Further, it will have an effect on purifying steps taken afterward and the final value recovery procedures

Keywords: K-feldspar, grinding, automated mineralogy, impurity, leaching

Procedia PDF Downloads 69
781 The Research on Diesel Bus Emissions in Ulaanbaatar City: Mongolia

Authors: Tsetsegmaa A., Bayarsuren B., Altantsetseg Ts.

Abstract:

To make the best decision on reducing harmful emissions from buses, we need to have a clear understanding of the current state of their actual emissions. The emissions from city buses running on high sulfur fuel, particularly particulate matter (PM) and nitrogen oxides (NOx) from the exhaust gases of conventional diesel engines, have been studied and measured with and without diesel particulate filter (DPF) in Ulaanbaatar city. The study was conducted by using the PEMS (Portable Emissions Measurement System) and gravimetric method in real traffic conditions. The obtained data were used to determine the actual emission rates and to evaluate the effectiveness of the selected particulate filters. Actual road and daily PM emissions from city buses were determined during the warm and cold seasons. A bus with an average daily mileage of 242 km was found to emit 166.155 g of PM into the city's atmosphere on average per day, with 141.3 g in summer and 175.8 g in winter. The actual PM of the city bus is 0.6866 g/km. The concentration of NOx in the exhaust gas averages 1410.94 ppm. The use of DPF reduced the exhaust gas opacity of 24 buses by an average of 97% and filtered a total of 340.4 kg of soot from these buses over a period of six months. Retrofitting an old conventional diesel engine with cassette-type silicon carbide (SiC) DPF, despite the laboriousness of cleaning, can significantly reduce particulate matter emissions. Innovation: First comprehensive road PM and NOx emission dataset and actual road emissions from public buses have been identified. PM and NOx mathematical model equations have been estimated as a function of the bus technical speed and engine revolution with and without DPF.

Keywords: conventional diesel, silicon carbide, real-time onboard measurements, particulate matter, diesel retrofit, fuel sulphur

Procedia PDF Downloads 149
780 Performance Analysis and Multi-Objective Optimization of a Kalina Cycle for Low-Temperature Applications

Authors: Sadegh Sadeghi, Negar Shabani

Abstract:

From a thermal point of view, zeotropic mixtures are likely to be more efficient than azeotropic fluids in low-temperature thermodynamic cycles due to their suitable boiling characteristics. In this study, performance of a low-temperature Kalina cycle with R717/water working fluid used in different existing power plants is mathematically investigated. To analyze the behavior of the cycle, mass conservation, energy conservation, and exergy balance equations are presented. With regard to the similarity in molar mass of R717 (17.03 gr/mol) and water (18.01 gr/mol), there is no need to alter the size of Kalina system components such as turbine and pump. To optimize the cycle energy and exergy efficiencies simultaneously, a constrained multi-objective optimization is carried out applying an Artificial Bee Colony algorithm. The main motivation behind using this algorithm lies on its robustness, reliability, remarkable precision and high–speed convergence rate in dealing with complicated constrained multi-objective problems. Convergence rates of the algorithm for calculating the optimal energy and exergy efficiencies are presented. Subsequently, due to the importance of exergy concept in Kalina cycles, exergy destructions occurring in the components are computed. Finally, the impacts of pressure, temperature, mass fraction and mass flow rate on the energy and exergy efficiencies are elaborately studied.

Keywords: artificial bee colony algorithm, binary zeotropic mixture, constrained multi-objective optimization, energy efficiency, exergy efficiency, Kalina cycle

Procedia PDF Downloads 148
779 Rapid Separation of Biomolecules and Neutral Analytes with a Cationic Stationary Phase by Capillary Electrochromatography

Authors: A. Aslihan Gokaltun, Ali Tuncel

Abstract:

The unique properties of capillary electrochromatography (CEC) such as high performance, high selectivity, low consumption of both reagents and analytes ensure this technique an attractive one for the separation of biomolecules including nucleosides and nucleotides, peptides, proteins, carbohydrates. Monoliths have become a well-established separation media for CEC in the format that can be compared to a single large 'particle' that does not include interparticular voids. Convective flow through the pores of monolith significantly accelerates the rate of mass transfer and enables a substantial increase in the speed of the separation. In this work, we propose a new approach for the preparation of cationic monolithic stationary phase for capillary electrochromatography. Instead of utilizing a charge bearing monomer during polymerization, the desired charge-bearing group is generated on the capillary monolith after polymerization by using the reactive moiety of the monolithic support via one-pot, simple reaction. Optimized monolithic column compensates the disadvantages of frequently used reversed phases, which are difficult for separation of polar solutes. Rapid separation and high column efficiencies are achieved for the separation of neutral analytes, nucleic acid bases and nucleosides in reversed phase mode. Capillary monolith showed satisfactory hydrodynamic permeability and mechanical stability with relative standard deviation (RSD) values below 2 %. A new promising, reactive support that has a 'ligand selection flexibility' due to its reactive functionality represent a new family of separation media for CEC.

Keywords: biomolecules, capillary electrochromatography, cationic monolith, neutral analytes

Procedia PDF Downloads 208
778 Surface Roughness Prediction Using Numerical Scheme and Adaptive Control

Authors: Michael K.O. Ayomoh, Khaled A. Abou-El-Hossein., Sameh F.M. Ghobashy

Abstract:

This paper proposes a numerical modelling scheme for surface roughness prediction. The approach is premised on the use of 3D difference analysis method enhanced with the use of feedback control loop where a set of adaptive weights are generated. The surface roughness values utilized in this paper were adapted from [1]. Their experiments were carried out using S55C high carbon steel. A comparison was further carried out between the proposed technique and those utilized in [1]. The experimental design has three cutting parameters namely: depth of cut, feed rate and cutting speed with twenty-seven experimental sample-space. The simulation trials conducted using Matlab software is of two sub-classes namely: prediction of the surface roughness readings for the non-boundary cutting combinations (NBCC) with the aid of the known surface roughness readings of the boundary cutting combinations (BCC). The following simulation involved the use of the predicted outputs from the NBCC to recover the surface roughness readings for the boundary cutting combinations (BCC). The simulation trial for the NBCC attained a state of total stability in the 7th iteration i.e. a point where the actual and desired roughness readings are equal such that error is minimized to zero by using a set of dynamic weights generated in every following simulation trial. A comparative study among the three methods showed that the proposed difference analysis technique with adaptive weight from feedback control, produced a much accurate output as against the abductive and regression analysis techniques presented in this.

Keywords: Difference Analysis, Surface Roughness; Mesh- Analysis, Feedback control, Adaptive weight, Boundary Element

Procedia PDF Downloads 613
777 Effect of Friction Pressure on the Properties of Friction Welded Aluminum–Ceramic Dissimilar Joints

Authors: Fares Khalfallah, Zakaria Boumerzoug, Selvarajan Rajakumar, Elhadj Raouache

Abstract:

The ceramic-aluminum bond is strongly present in industrial tools, due to the need to combine the properties of metals, such as ductility, thermal and electrical conductivity, with ceramic properties like high hardness, corrosion and wear resistance. In recent years, some joining techniques have been developed to achieve a good bonding between these materials such as brazing, diffusion bonding, ultrasonic joining and friction welding. In this work, AA1100 aluminum alloy rods were welded with Alumina 99.9 wt% ceramic rods, by friction welding. The effect of friction pressure on mechanical and structural properties of welded joints was studied. The welding was performed by direct friction welding machine. The welding samples were rotated at a constant rotational speed of 900 rpm, friction time of 4 sec, forging strength of 18 MPa, and forging time of 3 sec. Three different friction pressures were applied to 20, 34 and 45 MPa. The three-point bending test and Vickers microhardness measurements were used to evaluate the strength of the joints and investigate the mechanical properties of the welding area. The microstructure of joints was examined by optical microscopy (OM), scanning electron microscopy (SEM) and energy dispersive spectroscopy (EDS). The results show that bending strength increased, and then decreased after reaching a maximum value, with increasing friction pressure. The SEM observation shows that the increase in friction pressure led to the appearance of cracks in the microstructure of the interface area, which is decreasing the bending strength of joints.

Keywords: welding of ceramic to aluminum, friction welding, alumina, AA1100 aluminum alloy

Procedia PDF Downloads 121
776 The Design of a Vehicle Traffic Flow Prediction Model for a Gauteng Freeway Based on an Ensemble of Multi-Layer Perceptron

Authors: Tebogo Emma Makaba, Barnabas Ndlovu Gatsheni

Abstract:

The cities of Johannesburg and Pretoria both located in the Gauteng province are separated by a distance of 58 km. The traffic queues on the Ben Schoeman freeway which connects these two cities can stretch for almost 1.5 km. Vehicle traffic congestion impacts negatively on the business and the commuter’s quality of life. The goal of this paper is to identify variables that influence the flow of traffic and to design a vehicle traffic prediction model, which will predict the traffic flow pattern in advance. The model will unable motorist to be able to make appropriate travel decisions ahead of time. The data used was collected by Mikro’s Traffic Monitoring (MTM). Multi-Layer perceptron (MLP) was used individually to construct the model and the MLP was also combined with Bagging ensemble method to training the data. The cross—validation method was used for evaluating the models. The results obtained from the techniques were compared using predictive and prediction costs. The cost was computed using combination of the loss matrix and the confusion matrix. The predicted models designed shows that the status of the traffic flow on the freeway can be predicted using the following parameters travel time, average speed, traffic volume and day of month. The implications of this work is that commuters will be able to spend less time travelling on the route and spend time with their families. The logistics industry will save more than twice what they are currently spending.

Keywords: bagging ensemble methods, confusion matrix, multi-layer perceptron, vehicle traffic flow

Procedia PDF Downloads 334
775 Analysis of a Multiejector Cooling System in a Truck at Different Loads

Authors: Leonardo E. Pacheco, Carlos A. Díaz

Abstract:

An alternative way of addressing the difficult to recover the useless heat is through an ejector refrigeration cycle for vehicles applications. A group of thermo-compressor supply the mechanical compressor function at conventional refrigeration compression system. The thermo-compressor group recovers the thermal energy from waste streams (exhaust gases product in internal combustion motors, gases burned in wellhead among others) to eliminate the power consumption of the mechanical compressor. These types of alternative cooling system (air-conditioners) present a kind of advantages in both the increase in energy efficiency and the improvement of the COP of the system being studied from their its mechanical simplicity (decrease of moving parts). An ejector refrigeration cycle represents a significant step forward in the optimization of the efficient use of energy in the process of air conditioning and an alternative to reduce the environmental impacts. On one side, with the energy recycling decreases the temperature of the gases thrown into the atmosphere, which contributes to the principal beneficiaries of the average temperature of the planet. In parallel, mitigating the environmental impact caused by the production and handling of conventional cooling fluids commonly available in the market, causing the destruction of the ozone layer. This work had studied the operation of the multiejector cooling system for a truck with a 420 HP engine at different rotation speed. The operation condition limits and the COP of multi-ejector cooling systems applied in a truck are analyzed for a variable rpm range from to 800–1800 rpm.

Keywords: ejector system, exhaust gas, multiejector cooling system, recovery energy

Procedia PDF Downloads 252
774 Partnership Oriented Innovation Alliance Strategy Based on Market Feedback

Authors: Victor Romanov, Daria Efimenko

Abstract:

The focus on innovation in modern economy is the main factor in surviving business in a competitive environment. The innovations are based on the search and use of knowledge in a global context. Nowadays consumers and market demand are the main innovation drivers. This leads to build a business as a system with feedback, promptly restructuring production and innovation implementation in response to market demands. In modern knowledge economy, because of speed of technical progress, the product's lifecycle became much shorter, what makes more stringent requirements for innovation implementation on the enterprises of and therefore the possibility for enterprise for receiving extra income is decreasing. This circumstance imposes additional requirements for the replacement of obsolete products and the prompt release of innovative products to the market. The development of information technologies has led to the fact that only in the conditions of partnership and knowledge sharing with partners it is possible to update products quickly for innovative products. Many companies pay attention to updating innovations through the search for new partners, but the task of finding new partners presents some difficulties. The search for a suitable one includes several stages such as: determining the moment of innovation-critical, introducing a search, identifying search criteria, justifying and deciding on the choice of a partner. No less important is the question of how to manage an innovative product in response to a changing market. The article considers the problems of information support for the search for the source of innovation and partnership to decrease the time for implementation of novelty products.

Keywords: partnership, novelty, market feedback, alliance

Procedia PDF Downloads 187
773 Optimization of Bifurcation Performance on Pneumatic Branched Networks in next Generation Soft Robots

Authors: Van-Thanh Ho, Hyoungsoon Lee, Jaiyoung Ryu

Abstract:

Efficient pressure distribution within soft robotic systems, specifically to the pneumatic artificial muscle (PAM) regions, is essential to minimize energy consumption. This optimization involves adjusting reservoir pressure, pipe diameter, and branching network layout to reduce flow speed and pressure drop while enhancing flow efficiency. The outcome of this optimization is a lightweight power source and reduced mechanical impedance, enabling extended wear and movement. To achieve this, a branching network system was created by combining pipe components and intricate cross-sectional area variations, employing the principle of minimal work based on a complete virtual human exosuit. The results indicate that modifying the cross-sectional area of the branching network, gradually decreasing it, reduces velocity and enhances momentum compensation, preventing flow disturbances at separation regions. These optimized designs achieve uniform velocity distribution (uniformity index > 94%) prior to entering the connection pipe, with a pressure drop of less than 5%. The design must also consider the length-to-diameter ratio for fluid dynamic performance and production cost. This approach can be utilized to create a comprehensive PAM system, integrating well-designed tube networks and complex pneumatic models.

Keywords: pneumatic artificial muscles, pipe networks, pressure drop, compressible turbulent flow, uniformity flow, murray's law

Procedia PDF Downloads 63
772 Multi Response Optimization in Drilling Al6063/SiC/15% Metal Matrix Composite

Authors: Hari Singh, Abhishek Kamboj, Sudhir Kumar

Abstract:

This investigation proposes a grey-based Taguchi method to solve the multi-response problems. The grey-based Taguchi method is based on the Taguchi’s design of experimental method, and adopts Grey Relational Analysis (GRA) to transfer multi-response problems into single-response problems. In this investigation, an attempt has been made to optimize the drilling process parameters considering weighted output response characteristics using grey relational analysis. The output response characteristics considered are surface roughness, burr height and hole diameter error under the experimental conditions of cutting speed, feed rate, step angle, and cutting environment. The drilling experiments were conducted using L27 orthogonal array. A combination of orthogonal array, design of experiments and grey relational analysis was used to ascertain best possible drilling process parameters that give minimum surface roughness, burr height and hole diameter error. The results reveal that combination of Taguchi design of experiment and grey relational analysis improves surface quality of drilled hole.

Keywords: metal matrix composite, drilling, optimization, step drill, surface roughness, burr height, hole diameter error

Procedia PDF Downloads 308
771 Estimation of Twist Loss in the Weft Yarn during Air-Jet Weft Insertion

Authors: Muhammad Umair, Yasir Nawab, Khubab Shaker, Muhammad Maqsood, Adeel Zulfiqar, Danish Mahmood Baitab

Abstract:

Fabric is a flexible woven material consisting of a network of natural or artificial fibers often referred to as thread or yarn. Today fabrics are produced by weaving, braiding, knitting, tufting and non-woven. Weaving is a method of fabric production in which warp and weft yarns are interlaced perpendicular to each other. There is infinite number of ways for the interlacing of warp and weft yarn. Each way produces a different fabric structure. The yarns parallel to the machine direction are called warp yarns and the yarns perpendicular to the machine direction are called weft or filling yarns. Air jet weaving is the modern method of weft insertion and considered as high speed loom. The twist loss in air jet during weft insertion affects the strength. The aim of this study was to investigate the effect of twist change in weft yarn during air-jet weft insertion. A total number of 8 samples were produced using 1/1 plain and 3/1 twill weave design with two fabric widths having same loom settings. Two different types of yarns like cotton and PC blend were used. The effect of material type, weave design and fabric width on twist change of weft yarn was measured and discussed. Twist change in the different types of weft yarn and weave design was measured and compared the twist change in the weft yarn with the yarn before weft yarn insertion and twist loss is measured. Wider fabric leads to higher twist loss in the yarn.

Keywords: air jet loom, twist per inch, twist loss, weft yarn

Procedia PDF Downloads 390
770 Rainstorm Characteristics over the Northeastern Region of Thailand: Weather Radar Analysis

Authors: P. Intaracharoen, P. Chantraket, C. Detyothin, S. Kirtsaeng

Abstract:

Radar reflectivity data from Phimai weather radar station of DRRAA (Department of Royal Rainmaking and Agricultural Aviation) were used to analyzed the rainstorm characteristics via Thunderstorm Identification Tracking Analysis and Nowcasting (TITAN) algorithm. The Phimai weather radar station was situated at Nakhon Ratchasima province, northeastern Thailand. The data from 277 days of rainstorm events occurring from May 2016 to May 2017 were used to investigate temporal distribution characteristics of convective individual rainclouds. The important storm properties, structures, and their behaviors were analyzed by 9 variables as storm number, storm duration, storm volume, storm area, storm top, storm base, storm speed, storm orientation, and maximum storm reflectivity. The rainstorm characteristics were also examined by separating the data into two periods as wet and dry season followed by an announcement of TMD (Thai Meteorological Department), under the influence of southwest monsoon (SWM) and northeast monsoon (NEM). According to the characteristics of rainstorm results, it can be seen that rainstorms during the SWM influence were found to be the most potential rainstorms over northeastern region of Thailand. The SWM rainstorms are larger number of the storm (404, 140 no./day), storm area (34.09, 26.79 km²) and storm volume (95.43, 66.97 km³) than NEM rainstorms, respectively. For the storm duration, the average individual storm duration during the SWM and NEM was found a minor difference in both periods (47.6, 48.38 min) and almost all storm duration in both periods were less than 3 hours. The storm velocity was not exceeding 15 km/hr (13.34 km/hr for SWM and 10.67 km/hr for NEM). For the rainstorm reflectivity, it was found a little difference between wet and dry season (43.08 dBz for SWM and 43.72 dBz for NEM). It assumed that rainstorms occurred in both seasons have same raindrop size.

Keywords: rainstorm characteristics, weather radar, TITAN, Northeastern Thailand

Procedia PDF Downloads 182
769 Sentiment Analysis of Chinese Microblog Comments: Comparison between Support Vector Machine and Long Short-Term Memory

Authors: Xu Jiaqiao

Abstract:

Text sentiment analysis is an important branch of natural language processing. This technology is widely used in public opinion analysis and web surfing recommendations. At present, the mainstream sentiment analysis methods include three parts: sentiment analysis based on a sentiment dictionary, based on traditional machine learning, and based on deep learning. This paper mainly analyzes and compares the advantages and disadvantages of the SVM method of traditional machine learning and the Long Short-term Memory (LSTM) method of deep learning in the field of Chinese sentiment analysis, using Chinese comments on Sina Microblog as the data set. Firstly, this paper classifies and adds labels to the original comment dataset obtained by the web crawler, and then uses Jieba word segmentation to classify the original dataset and remove stop words. After that, this paper extracts text feature vectors and builds document word vectors to facilitate the training of the model. Finally, SVM and LSTM models are trained respectively. After accuracy calculation, it can be obtained that the accuracy of the LSTM model is 85.80%, while the accuracy of SVM is 91.07%. But at the same time, LSTM operation only needs 2.57 seconds, SVM model needs 6.06 seconds. Therefore, this paper concludes that: compared with the SVM model, the LSTM model is worse in accuracy but faster in processing speed.

Keywords: sentiment analysis, support vector machine, long short-term memory, Chinese microblog comments

Procedia PDF Downloads 81
768 Using LTE-Sim in New Hanover Decision Algorithm for 2-Tier Macrocell-Femtocell LTE Network

Authors: Umar D. M., Aminu A. M., Izaddeen K. Y.

Abstract:

Deployments of mini macrocell base stations also referred to as femtocells, improve the quality of service of indoor and outdoor users. Nevertheless, mobility management remains a key issue with regards to their deployment. This paper is leaned towards this issue, with an in-depth focus on the most important aspect of mobility management -handover. In handover management, making a handover decision in the LTE two-tier macrocell femtocell network is a crucial research area. Decision algorithms in this research are classified and comparatively analyzed according to received signal strength, user equipment speed, cost function, and interference. However, it was observed that most of the discussed decision algorithms fail to consider cell selection with hybrid access policy in a single macrocell multiple femtocell scenario, another observation was a majority of these algorithms lack the incorporation of user equipment residence parameter. Not including this parameter boosts the number of unnecessary handover occurrence. To deal with these issues, a sophisticated handover decision algorithm is proposed. The proposed algorithm considers the user’s velocity, received signal strength, residence time, as well as the femtocell base station’s access policy. Simulation results have shown that the proposed algorithm reduces the number of unnecessary handovers when compared to conventional received signal strength-based handover decision algorithm.

Keywords: user-equipment, radio signal service, long term evolution, mobility management, handoff

Procedia PDF Downloads 113
767 Conceptual Solution and Thermal Analysis of the Final Cooling Process of Biscuits in One Confectionary Factory in Serbia

Authors: Duško Salemović, Aleksandar Dedić, Matilda Lazić, Dragan Halas

Abstract:

The paper presents the conceptual solution for the final cooling of the chocolate dressing of biscuits in one confectionary factory in Serbia. The proposed concept solution was derived from the desired technological process of final cooling of biscuits and the required process parameters that were to be achieved, and which were an integral part of the project task. The desired process parameters for achieving proper hardening and coating formation are the exchanged amount of heat in the time unit between the two media (air and chocolate dressing), the speed of air inside the tunnel cooler, and the surface of all biscuits in contact with the air. These parameters were calculated in the paper. The final cooling of chocolate dressing on biscuits could be optimized by changing process parameters and dimensions of the tunnel cooler and looking for the appropriate values for them. The accurate temperature predictions and fluid flow analysis could be conducted by using heat balance and flow balance equations, having in mind the theory of similarity. Furthermore, some parameters were adopted from previous technology processes, such as the inlet temperature of biscuits and input air temperature. A thermal calculation was carried out, and it was demonstrated that the percentage error between the contact surface of the air and the chocolate biscuit topping, which is obtained from the heat balance and geometrically through the proposed conceptual solution, does not exceed 0.67%, which is a very good agreement. This enabled the quality of the cooling process of chocolate dressing applied on the biscuit and the hardness of its coating.

Keywords: chocolate dressing, air, cooling, heat balance

Procedia PDF Downloads 64
766 Study of the Persian Gulf’s and Oman Sea’s Numerical Tidal Currents

Authors: Fatemeh Sadat Sharifi

Abstract:

In this research, a barotropic model was employed to consider the tidal studies in the Persian Gulf and Oman Sea, where the only sufficient force was the tidal force. To do that, a finite-difference, free-surface model called Regional Ocean Modeling System (ROMS), was employed on the data over the Persian Gulf and Oman Sea. To analyze flow patterns of the region, the results of limited size model of The Finite Volume Community Ocean Model (FVCOM) were appropriated. The two points were determined since both are one of the most critical water body in case of the economy, biology, fishery, Shipping, navigation, and petroleum extraction. The OSU Tidal Prediction Software (OTPS) tide and observation data validated the modeled result. Next, tidal elevation and speed, and tidal analysis were interpreted. Preliminary results determine a significant accuracy in the tidal height compared with observation and OTPS data, declaring that tidal currents are highest in Hormuz Strait and the narrow and shallow region between Iranian coasts and Islands. Furthermore, tidal analysis clarifies that the M_2 component has the most significant value. Finally, the Persian Gulf tidal currents are divided into two branches: the first branch converts from south to Qatar and via United Arab Emirate rotates to Hormuz Strait. The secondary branch, in north and west, extends up to the highest point in the Persian Gulf and in the head of Gulf turns counterclockwise.

Keywords: numerical model, barotropic tide, tidal currents, OSU tidal prediction software, OTPS

Procedia PDF Downloads 122
765 Perforation Analysis of the Aluminum Alloy Sheets Subjected to High Rate of Loading and Heated Using Thermal Chamber: Experimental and Numerical Approach

Authors: A. Bendarma, T. Jankowiak, A. Rusinek, T. Lodygowski, M. Klósak, S. Bouslikhane

Abstract:

The analysis of the mechanical characteristics and dynamic behavior of aluminum alloy sheet due to perforation tests based on the experimental tests coupled with the numerical simulation is presented. The impact problems (penetration and perforation) of the metallic plates have been of interest for a long time. Experimental, analytical as well as numerical studies have been carried out to analyze in details the perforation process. Based on these approaches, the ballistic properties of the material have been studied. The initial and residual velocities laser sensor is used during experiments to obtain the ballistic curve and the ballistic limit. The energy balance is also reported together with the energy absorbed by the aluminum including the ballistic curve and ballistic limit. The high speed camera helps to estimate the failure time and to calculate the impact force. A wide range of initial impact velocities from 40 up to 180 m/s has been covered during the tests. The mass of the conical nose shaped projectile is 28 g, its diameter is 12 mm, and the thickness of the aluminum sheet is equal to 1.0 mm. The ABAQUS/Explicit finite element code has been used to simulate the perforation processes. The comparison of the ballistic curve was obtained numerically and was verified experimentally, and the failure patterns are presented using the optimal mesh densities which provide the stability of the results. A good agreement of the numerical and experimental results is observed.

Keywords: aluminum alloy, ballistic behavior, failure criterion, numerical simulation

Procedia PDF Downloads 304
764 Thermal and Mechanical Properties of Polycaprolactone-Soy Lecithin Modified Bentonite Nanocomposites

Authors: Danila Merino, Leandro N. Ludueña, Vera A. Alvarez

Abstract:

Clays are commonly used to reinforce polymeric materials. In order to modify them, long-chain quaternary-alkylammonium salts have been widely employed. However, the application of these clays in biological fields is limited by the toxicity and poor biocompatibility presented by these modifiers. Meanwhile, soy lecithin, acts as a natural biosurfactant and environment-friendly biomodifier. In this report, we analyse the effect of content of soy lecithin-modified bentonite on the properties of polycaprolactone (PCL) nanocomposites. Commercial grade PCL (CAPA FB 100) was supplied by Perstorp, with Mw = 100000 g/mol. Minarmco S.A. and Melar S.A supplied bentonite and soy lecithin, respectively. Clays with 18, 30 and 45 wt% of organic content were prepared by exchanging 4 g of Na-Bent with 1, 2 and 4 g of soy lecithin aqueous and acid solution (pH=1, with HCl) at 75ºC for 2 h. Then, they were washed and lyophilized for 72 h. Samples were labeled A, B and C. Nanocomposites with 1 and 2 wt.% of each clay were prepared by melt-intercalation followed by compression-moulding. An intensive Brabender type mixer with two counter-rotating roller rotors was used. Mixing temperature was 100 ºC; speed of rotation was 100 rpm. and mixing time was 10 min. Compression moulding was carried out in a hydraulic press under 75 Kg/mm2 for 10 minutes at 100 ºC. The thickness of the samples was about 1 mm. Thermal and mechanical properties were analysed. PCL nanocomposites with 1 and 2% of B presented the best mechanical properties. It was observed that an excessive organic content produced an increment on the rigidity of PCL, but caused a detrimental effect on the tensile strength and elongation at break of the nanocomposites. Thermogravimetrical analyses suggest that all reinforced samples have higher resistance to degradation than neat PCL.

Keywords: chemical modification, clay, nanocomposite, characterization

Procedia PDF Downloads 191
763 Cache Analysis and Software Optimizations for Faster on-Chip Network Simulations

Authors: Khyamling Parane, B. M. Prabhu Prasad, Basavaraj Talawar

Abstract:

Fast simulations are critical in reducing time to market in CMPs and SoCs. Several simulators have been used to evaluate the performance and power consumed by Network-on-Chips. Researchers and designers rely upon these simulators for design space exploration of NoC architectures. Our experiments show that simulating large NoC topologies take hours to several days for completion. To speed up the simulations, it is necessary to investigate and optimize the hotspots in simulator source code. Among several simulators available, we choose Booksim2.0, as it is being extensively used in the NoC community. In this paper, we analyze the cache and memory system behaviour of Booksim2.0 to accurately monitor input dependent performance bottlenecks. Our measurements show that cache and memory usage patterns vary widely based on the input parameters given to Booksim2.0. Based on these measurements, the cache configuration having least misses has been identified. To further reduce the cache misses, we use software optimization techniques such as removal of unused functions, loop interchanging and replacing post-increment operator with pre-increment operator for non-primitive data types. The cache misses were reduced by 18.52%, 5.34% and 3.91% by employing above technology respectively. We also employ thread parallelization and vectorization to improve the overall performance of Booksim2.0. The OpenMP programming model and SIMD are used for parallelizing and vectorizing the more time-consuming portions of Booksim2.0. Speedups of 2.93x and 3.97x were observed for the Mesh topology with 30 × 30 network size by employing thread parallelization and vectorization respectively.

Keywords: cache behaviour, network-on-chip, performance profiling, vectorization

Procedia PDF Downloads 186