Search results for: location routing problem
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 9201

Search results for: location routing problem

5151 Sloshing Response of Liquid in Prismatic Container under Oscillation

Authors: P. R. Maiti, S. K. Bhattacharyya

Abstract:

Sloshing is a physical phenomenon characterized by the oscillation of unrestrained free surface of liquid in a partially liquid filled container subjected to external excitation. Determination of sloshing frequency in container is important to avoid resonance condition of the system. The complex behavior of the free surface movement and its combined mode of vibration make difficulty for exact analysis of sloshing. In the present study, numerical analysis is carried out for a partially liquid filled tank under external forces. Boundary element approach is used to formulate the sloshing problem in two -dimensional prismatic container with potential flow. Effort has been made to find slosh response for two dimensional problems in partially liquid filled prismatic container.

Keywords: sloshing, boundary element method, prismatic container, oscillation

Procedia PDF Downloads 312
5150 An Approach to Autonomous Drones Using Deep Reinforcement Learning and Object Detection

Authors: K. R. Roopesh Bharatwaj, Avinash Maharana, Favour Tobi Aborisade, Roger Young

Abstract:

Presently, there are few cases of complete automation of drones and its allied intelligence capabilities. In essence, the potential of the drone has not yet been fully utilized. This paper presents feasible methods to build an intelligent drone with smart capabilities such as self-driving, and obstacle avoidance. It does this through advanced Reinforcement Learning Techniques and performs object detection using latest advanced algorithms, which are capable of processing light weight models with fast training in real time instances. For the scope of this paper, after researching on the various algorithms and comparing them, we finally implemented the Deep-Q-Networks (DQN) algorithm in the AirSim Simulator. In future works, we plan to implement further advanced self-driving and object detection algorithms, we also plan to implement voice-based speech recognition for the entire drone operation which would provide an option of speech communication between users (People) and the drone in the time of unavoidable circumstances. Thus, making drones an interactive intelligent Robotic Voice Enabled Service Assistant. This proposed drone has a wide scope of usability and is applicable in scenarios such as Disaster management, Air Transport of essentials, Agriculture, Manufacturing, Monitoring people movements in public area, and Defense. Also discussed, is the entire drone communication based on the satellite broadband Internet technology for faster computation and seamless communication service for uninterrupted network during disasters and remote location operations. This paper will explain the feasible algorithms required to go about achieving this goal and is more of a reference paper for future researchers going down this path.

Keywords: convolution neural network, natural language processing, obstacle avoidance, satellite broadband technology, self-driving

Procedia PDF Downloads 239
5149 On Confidence Intervals for the Difference between Inverse of Normal Means with Known Coefficients of Variation

Authors: Arunee Wongkhao, Suparat Niwitpong, Sa-aat Niwitpong

Abstract:

In this paper, we propose two new confidence intervals for the difference between the inverse of normal means with known coefficients of variation. One of these two confidence intervals for this problem is constructed based on the generalized confidence interval and the other confidence interval is constructed based on the closed form method of variance estimation. We examine the performance of these confidence intervals in terms of coverage probabilities and expected lengths via Monte Carlo simulation.

Keywords: coverage probability, expected length, inverse of normal mean, coefficient of variation, generalized confidence interval, closed form method of variance estimation

Procedia PDF Downloads 301
5148 Study on Roll Marks of Stainless Steel in Rolling Mill

Authors: Cai-Wan Chang-Jian, Han-Ting Tsai

Abstract:

In the processing industry of metal forming, rolling is the most used method of processing. In a cold rolling factory of stainless steel, there occurs a product defect on temper rolling process within cold rolling. It is called 'roll marks', which is a phenomenon of undesirable flatness problem. In this research, we performed a series of experimental measurements on the roll marks, and we used optical sensors to measure it and compared the vibration frequency of roll marks with the vibration frequency of key components in the skin pass mill. We found there is less correlation between the above mentioned data. Finally, we took measurement on the motor driver in rolling mill. We found that the undulation frequency of motor could match with the frequency of roll marks, and then we have confirmed that the motor’s undulation caused roll marks.

Keywords: roll mark, plane strain, rolling mill, stainless steel

Procedia PDF Downloads 445
5147 A Laundry Algorithm for Colored Textiles

Authors: H. E. Budak, B. Arslan-Ilkiz, N. Cakmakci, I. Gocek, U. K. Sahin, H. Acikgoz-Tufan, M. H. Arslan

Abstract:

The aim of this study is to design a novel laundry algorithm for colored textiles which have significant decoloring problem. During the experimental work, bleached knitted single jersey fabric made of 100% cotton and dyed with reactive dyestuff was utilized, since according to a conducted survey textiles made of cotton are the most demanded textile products in the textile market by the textile consumers and for coloration of textiles reactive dyestuffs are the ones that are the most commonly used in the textile industry for dyeing cotton-made products. Therefore, the fabric used in this study was selected and purchased in accordance with the survey results. The fabric samples cut out of this fabric were dyed with different dyeing parameters by using Remazol Brilliant Red 3BS dyestuff in Gyrowash machine at laboratory conditions. From the alternative reactive-dyed cotton fabric samples, the ones that have high tendency to color loss were determined and examined. Accordingly, the parameters of the dyeing process used for these fabric samples were evaluated and the dyeing process which was chosen to be used for causing high tendency to color loss for the cotton fabrics was determined in order to reveal the level of improvement in color loss during this study clearly. Afterwards, all of the untreated fabric samples cut out of the fabric purchased were dyed with the dyeing process selected. When dyeing process was completed, an experimental design was created for the laundering process by using Minitab® program considering temperature, time and mechanical action as parameters. All of the washing experiments were performed in domestic washing machine. 16 washing experiments were performed with 8 different experimental conditions and 2 repeats for each condition. After each of the washing experiments, water samples of the main wash of the laundering process were measured with UV spectrophotometer. The values obtained were compared with the calibration curve of the materials used for the dyeing process. The results of the washing experiments were statistically analyzed with Minitab® program. According to the results, the most suitable washing algorithm to be used in terms of the parameters temperature, time and mechanical action for domestic washing machines for minimizing fabric color loss was chosen. The laundry algorithm proposed in this study have the ability of minimalizing the problem of color loss of colored textiles in washing machines by eliminating the negative effects of the parameters of laundering process on color of textiles without compromising the fundamental effects of basic cleaning action being performed properly. Therefore, since fabric color loss is minimized with this washing algorithm, dyestuff residuals will definitely be lower in the grey water released from the laundering process. In addition to this, with this laundry algorithm it is possible to wash and clean other types of textile products with proper cleaning effect and minimized color loss.

Keywords: color loss, laundry algorithm, textiles, domestic washing process

Procedia PDF Downloads 340
5146 Analysis of the Contribution of Coastal and Marine Physical Factors to Oil Slick Movement: Case Study of Misrata, Libya

Authors: Abduladim Maitieg, Mark Johnson

Abstract:

Developing a coastal oil spill management plan for the Misratah coast is the motivating factor for building a database for coastal and marine systems and energy resources. Wind direction and speed, currents, bathymetry, coastal topography and offshore dynamics influence oil spill deposition in coastal water. Therefore, oceanographic and climatological data can be used to understand oil slick movement and potential oil deposits on shoreline area and the behaviour of oil spill trajectories on the sea surface. The purpose of this study is to investigate the effects of the coastal and marine physical factors under strong wave conditions and various bathymetric and coastal topography gradients in the western coastal area of Libya on the movement of oil slicks. The movement of oil slicks was computed using a GNOME simulation model based on current and wind speed/direction. The results in this paper show that (1) Oil slick might reach the Misratah shoreline area in two days in the summer and winter. Seasons. (2 ) The North coast of Misratah is the potential oil deposit area on the Misratah coast. (3) Tarball pollution was observed along the North coast of Misratah. (4) Two scenarios for the summer and the winter season were run, along the western coast of Libya . (5) The eastern coast is at a lower potential risk due to the influence of wind and current energy in the Gulf of Sidra. (6) The Misratah coastline is more vulnerable to oil spill movement in the summer than in winter seasons. (7) Oil slick takes from 2 to 5 days to reach the saltmarsh in the eastern Misratah coast. (8) Oil slick moves 300 km in 30 days from the spill resource location near the Libyan western border to the Misratah coast.(9) Bathymetric features have a profound effect on oil spill movement. (9)Oil dispersion simulations using GNOME are carried out taking into account high-resolution wind and current data.

Keywords: oil spill movement, coastal and marine physical factors, coast area, Libyan

Procedia PDF Downloads 217
5145 Modern Imputation Technique for Missing Data in Linear Functional Relationship Model

Authors: Adilah Abdul Ghapor, Yong Zulina Zubairi, Rahmatullah Imon

Abstract:

Missing value problem is common in statistics and has been of interest for years. This article considers two modern techniques in handling missing data for linear functional relationship model (LFRM) namely the Expectation-Maximization (EM) algorithm and Expectation-Maximization with Bootstrapping (EMB) algorithm using three performance indicators; namely the mean absolute error (MAE), root mean square error (RMSE) and estimated biased (EB). In this study, we applied the methods of imputing missing values in the LFRM. Results of the simulation study suggest that EMB algorithm performs much better than EM algorithm in both models. We also illustrate the applicability of the approach in a real data set.

Keywords: expectation-maximization, expectation-maximization with bootstrapping, linear functional relationship model, performance indicators

Procedia PDF Downloads 384
5144 Development of Pre-Mitigation Measures and Its Impact on Life-Cycle Cost of Facilities: Indian Scenario

Authors: Mahima Shrivastava, Soumya Kar, B. Swetha Malika, Lalu Saheb, M. Muthu Kumar, P. V. Ponambala Moorthi

Abstract:

Natural hazards and manmade destruction causes both economic and societal losses. Generalized pre-mitigation strategies introduced and adopted for prevention of disaster all over the world are capable of augmenting the resiliency and optimizing the life-cycle cost of facilities. In countries like India where varied topographical feature exists requires location specific mitigation measures and strategies to be followed for better enhancement by event-driven and code-driven approaches. Present state of vindication measures followed and adopted, lags dominance in accomplishing the required development. In addition, serious concern and debate over climate change plays a vital role in enhancing the need and requirement for the development of time bound adaptive mitigation measures. For the development of long-term sustainable policies incorporation of future climatic variation is inevitable. This will further assist in assessing the impact brought about by the climate change on life-cycle cost of facilities. This paper develops more definite region specific and time bound pre-mitigation measures, by reviewing the present state of mitigation measures in India and all over the world for improving life-cycle cost of facilities. For the development of region specific adoptive measures, Indian regions were divided based on multiple-calamity prone regions and geo-referencing tools were used to incorporate the effect of climate changes on life-cycle cost assessment. This study puts forward significant effort in establishing sustainable policies and helps decision makers in planning for pre-mitigation measures for different regions. It will further contribute towards evaluating the life cycle cost of facilities by adopting the developed measures.

Keywords: climate change, geo-referencing tools, life-cycle cost, multiple-calamity prone regions, pre-mitigation strategies, sustainable policies

Procedia PDF Downloads 370
5143 Performance of AquaCrop Model for Simulating Maize Growth and Yield Under Varying Sowing Dates in Shire Area, North Ethiopia

Authors: Teklay Tesfay, Gebreyesus Brhane Tesfahunegn, Abadi Berhane, Selemawit Girmay

Abstract:

Adjusting the proper sowing date of a crop at a particular location with a changing climate is an essential management option to maximize crop yield. However, determining the optimum sowing date for rainfed maize production through field experimentation requires repeated trials for many years in different weather conditions and crop management. To avoid such long-term experimentation to determine the optimum sowing date, crop models such as AquaCrop are useful. Therefore, the overall objective of this study was to evaluate the performance of AquaCrop model in simulating maize productivity under varying sowing dates. A field experiment was conducted for two consecutive cropping seasons by deploying four maize seed sowing dates in a randomized complete block design with three replications. Input data required to run this model are stored as climate, crop, soil, and management files in the AquaCrop database and adjusted through the user interface. Observed data from separate field experiments was used to calibrate and validate the model. AquaCrop model was validated for its performance in simulating the green canopy and aboveground biomass of maize for the varying sowing dates based on the calibrated parameters. Results of the present study showed that there was a good agreement (an overall R2 =, Ef= d= RMSE =) between measured and simulated values of the canopy cover and biomass yields. Considering the overall values of the statistical test indicators, the performance of the model to predict maize growth and biomass yield was successful, and so this is a valuable tool help for decision-making. Hence, this calibrated and validated model is suggested to use for determining optimum maize crop sowing date for similar climate and soil conditions to the study area, instead of conducting long-term experimentation.

Keywords: AquaCrop model, calibration, validation, simulation

Procedia PDF Downloads 52
5142 A Qualitative Approach to Engineering Design Issues, Problems, and Solutions

Authors: M. U. Arshid, M. A. Kamal

Abstract:

The engineering design process is the activities formulation, to help an engineer raising a plan with a specified goal and performance. The engineering design process is a multi-stage course of action including the conceptualization, research, feasibility studies, establishment of design parameters, preliminary and finally the detailed design. It is a progression from the abstract to the concrete; starting with probably abstract ideas about need, and thereafter elaborating detailed specifications of the object that would satisfy the needs, identified. Engineering design issues, problems, and solutions are discussed in this paper using qualitative approach from an information structure perspective. The objective is to identify the problems, to analyze them and propose solutions by integrating; innovation, practical experience, time and resource management, communications skills, isolating the problem in coordination with all stakeholders. Consequently, this would be beneficial for the engineering community to improve the Engineering design practices.

Keywords: engineering design, engineering design issues, innovation, public sector projects

Procedia PDF Downloads 334
5141 Comparison of Machine Learning Models for the Prediction of System Marginal Price of Greek Energy Market

Authors: Ioannis P. Panapakidis, Marios N. Moschakis

Abstract:

The Greek Energy Market is structured as a mandatory pool where the producers make their bid offers in day-ahead basis. The System Operator solves an optimization routine aiming at the minimization of the cost of produced electricity. The solution of the optimization problem leads to the calculation of the System Marginal Price (SMP). Accurate forecasts of the SMP can lead to increased profits and more efficient portfolio management from the producer`s perspective. Aim of this study is to provide a comparative analysis of various machine learning models such as artificial neural networks and neuro-fuzzy models for the prediction of the SMP of the Greek market. Machine learning algorithms are favored in predictions problems since they can capture and simulate the volatilities of complex time series.

Keywords: deregulated energy market, forecasting, machine learning, system marginal price

Procedia PDF Downloads 205
5140 Time Synchronization between the eNBs in E-UTRAN under the Asymmetric IP Network

Authors: M. Kollar, A. Zieba

Abstract:

In this paper, we present a method for a time synchronization between the two eNodeBs (eNBs) in E-UTRAN (Evolved Universal Terrestrial Radio Access) network. The two eNBs are cooperating in so-called inter eNB CA (Carrier Aggregation) case and connected via asymmetrical IP network. We solve the problem by using broadcasting signals generated in E-UTRAN as synchronization signals. The results show that the time synchronization with the proposed method is possible with the error significantly less than 1 ms which is sufficient considering the time transmission interval is 1 ms in E-UTRAN. This makes this method (with low complexity) more suitable than Network Time Protocol (NTP) in the mobile applications with generated broadcasting signals where time synchronization in asymmetrical network is required.

Keywords: IP scheduled throughput, E-UTRAN, Evolved Universal Terrestrial Radio Access Network, NTP, Network Time Protocol, assymetric network, delay

Procedia PDF Downloads 353
5139 Analyzing Social Media Discourses of Domestic Violence in Promoting Awareness and Support Seeking: An Exploratory Study

Authors: Sudha Subramani, Hua Wang

Abstract:

Domestic Violence (DV) against women is now recognized to be a serious and widespread problem worldwide. There is a growing concern that violence against women has a global public health impact, as well as a violation of human rights. From the existing statistical surveys, it is revealed that there exists a strong relationship between DV and health issues of women like bruising, lacerations, depression, anxiety, flashbacks, sleep disturbances, hyper-arousal, emotional distress, sexually transmitted diseases and so on. This social problem is still considered as behind the closed doors issue and stigmatized topic. Women conceal their sufferings from family and friends, as they experience a lack of trust in others, feelings of shame and embarrassment among the society. Hence, women survivors of DV experience some barriers in seeking the support of specialized services such as health care access, crisis support, and legal guidance. Fortunately, with the popularity of social media like Facebook and Twitter, people share their opinions and emotional feelings to seek the social and emotional support, for sympathetic encouragement, to show compassion and empathy among the public. Considering the DV, social media plays a predominant role in creating the awareness and promoting the support services to the public, as we live in the golden era of social media. The various professional people like the public health researchers, clinicians, psychologists, social workers, national family health organizations, lawyers, and victims or their family and friends share the unprecedentedly valuable information (personal opinions and experiences) in a single platform to improve the social welfare of the community. Though each tweet or post contains a less informational value, the consolidation of millions of messages can generate actionable knowledge and provide valuable insights about the public opinion in general. Hence, this paper reports on an exploratory analysis of the effectiveness of social media for unobtrusive assessment of attitudes and awareness towards DV. In this paper, mixed methods such as qualitative analysis and text mining approaches are used to understand the social media disclosures of DV through the lenses of opinion sharing, anonymity, and support seeking. The results of this study could be helpful to avoid the cost of wide scale surveys, while still maintaining appropriate research conditions is to leverage the abundance of data publicly available on the web. Also, this analysis with data enrichment and consolidation would be useful in assisting advocacy and national family health organizations to provide information about resources and support, raise awareness and counter common stigmatizing attitudes about DV.

Keywords: domestic violence, social media, social stigma and support, women health

Procedia PDF Downloads 275
5138 21st Century Islam: Global Challenges of Islamic Representation and Knowledge Acquisition

Authors: M. M. Muhammed, O. Khuzaima

Abstract:

This research examined and outlined some of the challenges facing Islam and Muslims in the 21st century, considering global Islamic representation and knowledge acquisition as key objectives. It was observed that the Western media misrepresentation of Islam and the Western ethos embodied by the acquisition of western civilisation are major challenges faced by Islam and Muslims today. The problem of sectarianism, decline in the socio-economic power of Muslim communities and the archaic nature of the Islamic creed were recorded as major actors to the evolving global Islamic issues. It was therefore concluded that Islam is not the reason for these challenges, rather the action of some Muslims and non-Muslims were the contributing factors to the pandemics faced by Islam and Muslims. Some relevant recommendations were made to the Islamic world that could serve as effectual solutions to these lingering problems.

Keywords: Islam, challenges, representation, knowledge, century, global, twenty-first

Procedia PDF Downloads 223
5137 Early Prediction of Diseases in a Cow for Cattle Industry

Authors: Ghufran Ahmed, Muhammad Osama Siddiqui, Shahbaz Siddiqui, Rauf Ahmad Shams Malick, Faisal Khan, Mubashir Khan

Abstract:

In this paper, a machine learning-based approach for early prediction of diseases in cows is proposed. Different ML algos are applied to extract useful patterns from the available dataset. Technology has changed today’s world in every aspect of life. Similarly, advanced technologies have been developed in livestock and dairy farming to monitor dairy cows in various aspects. Dairy cattle monitoring is crucial as it plays a significant role in milk production around the globe. Moreover, it has become necessary for farmers to adopt the latest early prediction technologies as the food demand is increasing with population growth. This highlight the importance of state-ofthe-art technologies in analyzing how important technology is in analyzing dairy cows’ activities. It is not easy to predict the activities of a large number of cows on the farm, so, the system has made it very convenient for the farmers., as it provides all the solutions under one roof. The cattle industry’s productivity is boosted as the early diagnosis of any disease on a cattle farm is detected and hence it is treated early. It is done on behalf of the machine learning output received. The learning models are already set which interpret the data collected in a centralized system. Basically, we will run different algorithms on behalf of the data set received to analyze milk quality, and track cows’ health, location, and safety. This deep learning algorithm draws patterns from the data, which makes it easier for farmers to study any animal’s behavioral changes. With the emergence of machine learning algorithms and the Internet of Things, accurate tracking of animals is possible as the rate of error is minimized. As a result, milk productivity is increased. IoT with ML capability has given a new phase to the cattle farming industry by increasing the yield in the most cost-effective and time-saving manner.

Keywords: IoT, machine learning, health care, dairy cows

Procedia PDF Downloads 55
5136 Quantum Mechanics as A Limiting Case of Relativistic Mechanics

Authors: Ahmad Almajid

Abstract:

The idea of unifying quantum mechanics with general relativity is still a dream for many researchers, as physics has only two paths, no more. Einstein's path, which is mainly based on particle mechanics, and the path of Paul Dirac and others, which is based on wave mechanics, the incompatibility of the two approaches is due to the radical difference in the initial assumptions and the mathematical nature of each approach. Logical thinking in modern physics leads us to two problems: - In quantum mechanics, despite its success, the problem of measurement and the problem of wave function interpretation is still obscure. - In special relativity, despite the success of the equivalence of rest-mass and energy, but at the speed of light, the fact that the energy becomes infinite is contrary to logic because the speed of light is not infinite, and the mass of the particle is not infinite too. These contradictions arise from the overlap of relativistic and quantum mechanics in the neighborhood of the speed of light, and in order to solve these problems, one must understand well how to move from relativistic mechanics to quantum mechanics, or rather, to unify them in a way different from Dirac's method, in order to go along with God or Nature, since, as Einstein said, "God doesn't play dice." From De Broglie's hypothesis about wave-particle duality, Léon Brillouin's definition of the new proper time was deduced, and thus the quantum Lorentz factor was obtained. Finally, using the Euler-Lagrange equation, we come up with new equations in quantum mechanics. In this paper, the two problems in modern physics mentioned above are solved; it can be said that this new approach to quantum mechanics will enable us to unify it with general relativity quite simply. If the experiments prove the validity of the results of this research, we will be able in the future to transport the matter at speed close to the speed of light. Finally, this research yielded three important results: 1- Lorentz quantum factor. 2- Planck energy is a limited case of Einstein energy. 3- Real quantum mechanics, in which new equations for quantum mechanics match and exceed Dirac's equations, these equations have been reached in a completely different way from Dirac's method. These equations show that quantum mechanics is a limited case of relativistic mechanics. At the Solvay Conference in 1927, the debate about quantum mechanics between Bohr, Einstein, and others reached its climax, while Bohr suggested that if particles are not observed, they are in a probabilistic state, then Einstein said his famous claim ("God does not play dice"). Thus, Einstein was right, especially when he didn't accept the principle of indeterminacy in quantum theory, although experiments support quantum mechanics. However, the results of our research indicate that God really does not play dice; when the electron disappears, it turns into amicable particles or an elastic medium, according to the above obvious equations. Likewise, Bohr was right also, when he indicated that there must be a science like quantum mechanics to monitor and study the motion of subatomic particles, but the picture in front of him was blurry and not clear, so he resorted to the probabilistic interpretation.

Keywords: lorentz quantum factor, new, planck’s energy as a limiting case of einstein’s energy, real quantum mechanics, new equations for quantum mechanics

Procedia PDF Downloads 67
5135 Towards a Resources Provisioning for Dynamic Workflows in the Cloud

Authors: Fairouz Fakhfakh, Hatem Hadj Kacem, Ahmed Hadj Kacem

Abstract:

Cloud computing offers a new model of service provisioning for workflow applications, thanks to its elasticity and its paying model. However, it presents various challenges that need to be addressed in order to be efficiently utilized. The resources provisioning problem for workflow applications has been widely studied. Nevertheless, the existing works did not consider the change in workflow instances while they are being executed. This functionality has become a major requirement to deal with unusual situations and evolution. This paper presents a first step towards the resources provisioning for a dynamic workflow. In fact, we propose a provisioning algorithm which minimizes the overall workflow execution cost, while meeting a deadline constraint. Then, we extend it to support the dynamic adding of tasks. Experimental results show that our proposed heuristic demonstrates a significant reduction in resources cost by using a consolidation process.

Keywords: cloud computing, resources provisioning, dynamic workflow, workflow applications

Procedia PDF Downloads 279
5134 Software Quality Assurance in Network Security using Cryptographic Techniques

Authors: Sidra Shabbir, Ayesha Manzoor, Mehreen Sirshar

Abstract:

The use of the network communication has imposed serious threats to the security of assets over the network. Network security is getting more prone to active and passive attacks which may result in serious consequences to data integrity, confidentiality and availability. Various cryptographic techniques have been proposed in the past few years to combat with the concerned problem by ensuring quality but in order to have a fully secured network; a framework of new cryptosystem was needed. This paper discusses certain cryptographic techniques which have shown far better improvement in the network security with enhanced quality assurance. The scope of this research paper is to cover the security pitfalls in the current systems and their possible solutions based on the new cryptosystems. The development of new cryptosystem framework has paved a new way to the widespread network communications with enhanced quality in network security.

Keywords: cryptography, network security, encryption, decryption, integrity, confidentiality, security algorithms, elliptic curve cryptography

Procedia PDF Downloads 726
5133 The Impact of an Interactive E-Book on Mathematics Reading and Spatial Ability in Middle School Students

Authors: Abebayehu Yohannes, Hsiu-Ling Chen, Chiu-Chen Chang

Abstract:

Mathematics reading and spatial ability are important learning components in mathematics education. However, many students struggle to understand real-world problems and lack the spatial ability to form internal imagery. To cope with this problem, in this study, an interactive e-book was developed. The result indicated that both groups had a significant increase in the mathematics reading ability test, and a significant difference was observed in the overall mathematics reading score in favor of the experimental group. In addition, the interactive e-book learning mode had significant impacts on students’ spatial ability. It was also found that the richness of content with visual and interactive elements provided in the interactive e-book enhanced students’ satisfaction with the teaching material.

Keywords: interactive e-books, spatial ability, mathematics reading, satisfaction, three view

Procedia PDF Downloads 175
5132 Trusted Neural Network: Reversibility in Neural Networks for Network Integrity Verification

Authors: Malgorzata Schwab, Ashis Kumer Biswas

Abstract:

In this concept paper, we explore the topic of Reversibility in Neural Networks leveraged for Network Integrity Verification and crafted the term ''Trusted Neural Network'' (TNN), paired with the API abstraction around it, to embrace the idea formally. This newly proposed high-level generalizable TNN model builds upon the Invertible Neural Network architecture, trained simultaneously in both forward and reverse directions. This allows for the original system inputs to be compared with the ones reconstructed from the outputs in the reversed flow to assess the integrity of the end-to-end inference flow. The outcome of that assessment is captured as an Integrity Score. Concrete implementation reflecting the needs of specific problem domains can be derived from this general approach and is demonstrated in the experiments. The model aspires to become a useful practice in drafting high-level systems architectures which incorporate AI capabilities.

Keywords: trusted, neural, invertible, API

Procedia PDF Downloads 138
5131 A Performance Model for Designing Network in Reverse Logistic

Authors: S. Dhib, S. A. Addouche, T. Loukil, A. Elmhamedi

Abstract:

In this paper, a reverse supply chain network is investigated for a decision making. This decision is surrounded by complex flows of returned products, due to the increasing quantity, the type of returned products and the variety of recovery option products (reuse, recycling, and refurbishment). The most important problem in the reverse logistic network (RLN) is to orient returned products to the suitable type of recovery option. However, returned products orientations from collect sources to the recovery disposition have not well considered in performance model. In this study, we propose a performance model for designing a network configuration on reverse logistics. Conceptual and analytical models are developed with taking into account operational, economic and environmental factors on designing network.

Keywords: reverse logistics, network design, performance model, open loop configuration

Procedia PDF Downloads 431
5130 Critical and Strategic Issues in Compensation, Staffing and Personnel Management in Nigeria

Authors: Shonuga Olajumoke Adedoyinsola

Abstract:

Staffing and Compensation are at the core of any employment exchange, and they serve as the defining characteristics of any employment relationship. Most organizations understand the benefits that a longer term approach to staff planning can bring and the answer to this problem lies not in trying to implement the traditional approach more effectively, but in implementing a completely different kind of process for strategic staffing. The study focuses on critical points of compensation, staffing and personnel management. The fundamentals of these programs include the elements of vision, potential, communication and motivation. The aim of the paper is to identify the most important attributes of compensation and incentives, staffing and personnel management. Research method is the analysis and synthesis of scientific literature, logical, comparative and graphic representation. On the basis of analysis, the author presents the models of these systems for positive employee attitudes and behaviors.

Keywords: compensation, employees, incentives, staffing, personnel management

Procedia PDF Downloads 288
5129 Comparison of Different Data Acquisition Techniques for Shape Optimization Problems

Authors: Attila Vámosi, Tamás Mankovits, Dávid Huri, Imre Kocsis, Tamás Szabó

Abstract:

Non-linear FEM calculations are indispensable when important technical information like operating performance of a rubber component is desired. Rubber bumpers built into air-spring structures may undergo large deformations under load, which in itself shows non-linear behavior. The changing contact range between the parts and the incompressibility of the rubber increases this non-linear behavior further. The material characterization of an elastomeric component is also a demanding engineering task. The shape optimization problem of rubber parts led to the study of FEM based calculation processes. This type of problems was posed and investigated by several authors. In this paper the time demand of certain calculation methods are studied and the possibilities of time reduction is presented.

Keywords: rubber bumper, data acquisition, finite element analysis, support vector regression

Procedia PDF Downloads 466
5128 The Solid-Phase Sensor Systems for Fluorescent and SERS-Recognition of Neurotransmitters for Their Visualization and Determination in Biomaterials

Authors: Irina Veselova, Maria Makedonskaya, Olga Eremina, Alexandr Sidorov, Eugene Goodilin, Tatyana Shekhovtsova

Abstract:

Such catecholamines as dopamine, norepinephrine, and epinephrine are the principal neurotransmitters in the sympathetic nervous system. Catecholamines and their metabolites are considered to be important markers of socially significant diseases such as atherosclerosis, diabetes, coronary heart disease, carcinogenesis, Alzheimer's and Parkinson's diseases. Currently, neurotransmitters can be studied via electrochemical and chromatographic techniques that allow their characterizing and quantification, although these techniques can only provide crude spatial information. Besides, the difficulty of catecholamine determination in biological materials is associated with their low normal concentrations (~ 1 nM) in biomaterials, which may become even one more order lower because of some disorders. In addition, in blood they are rapidly oxidized by monoaminooxidases from thrombocytes and, for this reason, the determination of neurotransmitter metabolism indicators in an organism should be very rapid (15—30 min), especially in critical states. Unfortunately, modern instrumental analysis does not offer a complex solution of this problem: despite its high sensitivity and selectivity, HPLC-MS cannot provide sufficiently rapid analysis, while enzymatic biosensors and immunoassays for the determination of the considered analytes lack sufficient sensitivity and reproducibility. Fluorescent and SERS-sensors remain a compelling technology for approaching the general problem of selective neurotransmitter detection. In recent years, a number of catecholamine sensors have been reported including RNA aptamers, fluorescent ribonucleopeptide (RNP) complexes, and boronic acid based synthetic receptors and the sensor operated in a turn-off mode. In this work we present the fluorescent and SERS turn-on sensor systems based on the bio- or chemorecognizing nanostructured films {chitosan/collagen-Tb/Eu/Cu-nanoparticles-indicator reagents} that provide the selective recognition, visualization, and sensing of the above mentioned catecholamines on the level of nanomolar concentrations in biomaterials (cell cultures, tissue etc.). We have (1) developed optically transparent porous films and gels of chitosan/collagen; (2) ensured functionalization of the surface by molecules-'recognizers' (by impregnation and immobilization of components of the indicator systems: biorecognizing and auxiliary reagents); (3) performed computer simulation for theoretical prediction and interpretation of some properties of the developed materials and obtained analytical signals in biomaterials. We are grateful for the financial support of this research from Russian Foundation for Basic Research (grants no. 15-03-05064 a, and 15-29-01330 ofi_m).

Keywords: biomaterials, fluorescent and SERS-recognition, neurotransmitters, solid-phase turn-on sensor system

Procedia PDF Downloads 398
5127 Physical and Thermo-Physical Properties of High Strength Concrete Containing Raw Rice Husk after High Temperature Effect

Authors: B. Akturk, N. Yuzer, N. Kabay

Abstract:

High temperature is one of the most detrimental effects that cause important changes in concrete’s mechanical, physical, and thermo-physical properties. As a result of these changes, especially high strength concrete (HSC), may exhibit damages such as cracks and spallings. To overcome this problem, incorporating polymer fibers such as polypropylene (PP) in concrete is a very well-known method. In this study, using RRH as a sustainable material instead of PP fiber in HSC to prevent spallings and improve physical and thermo-physical properties were investigated. Therefore, seven HSC mixtures with 0.25 water to binder ratio were prepared, incorporating silica fume and blast furnace slag. PP and RRH were used at 0.2-0.5% and 0.5-3% by weight of cement, respectively. All specimens were subjected to high temperatures (20 (control), 300, 600 and 900˚C) with a heating rate of 2.5˚C/min and after cooling, residual physical and thermo-physical properties were determined.

Keywords: high temperature, high strength concrete, polypropylene fiber, raw rice husk, thermo-physical properties

Procedia PDF Downloads 261
5126 Preoperative versus Postoperative Radiation Therapy in Patients with Soft Tissue Sarcoma of the Extremity

Authors: AliAkbar Hafezi, Jalal Taherian, Jamshid Abedi, Mahsa Elahi, Behnam Kadkhodaei

Abstract:

Background: Soft tissue sarcomas (STS) are generally treated with a combination of limb preservation surgery and radiation therapy. Today, preoperative radiation therapy is considered for accurate treatment volume and smaller field size. Therefore, this study was performed to compare preoperative with postoperative radiation therapy in patients with extremity STS. Methods: In this non-randomized clinical trial, patients with localized extremity STS referred to the orthopedic clinics in Iran from 2021 to 2023 were studied. Patients were randomly divided into two groups: preoperative and postoperative radiation therapy. The two groups of patients were compared in terms of acute (wound dehiscence and infection) and late (limb edema, subcutaneous fibrosis, and joint stiffness) complications and their severity, as well as local recurrence and other one-year outcomes. Results: A total of 80 patients with localized extremity STS were evaluated in two treatment groups. The groups were matched in terms of age, sex, history of diabetes mellitus, hypertension, smoking, involved side, involved extremity, lesion location, and tumor histopathology. The acute complications of treatment in the two groups of patients did not differ significantly (P > 0.05). Of the late complications, only joint stiffness between the two groups had significant statistical differences (P < 0.001). The severity of all three late complications in the postoperative radiation therapy group was significantly higher (P < 0.05). There was no significant difference between the two groups in terms of the rate of local recurrence of other one-year outcomes (P > 0.05). Conclusion: This study showed that in patients with localized extremity STS, the two therapeutic approaches of adjuvant and neoadjuvant radiation therapy did not differ significantly in terms of local recurrence and distant metastasis during the one-year follow-up period and due to fewer late complications in preoperative radiotherapy group, this treatment approach can be a better choice than postoperative radiation therapy.

Keywords: soft tissue sarcoma, extremity, preoperative radiation therapy, postoperative radiation therapy

Procedia PDF Downloads 35
5125 Design of Jumping Structure of Spherical Robot Based on Archimedes' Helix

Authors: Zhang Zijian

Abstract:

Nowadays, spherical robots have played an important role in many fields, but the insufficient ability of obstacle surmounting limits their wider application fields. To solve this problem, a jumping system of a spherical robot is designed based on Archimedes helix. The jumping system of the robot utilizes the characteristics of Archimedes helix and isovelocity helix to achieve constant speed and stable contraction, which ensures the stability of the system. Also, the jumping action of the robot is realized by instantaneous release of elastic potential energy. In order to verify the effectiveness of the jumping system, we designed a spherical robot and its jumping system. The experimental results show that the jumping system has the advantages of light weight, small size, high energy conversion efficiency, and can realize the spherical jumping function.

Keywords: hopping mechanism, Archimedes' Helix, hopping robot, spherical robot

Procedia PDF Downloads 126
5124 Parametric Analysis of Solid Oxide Fuel Cell Using Lattice Boltzmann Method

Authors: Abir Yahya, Hacen Dhahri, Khalifa Slimi

Abstract:

The present paper deals with a numerical simulation of temperature field inside a solid oxide fuel cell (SOFC) components. The temperature distribution is investigated using a co-flow planar SOFC comprising the air and fuel channel and two-ceramic electrodes, anode and cathode, separated by a dense ceramic electrolyte. The Lattice Boltzmann method (LBM) is used for the numerical simulation of the physical problem. The effects of inlet temperature, anode thermal conductivity and current density on temperature distribution are discussed. It was found that temperature distribution is very sensitive to the inlet temperature and the current density.

Keywords: heat sources, Lattice Boltzmann method, solid oxide fuel cell, temperature

Procedia PDF Downloads 301
5123 Comparison of Water Equivalent Ratio of Several Dosimetric Materials in Proton Therapy Using Monte Carlo Simulations and Experimental Data

Authors: M. R. Akbari , H. Yousefnia, E. Mirrezaei

Abstract:

Range uncertainties of protons are currently a topic of interest in proton therapy. Two of the parameters that are often used to specify proton range are water equivalent thickness (WET) and water equivalent ratio (WER). Since WER values for a specific material is nearly constant at different proton energies, it is a more useful parameter to compare. In this study, WER values were calculated for different proton energies in polymethyl methacrylate (PMMA), polystyrene (PS) and aluminum (Al) using FLUKA and TRIM codes. The results were compared with analytical, experimental and simulated SEICS code data obtained from the literature. In FLUKA simulation, a cylindrical phantom, 1000 mm in height and 300 mm in diameter, filled with the studied materials was simulated. A typical mono-energetic proton pencil beam in a wide range of incident energies usually applied in proton therapy (50 MeV to 225 MeV) impinges normally on the phantom. In order to obtain the WER values for the considered materials, cylindrical detectors, 1 mm in height and 20 mm in diameter, were also simulated along the beam trajectory in the phantom. In TRIM calculations, type of projectile, energy and angle of incidence, type of target material and thickness should be defined. The mode of 'detailed calculation with full damage cascades' was selected for proton transport in the target material. The biggest difference in WER values between the codes was 3.19%, 1.9% and 0.67% for Al, PMMA and PS, respectively. In Al and PMMA, the biggest difference between each code and experimental data was 1.08%, 1.26%, 2.55%, 0.94%, 0.77% and 0.95% for SEICS, FLUKA and SRIM, respectively. FLUKA and SEICS had the greatest agreement (≤0.77% difference in PMMA and ≤1.08% difference in Al, respectively) with the available experimental data in this study. It is concluded that, FLUKA and TRIM codes have capability for Bragg curves simulation and WER values calculation in the studied materials. They can also predict Bragg peak location and range of proton beams with acceptable accuracy.

Keywords: water equivalent ratio, dosimetric materials, proton therapy, Monte Carlo simulations

Procedia PDF Downloads 309
5122 Hardware Implementation on Field Programmable Gate Array of Two-Stage Algorithm for Rough Set Reduct Generation

Authors: Tomasz Grzes, Maciej Kopczynski, Jaroslaw Stepaniuk

Abstract:

The rough sets theory developed by Prof. Z. Pawlak is one of the tools that can be used in the intelligent systems for data analysis and processing. Banking, medicine, image recognition and security are among the possible fields of utilization. In all these fields, the amount of the collected data is increasing quickly, but with the increase of the data, the computation speed becomes the critical factor. Data reduction is one of the solutions to this problem. Removing the redundancy in the rough sets can be achieved with the reduct. A lot of algorithms of generating the reduct were developed, but most of them are only software implementations, therefore have many limitations. Microprocessor uses the fixed word length, consumes a lot of time for either fetching as well as processing of the instruction and data; consequently, the software based implementations are relatively slow. Hardware systems don’t have these limitations and can process the data faster than a software. Reduct is the subset of the decision attributes that provides the discernibility of the objects. For the given decision table there can be more than one reduct. Core is the set of all indispensable condition attributes. None of its elements can be removed without affecting the classification power of all condition attributes. Moreover, every reduct consists of all the attributes from the core. In this paper, the hardware implementation of the two-stage greedy algorithm to find the one reduct is presented. The decision table is used as an input. Output of the algorithm is the superreduct which is the reduct with some additional removable attributes. First stage of the algorithm is calculating the core using the discernibility matrix. Second stage is generating the superreduct by enriching the core with the most common attributes, i.e., attributes that are more frequent in the decision table. Described above algorithm has two disadvantages: i) generating the superreduct instead of reduct, ii) additional first stage may be unnecessary if the core is empty. But for the systems focused on the fast computation of the reduct the first disadvantage is not the key problem. The core calculation can be achieved with a combinational logic block, and thus add respectively little time to the whole process. Algorithm presented in this paper was implemented in Field Programmable Gate Array (FPGA) as a digital device consisting of blocks that process the data in a single step. Calculating the core is done by the comparators connected to the block called 'singleton detector', which detects if the input word contains only single 'one'. Calculating the number of occurrences of the attribute is performed in the combinational block made up of the cascade of the adders. The superreduct generation process is iterative and thus needs the sequential circuit for controlling the calculations. For the research purpose, the algorithm was also implemented in C language and run on a PC. The times of execution of the reduct calculation in a hardware and software were considered. Results show increase in the speed of data processing.

Keywords: data reduction, digital systems design, field programmable gate array (FPGA), reduct, rough set

Procedia PDF Downloads 212