Search results for: entertainment computing
594 Modeling and Simulation of the Tripod Gait of a Hexapod Robot
Authors: El Hansali Hasnaa, Bennani Mohammed
Abstract:
Hexapod legged robot’s missions, particularly in irregular and dangerous areas, require high stability and high precision. In this paper, we consider the rectangular architecture body of legged robots with six legs distributed symmetrically along two sides, each leg contains three degrees of freedom for greater mobility. The aim of this work is planning tripod gait trajectory, based on the computing of the kinematic model to determine the joint variables in the lifting and the propelling phases. For this, appropriate coordinate frames are attached to the body and legs in order to obtain clear representation and efficient generation of the system equations. A simulation in MATLAB software platform is developed to confirm the kinematic model and various trajectories to the tripod gait adopted by the hexapod robot in its locomotion.Keywords: hexapod legged robot, inverse kinematic model, simulation in MATLAB, tripod gait
Procedia PDF Downloads 277593 Monte Carlo Simulation of Magnetic Properties in Bit Patterned Media
Authors: O. D. Arbeláez-Echeverri, E. Restrepo-Parra, J. C. Riano-Rojas
Abstract:
A two dimensional geometric model of Bit Patterned Media is proposed, the model is based on the crystal structure of the materials commonly used to produce the nano islands in bit patterned materials and the possible defects that may arise from the interaction between the nano islands and the matrix material. The dynamic magnetic properties of the material are then computed using time aware integration methods for the multi spin Hamiltonian. The Hamiltonian takes into account both the spatial and topological disorder of the sample as well as the high perpendicular anisotropy that is pursued when building bit patterned media. The main finding of the research was the possibility of replicating the results of previous experiments on similar materials and the ability of computing the switching field distribution given the geometry of the material and the parameters required by the model.Keywords: nanostructures, Monte Carlo, pattern media, magnetic properties
Procedia PDF Downloads 503592 Computing Some Topological Descriptors of Single-Walled Carbon Nanotubes
Authors: Amir Bahrami
Abstract:
In the fields of chemical graph theory, molecular topology, and mathematical chemistry, a topological index or a descriptor index also known as a connectivity index is a type of a molecular descriptor that is calculated based on the molecular graph of a chemical compound. Topological indices are numerical parameters of a graph which characterize its topology and are usually graph invariant. Topological indices are used for example in the development of quantitative structure-activity relationships (QSARs) in which the biological activity or other properties of molecules are correlated with their chemical structure. In this paper some descriptor index (descriptor index) of single-walled carbon nanotubes, is determined.Keywords: chemical graph theory, molecular topology, molecular descriptor, single-walled carbon nanotubes
Procedia PDF Downloads 338591 Autonomic Recovery Plan with Server Virtualization
Authors: S. Hameed, S. Anwer, M. Saad, M. Saady
Abstract:
For autonomic recovery with server virtualization, a cogent plan that includes recovery techniques and backups with virtualized servers can be developed instead of assigning an idle server to backup operations. In addition to hardware cost reduction and data center trail, the disaster recovery plan can ensure system uptime and to meet objectives of high availability, recovery time, recovery point, server provisioning, and quality of services. This autonomic solution would also support disaster management, testing, and development of the recovery site. In this research, a workflow plan is proposed for supporting disaster recovery with virtualization providing virtual monitoring, requirements engineering, solution decision making, quality testing, and disaster management. This recovery model would make disaster recovery a lot easier, faster, and less error prone.Keywords: autonomous intelligence, disaster recovery, cloud computing, server virtualization
Procedia PDF Downloads 162590 Effect of Ionized Plasma Medium on the Radiation of a Rectangular Microstrip Antenna on Ferrite Substrate
Authors: Ayman Al Sawalha
Abstract:
This paper presents theoretical investigations on the radiation of rectangular microstrip antenna printed on a magnetized ferrite substrate Ni0.62Co0.02Fe1.948O4 in the presence of ionized plasma medium. The theoretical study of rectangular microstrip antenna in free space is carried out by applying the transmission line model combining with potential function techniques while hydrodynamic theory is used for it is analysis in plasma medium. By taking the biased and unbiased ferrite cases, far-field radiation patterns in free space and plasma medium are obtained which in turn are applied in computing radiated power, directivity, quality factor and bandwidth of antenna. It is found that the presence of plasma medium affects the performance of rectangular microstrip antenna structure significantly.Keywords: ferrite, microstrip antenna, plasma, radiation
Procedia PDF Downloads 323589 Fast and Non-Invasive Patient-Specific Optimization of Left Ventricle Assist Device Implantation
Authors: Huidan Yu, Anurag Deb, Rou Chen, I-Wen Wang
Abstract:
The use of left ventricle assist devices (LVADs) in patients with heart failure has been a proven and effective therapy for patients with severe end-stage heart failure. Due to the limited availability of suitable donor hearts, LVADs will probably become the alternative solution for patient with heart failure in the near future. While the LVAD is being continuously improved toward enhanced performance, increased device durability, reduced size, a better understanding of implantation management becomes critical in order to achieve better long-term blood supplies and less post-surgical complications such as thrombi generation. Important issues related to the LVAD implantation include the location of outflow grafting (OG), the angle of the OG, the combination between LVAD and native heart pumping, uniform or pulsatile flow at OG, etc. We have hypothesized that an optimal implantation of LVAD is patient specific. To test this hypothesis, we employ a novel in-house computational modeling technique, named InVascular, to conduct a systematic evaluation of cardiac output at aortic arch together with other pertinent hemodynamic quantities for each patient under various implantation scenarios aiming to get an optimal implantation strategy. InVacular is a powerful computational modeling technique that integrates unified mesoscale modeling for both image segmentation and fluid dynamics with the cutting-edge GPU parallel computing. It first segments the aortic artery from patient’s CT image, then seamlessly feeds extracted morphology, together with the velocity wave from Echo Ultrasound image of the same patient, to the computation model to quantify 4-D (time+space) velocity and pressure fields. Using one NVIDIA Tesla K40 GPU card, InVascular completes a computation from CT image to 4-D hemodynamics within 30 minutes. Thus it has the great potential to conduct massive numerical simulation and analysis. The systematic evaluation for one patient includes three OG anastomosis (ascending aorta, descending thoracic aorta, and subclavian artery), three combinations of LVAD and native heart pumping (1:1, 1:2, and 1:3), three angles of OG anastomosis (inclined upward, perpendicular, and inclined downward), and two LVAD inflow conditions (uniform and pulsatile). The optimal LVAD implantation is suggested through a comprehensive analysis of the cardiac output and related hemodynamics from the simulations over the fifty-four scenarios. To confirm the hypothesis, 5 random patient cases will be evaluated.Keywords: graphic processing unit (GPU) parallel computing, left ventricle assist device (LVAD), lumped-parameter model, patient-specific computational hemodynamics
Procedia PDF Downloads 133588 Diagnosis of Diabetes Using Computer Methods: Soft Computing Methods for Diabetes Detection Using Iris
Authors: Piyush Samant, Ravinder Agarwal
Abstract:
Complementary and Alternative Medicine (CAM) techniques are quite popular and effective for chronic diseases. Iridology is more than 150 years old CAM technique which analyzes the patterns, tissue weakness, color, shape, structure, etc. for disease diagnosis. The objective of this paper is to validate the use of iridology for the diagnosis of the diabetes. The suggested model was applied in a systemic disease with ocular effects. 200 subject data of 100 each diabetic and non-diabetic were evaluated. Complete procedure was kept very simple and free from the involvement of any iridologist. From the normalized iris, the region of interest was cropped. All 63 features were extracted using statistical, texture analysis, and two-dimensional discrete wavelet transformation. A comparison of accuracies of six different classifiers has been presented. The result shows 89.66% accuracy by the random forest classifier.Keywords: complementary and alternative medicine, classification, iridology, iris, feature extraction, disease prediction
Procedia PDF Downloads 407587 Portrait of Musical Creativity or Indolence: A Critique of Unoka Character in Achebe’s Things Fall Apart
Authors: Ebim Matthew Abua
Abstract:
In Chinua Achebe’s Things Fall Apart (henceforth, TFA), the character Unoka is portrayed as a weakling and indolent person even when he was a creative artist, a talented musician, and a mathematician. His lack of achievement becomes the barometer for measuring his success. Right from time, music is considered to be of great significance because of its capacity to recreate and retell social events. To this end, music is both a social act and a creative art. As a social act, music is a discursive medium that exploits the dynamics of art in its evaluation of society. Music is so much a part of human existence that its presence in literature can help readers relate to fictional situations and characters. In this paper, the character Unoka is examined against the backdrop of his musical proclivities and his contributions to the overall development of TFA. Unfortunately, Achebe’s Things Fall Apart, a product of artistic creativity, portrays the personality of Unoka as lazy and uninspiring because he (Unoka) is a musician who is busy playing his flute and hardly doing anything productive. This paper is significant because it examines the literary and or linguistic depiction of Unoka and the aftermath of that depiction on the entire novel and, by extension, the larger society. Methodologically, this paper adopted the qualitative approach from the ethnography of communication (EOC), which is the analysis of communication within the wider context of the social and cultural practices and beliefs of the members of a particular culture or speech community. The aim of this qualitative research method includes the ability to discern which communication acts and/or codes are important to different groups, what types of meanings groups apply to different communication events, and how group members learn these codes to provide insight into particular communities. The study reveals that the people of Umuofia were mono-directional in their economy, and there was no room for diversification. One was either a farmer, or such a person was relegated to the background. Unoka, taking up a new challenge of diversifying the economy from the perspective of entertainment, was viewed as a misnomer. This study opens the door to other areas of studying Achebe’s epic novel apart from the critical works of literary artists that have been dished out over time.Keywords: literature, popular culture, unoka, things fall apart
Procedia PDF Downloads 59586 Fat-Tail Test of Regulatory DNA Sequences
Authors: Jian-Jun Shu
Abstract:
The statistical properties of CRMs are explored by estimating similar-word set occurrence distribution. It is observed that CRMs tend to have a fat-tail distribution for similar-word set occurrence. Thus, the fat-tail test with two fatness coefficients is proposed to distinguish CRMs from non-CRMs, especially from exons. For the first fatness coefficient, the separation accuracy between CRMs and exons is increased as compared with the existing content-based CRM prediction method – fluffy-tail test. For the second fatness coefficient, the computing time is reduced as compared with fluffy-tail test, making it very suitable for long sequences and large data-base analysis in the post-genome time. Moreover, these indexes may be used to predict the CRMs which have not yet been observed experimentally. This can serve as a valuable filtering process for experiment.Keywords: statistical approach, transcription factor binding sites, cis-regulatory modules, DNA sequences
Procedia PDF Downloads 290585 Exploration of Various Metrics for Partitioning of Cellular Automata Units for Efficient Reconfiguration of Field Programmable Gate Arrays (FPGAs)
Authors: Peter Tabatt, Christian Siemers
Abstract:
Using FPGA devices to improve the behavior of time-critical parts of embedded systems is a proven concept for years. With reconfigurable FPGA devices, the logical blocks can be partitioned and grouped into static and dynamic parts. The dynamic parts can be reloaded 'on demand' at runtime. This work uses cellular automata, which are constructed through compilation from (partially restricted) ANSI-C sources, to determine the suitability of various metrics for optimal partitioning. Significant metrics, in this case, are for example the area on the FPGA device for the partition, the pass count for loop constructs and communication characteristics to other partitions. With successful partitioning, it is possible to use smaller FPGA devices for the same requirements as with not reconfigurable FPGA devices or – vice versa – to use the same FPGAs for larger programs.Keywords: reconfigurable FPGA, cellular automata, partitioning, metrics, parallel computing
Procedia PDF Downloads 271584 Investigating the Form of the Generalised Equations of Motion of the N-Bob Pendulum and Computing Their Solution Using MATLAB
Authors: Divij Gupta
Abstract:
Pendular systems have a range of both mathematical and engineering applications, ranging from modelling the behaviour of a continuous mass-density rope to utilisation as Tuned Mass Dampers (TMD). Thus, it is of interest to study the differential equations governing the motion of such systems. Here we attempt to generalise these equations of motion for the plane compound pendulum with a finite number of N point masses. A Lagrangian approach is taken, and we attempt to find the generalised form for the Euler-Lagrange equations of motion for the i-th bob of the N -bob pendulum. The co-ordinates are parameterized as angular quantities to reduce the number of degrees of freedom from 2N to N to simplify the form of the equations. We analyse the form of these equations up to N = 4 to determine the general form of the equation. We also develop a MATLAB program to compute a solution to the system for a given input value of N and a given set of initial conditions.Keywords: classical mechanics, differential equation, lagrangian analysis, pendulum
Procedia PDF Downloads 208583 Low-Cost Fog Edge Computing for Smart Power Management and Home Automation
Authors: Belkacem Benadda, Adil Benabdellah, Boutheyna Souna
Abstract:
The Internet of Things (IoT) is an unprecedented creation. Electronics objects are now able to interact, share, respond and adapt to their environment on a much larger basis. Actual spread of these modern means of connectivity and solutions with high data volume exchange are affecting our ways of life. Accommodation is becoming an intelligent living space, not only suited to the people circumstances and desires, but also to systems constraints to make daily life simpler, cheaper, increase possibilities and achieve a higher level of services and luxury. In this paper we are as Internet access, teleworking, consumption monitoring, information search, etc.). This paper addresses the design and integration of a smart home, it also purposes an IoT solution that allows smart power consumption based on measurements from power-grid and deep learning analysis.Keywords: array sensors, IoT, power grid, FPGA, embedded
Procedia PDF Downloads 116582 DAG Design and Tradeoff for Full Live Virtual Machine Migration over XIA Network
Authors: Dalu Zhang, Xiang Jin, Dejiang Zhou, Jianpeng Wang, Haiying Jiang
Abstract:
Traditional TCP/IP network is showing lots of shortages and research for future networks is becoming a hotspot. FIA (Future Internet Architecture) and FIA-NP (Next Phase) are supported by US NSF for future Internet designing. Moreover, virtual machine migration is a significant technique in cloud computing. As a network application, it should also be supported in XIA (expressive Internet Architecture), which is in both FIA and FIA-NP projects. This paper is an experimental study aims at verifying the feasibility of VM migration over XIA. We present three ways to maintain VM connectivity and communication states concerning DAG design and routing table modification. VM migration experiments are conducted intra-AD and inter-AD with KVM instances. The procedure is achieved by a migration control protocol which is suitable for the characters of XIA. Evaluation results show that our solutions can well supports full live VM migration over XIA network respectively, keeping services seamless.Keywords: DAG, downtime, virtual machine migration, XIA
Procedia PDF Downloads 855581 SciPaaS: a Scientific Execution Platform for the Cloud
Authors: Wesley H. Brewer, John C. Sanford
Abstract:
SciPaaS is a prototype development of an execution platform/middleware designed to make it easy for scientists to rapidly deploy their scientific applications (apps) to the cloud. It provides all the necessary infrastructure for running typical IXP (Input-eXecute-Plot) style apps, including: a web interface, post-processing and plotting capabilities, job scheduling, real-time monitoring of running jobs, and even a file/case manager. In this paper, first the system architecture is described and then is demonstrated for a two scientific applications: (1) a simple finite-difference solver of the inviscid Burger’s equation, and (2) Mendel’s Accountant—a forward-time population genetics simulation model. The implications of the prototype are discussed in terms of ease-of-use and deployment options, especially in cloud environments.Keywords: web-based simulation, cloud computing, Platform-as-a-Service (PaaS), rapid application development (RAD), population genetics
Procedia PDF Downloads 590580 The Impact of Artificial Intelligence on Journalism and Mass Communication
Authors: Saad Zagloul Shokri Melika
Abstract:
The London College of Communication is one of the only universities in the world to offer a lifestyle journalism master’s degree. A hybrid originally constructed largely out of a generic journalism program crossed with numerous cultural studies approaches, the degree has developed into a leading lifestyle journalism education attracting students worldwide. This research project seeks to present a framework for structuring the degree as well as to understand how students in this emerging field of study value the program. While some researchers have addressed questions about journalism and higher education, none have looked specifically at the increasingly important genre of lifestyle journalism, which Folker Hanusch defines as including notions of consumerism and critique among other identifying traits. Lifestyle journalism, itself poorly researched by scholars, can relate to topics including travel, fitness, and entertainment, and as such, arguably a lifestyle journalism degree should prepare students to engage with these topics. This research uses the existing Masters of Arts and Lifestyle Journalism at the London College of Communications as a case study to examine the school’s approach. Furthering Hanusch’s original definition, this master’s program attempts to characterizes lifestyle journalism by a specific voice or approach, as reflected in the diversity of student’s final projects. This framework echoes the ethos and ideas of the university, which focuses on creativity, design, and experimentation. By analyzing the current degree as well as student feedback, this research aims to assist future educators in pursuing the often neglected field of lifestyle journalism. Through a discovery of the unique mix of practical coursework, theoretical lessons, and broad scope of student work presented in this degree program, researchers strive to develop a framework for lifestyle journalism education, referring to Mark Deuze’s ten questions for journalism education development. While Hanusch began the discussion to legitimize the study of lifestyle journalism, this project strives to go one step further and open up a discussion about teaching of lifestyle journalism at the university level.Keywords: Journalism, accountability, education, television, publicdearth, investigative, journalism, Nigeria, journalismeducation, lifestyle, university
Procedia PDF Downloads 44579 Performance Analysis and Optimization for Diagonal Sparse Matrix-Vector Multiplication on Machine Learning Unit
Authors: Qiuyu Dai, Haochong Zhang, Xiangrong Liu
Abstract:
Diagonal sparse matrix-vector multiplication is a well-studied topic in the fields of scientific computing and big data processing. However, when diagonal sparse matrices are stored in DIA format, there can be a significant number of padded zero elements and scattered points, which can lead to a degradation in the performance of the current DIA kernel. This can also lead to excessive consumption of computational and memory resources. In order to address these issues, the authors propose the DIA-Adaptive scheme and its kernel, which leverages the parallel instruction sets on MLU. The researchers analyze the effect of allocating a varying number of threads, clusters, and hardware architectures on the performance of SpMV using different formats. The experimental results indicate that the proposed DIA-Adaptive scheme performs well and offers excellent parallelism.Keywords: adaptive method, DIA, diagonal sparse matrices, MLU, sparse matrix-vector multiplication
Procedia PDF Downloads 134578 A Conceptual Framework of Digital Twin for Homecare
Authors: Raja Omman Zafar, Yves Rybarczyk, Johan Borg
Abstract:
This article proposes a conceptual framework for the application of digital twin technology in home care. The main goal is to bridge the gap between advanced digital twin concepts and their practical implementation in home care. This study uses a literature review and thematic analysis approach to synthesize existing knowledge and proposes a structured framework suitable for homecare applications. The proposed framework integrates key components such as IoT sensors, data-driven models, cloud computing, and user interface design, highlighting the importance of personalized and predictive homecare solutions. This framework can significantly improve the efficiency, accuracy, and reliability of homecare services. It paves the way for the implementation of digital twins in home care, promoting real-time monitoring, early intervention, and better outcomes.Keywords: digital twin, homecare, older adults, healthcare, IoT, artificial intelligence
Procedia PDF Downloads 71577 Separating Landform from Noise in High-Resolution Digital Elevation Models through Scale-Adaptive Window-Based Regression
Authors: Anne M. Denton, Rahul Gomes, David W. Franzen
Abstract:
High-resolution elevation data are becoming increasingly available, but typical approaches for computing topographic features, like slope and curvature, still assume small sliding windows, for example, of size 3x3. That means that the digital elevation model (DEM) has to be resampled to the scale of the landform features that are of interest. Any higher resolution is lost in this resampling. When the topographic features are computed through regression that is performed at the resolution of the original data, the accuracy can be much higher, and the reported result can be adjusted to the length scale that is relevant locally. Slope and variance are calculated for overlapping windows, meaning that one regression result is computed per raster point. The number of window centers per area is the same for the output as for the original DEM. Slope and variance are computed by performing regression on the points in the surrounding window. Such an approach is computationally feasible because of the additive nature of regression parameters and variance. Any doubling of window size in each direction only takes a single pass over the data, corresponding to a logarithmic scaling of the resulting algorithm as a function of the window size. Slope and variance are stored for each aggregation step, allowing the reported slope to be selected to minimize variance. The approach thereby adjusts the effective window size to the landform features that are characteristic to the area within the DEM. Starting with a window size of 2x2, each iteration aggregates 2x2 non-overlapping windows from the previous iteration. Regression results are stored for each iteration, and the slope at minimal variance is reported in the final result. As such, the reported slope is adjusted to the length scale that is characteristic of the landform locally. The length scale itself and the variance at that length scale are also visualized to aid in interpreting the results for slope. The relevant length scale is taken to be half of the window size of the window over which the minimum variance was achieved. The resulting process was evaluated for 1-meter DEM data and for artificial data that was constructed to have defined length scales and added noise. A comparison with ESRI ArcMap was performed and showed the potential of the proposed algorithm. The resolution of the resulting output is much higher and the slope and aspect much less affected by noise. Additionally, the algorithm adjusts to the scale of interest within the region of the image. These benefits are gained without additional computational cost in comparison with resampling the DEM and computing the slope over 3x3 images in ESRI ArcMap for each resolution. In summary, the proposed approach extracts slope and aspect of DEMs at the lengths scales that are characteristic locally. The result is of higher resolution and less affected by noise than existing techniques.Keywords: high resolution digital elevation models, multi-scale analysis, slope calculation, window-based regression
Procedia PDF Downloads 129576 Managers’ Mobile Information Behavior in an Openness Paradigm Era
Authors: Abd Latif Abdul Rahman, Zuraidah Arif, Muhammad Faizal Iylia, Mohd Ghazali, Asmadi Mohammed Ghazali
Abstract:
Mobile information is a significant access point for human information activities. Theories and models of human information behavior have developed over several decades but have not yet considered the role of the user’s computing device in digital information interactions. This paper reviews the literature that leads to developing a conceptual framework of a study on the managers mobile information behavior. Based on the literature review, dimensions of mobile information behavior are identified, namely, dimension information needs, dimension information access, information retrieval and dimension of information use. The study is significant to understand the nature of librarians’ behavior in searching, retrieving and using information via the mobile device. Secondly, the study would provide suggestions about various kinds of mobile applications which organization can provide for their staff to improve their services.Keywords: mobile information behavior, information behavior, mobile information, mobile devices
Procedia PDF Downloads 349575 Evaluating Service Trustworthiness for Service Selection in Cloud Environment
Authors: Maryam Amiri, Leyli Mohammad-Khanli
Abstract:
Cloud computing is becoming increasingly popular and more business applications are moving to cloud. In this regard, services that provide similar functional properties are increasing. So, the ability to select a service with the best non-functional properties, corresponding to the user preference, is necessary for the user. This paper presents an Evaluation Framework of Service Trustworthiness (EFST) that evaluates the trustworthiness of equivalent services without need to additional invocations of them. EFST extracts user preference automatically. Then, it assesses trustworthiness of services in two dimensions of qualitative and quantitative metrics based on the experiences of past usage of services. Finally, EFST determines the overall trustworthiness of services using Fuzzy Inference System (FIS). The results of experiments and simulations show that EFST is able to predict the missing values of Quality of Service (QoS) better than other competing approaches. Also, it propels users to select the most appropriate services.Keywords: user preference, cloud service, trustworthiness, QoS metrics, prediction
Procedia PDF Downloads 287574 Multi-Scaled Non-Local Means Filter for Medical Images Denoising: Empirical Mode Decomposition vs. Wavelet Transform
Authors: Hana Rabbouch
Abstract:
In recent years, there has been considerable growth of denoising techniques mainly devoted to medical imaging. This important evolution is not only due to the progress of computing techniques, but also to the emergence of multi-resolution analysis (MRA) on both mathematical and algorithmic bases. In this paper, a comparative study is conducted between the two best-known MRA-based decomposition techniques: the Empirical Mode Decomposition (EMD) and the Discrete Wavelet Transform (DWT). The comparison is carried out in a framework of multi-scale denoising, where a Non-Local Means (NLM) filter is performed scale-by-scale to a sample of benchmark medical images. The results prove the effectiveness of the multiscaled denoising, especially when the NLM filtering is coupled with the EMD.Keywords: medical imaging, non local means, denoising, multiscaled analysis, empirical mode decomposition, wavelets
Procedia PDF Downloads 141573 Cryptography and Cryptosystem a Panacea to Security Risk in Wireless Networking
Authors: Modesta E. Ezema, Chikwendu V. Alabekee, Victoria N. Ishiwu, Ifeyinwa NwosuArize, Chinedu I. Nwoye
Abstract:
The advent of wireless networking in computing technology cannot be overemphasized, it opened up easy accessibility to information resources, networking made easier and brought internet accessibility to our doorsteps, but despite all these, some mishap came in with it that is causing mayhem in today ‘s overall information security. The cyber criminals will always compromise the integrity of a message that is not encrypted or that is encrypted with a weak algorithm.In other to correct the mayhem, this study focuses on cryptosystem and cryptography. This ensures end to end crypt messaging. The study of various cryptographic algorithms, as well as the techniques and applications of the cryptography for efficiency, were all considered in the work., present and future applications of cryptography were dealt with as well as Quantum Cryptography was exposed as the current and the future area in the development of cryptography. An empirical study was conducted to collect data from network users.Keywords: algorithm, cryptography, cryptosystem, network
Procedia PDF Downloads 349572 Short Life Cycle Time Series Forecasting
Authors: Shalaka Kadam, Dinesh Apte, Sagar Mainkar
Abstract:
The life cycle of products is becoming shorter and shorter due to increased competition in market, shorter product development time and increased product diversity. Short life cycles are normal in retail industry, style business, entertainment media, and telecom and semiconductor industry. The subject of accurate forecasting for demand of short lifecycle products is of special enthusiasm for many researchers and organizations. Due to short life cycle of products the amount of historical data that is available for forecasting is very minimal or even absent when new or modified products are launched in market. The companies dealing with such products want to increase the accuracy in demand forecasting so that they can utilize the full potential of the market at the same time do not oversupply. This provides the challenge to develop a forecasting model that can forecast accurately while handling large variations in data and consider the complex relationships between various parameters of data. Many statistical models have been proposed in literature for forecasting time series data. Traditional time series forecasting models do not work well for short life cycles due to lack of historical data. Also artificial neural networks (ANN) models are very time consuming to perform forecasting. We have studied the existing models that are used for forecasting and their limitations. This work proposes an effective and powerful forecasting approach for short life cycle time series forecasting. We have proposed an approach which takes into consideration different scenarios related to data availability for short lifecycle products. We then suggest a methodology which combines statistical analysis with structured judgement. Also the defined approach can be applied across domains. We then describe the method of creating a profile from analogous products. This profile can then be used for forecasting products with historical data of analogous products. We have designed an application which combines data, analytics and domain knowledge using point-and-click technology. The forecasting results generated are compared using MAPE, MSE and RMSE error scores. Conclusion: Based on the results it is observed that no one approach is sufficient for short life-cycle forecasting and we need to combine two or more approaches for achieving the desired accuracy.Keywords: forecast, short life cycle product, structured judgement, time series
Procedia PDF Downloads 358571 Multi-Level Priority Based Task Scheduling Algorithm for Workflows in Cloud Environment
Authors: Anju Bala, Inderveer Chana
Abstract:
Task scheduling is the key concern for the execution of performance-driven workflow applications. As efficient scheduling can have major impact on the performance of the system, task scheduling is often chosen for assigning the request to resources in an efficient way based on cloud resource characteristics. In this paper, priority based task scheduling algorithm has been proposed that prioritizes the tasks based on the length of the instructions. The proposed scheduling approach prioritize the tasks of Cloud applications according to the limits set by six sigma control charts based on dynamic threshold values. Further, the proposed algorithm has been validated through the CloudSim toolkit. The experimental results demonstrate that the proposed algorithm is effective for handling multiple task lists from workflows and in considerably reducing Makespan and Execution time.Keywords: cloud computing, priority based scheduling, task scheduling, VM allocation
Procedia PDF Downloads 517570 Automatic Number Plate Recognition System Based on Deep Learning
Authors: T. Damak, O. Kriaa, A. Baccar, M. A. Ben Ayed, N. Masmoudi
Abstract:
In the last few years, Automatic Number Plate Recognition (ANPR) systems have become widely used in the safety, the security, and the commercial aspects. Forethought, several methods and techniques are computing to achieve the better levels in terms of accuracy and real time execution. This paper proposed a computer vision algorithm of Number Plate Localization (NPL) and Characters Segmentation (CS). In addition, it proposed an improved method in Optical Character Recognition (OCR) based on Deep Learning (DL) techniques. In order to identify the number of detected plate after NPL and CS steps, the Convolutional Neural Network (CNN) algorithm is proposed. A DL model is developed using four convolution layers, two layers of Maxpooling, and six layers of fully connected. The model was trained by number image database on the Jetson TX2 NVIDIA target. The accuracy result has achieved 95.84%.Keywords: ANPR, CS, CNN, deep learning, NPL
Procedia PDF Downloads 306569 The UAV Feasibility Trajectory Prediction Using Convolution Neural Networks
Authors: Adrien Marque, Daniel Delahaye, Pierre Maréchal, Isabelle Berry
Abstract:
Wind direction and uncertainty are crucial in aircraft or unmanned aerial vehicle trajectories. By computing wind covariance matrices on each spatial grid point, these spatial grids can be defined as images with symmetric positive definite matrix elements. A data pre-processing step, a specific convolution, a specific max-pooling, and a specific flatten layers are implemented to process such images. Then, the neural network is applied to spatial grids, whose elements are wind covariance matrices, to solve classification problems related to the feasibility of unmanned aerial vehicles based on wind direction and wind uncertainty.Keywords: wind direction, uncertainty level, unmanned aerial vehicle, convolution neural network, SPD matrices
Procedia PDF Downloads 49568 The Effectiveness of the Family-Centered Sensory and Motor Interactive Games Program on Strengthening the Developmental and Motor Skills of Children aged 12 to 24 Months Who Have a Prior History of Low Birth Weight
Authors: Seyede Soraya Alavinezhad, Gholam Ali Afrooz, Seyedsaeid Sajjadianari
Abstract:
The purpose of this study was to assess the efficacy of a family-centered sensory and motor interactive activities program in enhancing the motor and developmental abilities of infants between the ages of 12 and 24 months who have a medical history of low birth weight. The design of the study was a combined method (qualitative and quantitative). The statistical population comprised infants between the ages of 12 and 24 months who had a documented history of low birth weight in Tehran in 2022. The study sample comprised twenty-eight infants, ranging in age from twelve to twenty-four months, whose mothers were selected using a readily available sampling method. The participants were allocated into two groups—experimental and control—at random. The Children's Developmental Screening Scale, the third edition of Ages and Stages Questionnaires (ASQ3TM), was utilized in both cohorts. Two sessions of the family-centered program for mothers and sixteen sessions for children in the experimental group were taken into account. The statistical analysis software SPSS version 26 was utilized to analyze the data. Initially, the descriptive analysis of the variables, the normality of the assumptions, and the equality of the variance of the variables in the groups were examined. Subsequently, univariate analysis of covariance was employed to examine research hypotheses. The results of the covariance analysis demonstrated that the family-centered interactive activities program for sensory and motor development was effective. A significant difference has been observed between the experimental and control groups with regard to developmental skills between the pre-test and post-test (P<0.005). Motor and developmental skills among children aged 12 to 24 months with a history of low birth weight can be enhanced through entertainment programs that incorporate suitable structure, according to the findings of this study. It is recommended that future research investigate the efficacy of this program on children of average weight and conduct longitudinal studies.Keywords: children, developmental skills, low birth weight, sensory and motor interactive games program
Procedia PDF Downloads 20567 From Myth to Screen: A Cultural Criticism of the Adaptation of Nordic Mythology in Marvel Cinematic Universe’s Thor Trilogy
Authors: Vathya Anindita Putri, Henny Saptatia Drajati Nugrahani
Abstract:
This research aims to explore the representation of Nordic mythology in the commercial film titled “Thor” produced by the Marvel Cinematic Universe. First, the Nordic mythology adaptation and representation in “Thor” compared to other media. Second, the importance of using the mise en scene technique, the comprehensive portrayal of Nordic mythology and the audience's experiences in enjoying the film. This research is conducted using qualitative methods. The two research questions are analyzed using three theories: Adaptation theory by Robert Stam, Mise en Scene theory by Jean-Luc Godard, and Cultural Criticism theory by Michel Foucault. Robert Stam emphasizes the importance of social and historical in understanding film adaptations. Film adaptations always occur in a specific cultural and historical context; therefore, authors and producers must consider these factors when creating a successful adaptation. Jean-Luc Godard uses the “politiques des auteurs” approach to understand that films are not just cultural products made for entertainment, but they are works of art by authors and directors. It is important to explore how authors and directors convey their ideas and emotions in their films, in this case, a film set in Nordic mythology. Foucault takes an approach to analyzing power that considers how power operates and influences social relationships in a specific context. Foucault’s theory is used to analyze how the representation of Nordic mythology is used as an instrument of power by the Marvel Cinematic Universe to influence how the audience views Nordic mythology. The initial findings of this research are that the fusion of Nordic mythology with modern superhero storytelling in the film “Thor” produced by Marvel, is successful. The film contains conflicts in the modern world and represents the symbolism of Nordic mythology. The rich and interesting atmosphere of Nordic mythology is presented through epic battle scenes, captivating character roles, and the use of visual effects that make the film more vivid and real.Keywords: adaptation theory, cultural criticism theory, film criticism, Marvel cinematic universe, Mise en Scene theory, Nordic mythology
Procedia PDF Downloads 86566 Application of Deep Learning in Top Pair and Single Top Quark Production at the Large Hadron Collider
Authors: Ijaz Ahmed, Anwar Zada, Muhammad Waqas, M. U. Ashraf
Abstract:
We demonstrate the performance of a very efficient tagger applies on hadronically decaying top quark pairs as signal based on deep neural network algorithms and compares with the QCD multi-jet background events. A significant enhancement of performance in boosted top quark events is observed with our limited computing resources. We also compare modern machine learning approaches and perform a multivariate analysis of boosted top-pair as well as single top quark production through weak interaction at √s = 14 TeV proton-proton Collider. The most relevant known background processes are incorporated. Through the techniques of Boosted Decision Tree (BDT), likelihood and Multlayer Perceptron (MLP) the analysis is trained to observe the performance in comparison with the conventional cut based and count approachKeywords: top tagger, multivariate, deep learning, LHC, single top
Procedia PDF Downloads 111565 Comfort Sensor Using Fuzzy Logic and Arduino
Authors: Samuel John, S. Sharanya
Abstract:
Automation has become an important part of our life. It has been used to control home entertainment systems, changing the ambience of rooms for different events etc. One of the main parameters to control in a smart home is the atmospheric comfort. Atmospheric comfort mainly includes temperature and relative humidity. In homes, the desired temperature of different rooms varies from 20 °C to 25 °C and relative humidity is around 50%. However, it varies widely. Hence, automated measurement of these parameters to ensure comfort assumes significance. To achieve this, a fuzzy logic controller using Arduino was developed using MATLAB. Arduino is an open source hardware consisting of a 24 pin ATMEGA chip (atmega328), 14 digital input /output pins and an inbuilt ADC. It runs on 5v and 3.3v power supported by a board voltage regulator. Some of the digital pins in Aruduino provide PWM (pulse width modulation) signals, which can be used in different applications. The Arduino platform provides an integrated development environment, which includes support for c, c++ and java programming languages. In the present work, soft sensor was introduced in this system that can indirectly measure temperature and humidity and can be used for processing several measurements these to ensure comfort. The Sugeno method (output variables are functions or singleton/constant, more suitable for implementing on microcontrollers) was used in the soft sensor in MATLAB and then interfaced to the Arduino, which is again interfaced to the temperature and humidity sensor DHT11. The temperature-humidity sensor DHT11 acts as the sensing element in this system. Further, a capacitive humidity sensor and a thermistor were also used to support the measurement of temperature and relative humidity of the surrounding to provide a digital signal on the data pin. The comfort sensor developed was able to measure temperature and relative humidity correctly. The comfort percentage was calculated and accordingly the temperature in the room was controlled. This system was placed in different rooms of the house to ensure that it modifies the comfort values depending on temperature and relative humidity of the environment. Compared to the existing comfort control sensors, this system was found to provide an accurate comfort percentage. Depending on the comfort percentage, the air conditioners and the coolers in the room were controlled. The main highlight of the project is its cost efficiency.Keywords: arduino, DHT11, soft sensor, sugeno
Procedia PDF Downloads 312