Search results for: concurrent computing
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 1165

Search results for: concurrent computing

775 A Convergent Interacting Particle Method for Computing Kpp Front Speeds in Random Flows

Authors: Tan Zhang, Zhongjian Wang, Jack Xin, Zhiwen Zhang

Abstract:

We aim to efficiently compute the spreading speeds of reaction-diffusion-advection (RDA) fronts in divergence-free random flows under the Kolmogorov-Petrovsky-Piskunov (KPP) nonlinearity. We study a stochastic interacting particle method (IPM) for the reduced principal eigenvalue (Lyapunov exponent) problem of an associated linear advection-diffusion operator with spatially random coefficients. The Fourier representation of the random advection field and the Feynman-Kac (FK) formula of the principal eigenvalue (Lyapunov exponent) form the foundation of our method implemented as a genetic evolution algorithm. The particles undergo advection-diffusion and mutation/selection through a fitness function originated in the FK semigroup. We analyze the convergence of the algorithm based on operator splitting and present numerical results on representative flows such as 2D cellular flow and 3D Arnold-Beltrami-Childress (ABC) flow under random perturbations. The 2D examples serve as a consistency check with semi-Lagrangian computation. The 3D results demonstrate that IPM, being mesh-free and self-adaptive, is simple to implement and efficient for computing front spreading speeds in the advection-dominated regime for high-dimensional random flows on unbounded domains where no truncation is needed.

Keywords: KPP front speeds, random flows, Feynman-Kac semigroups, interacting particle method, convergence analysis

Procedia PDF Downloads 35
774 Intelligent System of the Grinding Robot for Spiral Welded Pipe

Authors: Getachew Demeissie Ayalew, Yongtao Sun, Yang Yang

Abstract:

The spiral welded pipe manufacturing industry requires strict production standards for automated grinders for welding seams. However, traditional grinding machines in this sector are insufficient due to a lack of quality control protocols and inconsistent performance. This research aims to improve the quality of spiral welded pipes by developing intelligent automated abrasive belt grinding equipment. The system has equipped with six degrees of freedom (6 DOF) KUKA KR360 industrial robots, enabling concurrent grinding operations on both internal and external welds. The grinding robot control system is designed with a PLC, and a human-machine interface (HMI) system is employed for operations. The system includes an electric speed controller, data connection card, DC driver, analog amplifier, and HMI for input data. This control system enables the grinding of spiral welded pipe. It ensures consistent production quality and cost-effectiveness by reducing the product life cycle and minimizing risks in the working environment.

Keywords: Intelligent Systems, Spiral Welded Pipe, Grinding, Industrial Robot, End-Effector, PLC Controller System, 3D Laser Sensor, HMI.

Procedia PDF Downloads 271
773 'Spare the Rod and Spoil the Child': The Criminal Career of an Armed Robber

Authors: Mahlogonolo Stephina Thobane

Abstract:

The aim of the study upon which this article is based was “to evaluate the possibility of using criminal career research in the development and evaluation of crime control strategies, particularly for armed robberies.” The research employed a concurrent triangulation mixed-method approach where quantitative and qualitative data were collected concurrently but analysed separately through the use of SPSS and Atlas.ti respectively. Forty offenders incarcerated at six correctional centres around the Gauteng province of South Africa for robbery with aggravating circumstances were interviewed as research participants. Since the researcher had no prior information on the total number of the population, purposive sampling (i.e. snowballing) was executed to draw the sample. This research found that offenders launched their criminal career at a very young age of, 11 years, by committing petty crimes such as theft and then, as they grew older, they progressed to more serious and violent crimes such as vehicle hijacking and Cash-in-Transit (CIT) robberies. Thus, it is pivotal that those responsible for developing crime prevention policies focus on interrupting the root causes of crime in the early stages of one’s life in order to prevent continuation of delinquent behaviour from childhood to adolescence and adulthood.

Keywords: criminal career, robbery with aggravating circumstances, cash-in-transit robbery, criminal career research

Procedia PDF Downloads 435
772 Effective Nutrition Label Use on Smartphones

Authors: Vladimir Kulyukin, Tanwir Zaman, Sarat Kiran Andhavarapu

Abstract:

Research on nutrition label use identifies four factors that impede comprehension and retention of nutrition information by consumers: label’s location on the package, presentation of information within the label, label’s surface size, and surrounding visual clutter. In this paper, a system is presented that makes nutrition label use more effective for nutrition information comprehension and retention. The system’s front end is a smartphone application. The system’s back end is a four node Linux cluster for image recognition and data storage. Image frames captured on the smartphone are sent to the back end for skewed or aligned barcode recognition. When barcodes are recognized, corresponding nutrition labels are retrieved from a cloud database and presented to the user on the smartphone’s touchscreen. Each displayed nutrition label is positioned centrally on the touchscreen with no surrounding visual clutter. Wikipedia links to important nutrition terms are embedded to improve comprehension and retention of nutrition information. Standard touch gestures (e.g., zoom in/out) available on mainstream smartphones are used to manipulate the label’s surface size. The nutrition label database currently includes 200,000 nutrition labels compiled from public web sites by a custom crawler. Stress test experiments with the node cluster are presented. Implications for proactive nutrition management and food policy are discussed.

Keywords: mobile computing, cloud computing, nutrition label use, nutrition management, barcode scanning

Procedia PDF Downloads 358
771 BROTHERS: World-class Ergonomic Sofa Development

Authors: Aminur Rahman

Abstract:

The Unique feature of BROTHERS Furniture sofa stands in ergonomic Design, skilled hand work and art work. Present world market is passing through a contentious competitive situation that is rapidly and dramatic. Competitive strategy concerns how to create competitive advantage in upholstery businesses. In order to competitive advantage in upholstery sofa market, Design and develop a sofa that have to ergonomic features. Design an ergonomic upholstery sofa knowing and understanding the appropriate seat depth, seat height, angle between Seat & back, back height which is concurrent market demand, world class sofa has to incorporate ergonomic factors. The study the relationships between human, seat and context variables comfort and discomfort. We must have conduct market survey among users whose are need and use sofa. Health & safety factors should be examined from a variety of angle. An attractive design and meet customer requirements, ergonomically fit should be considered for sofa development. This paper will explain how to design & develop sofa’s as per standard specifications which have ergonomic features for users all over the world.

Keywords: ergonomics, angle between seat & back, standard dimension, seat comfort

Procedia PDF Downloads 127
770 Artificially Intelligent Context Aware Personal Computer Assistant (ACPCA)

Authors: Abdul Mannan Akhtar

Abstract:

In this paper a novel concept of a self learning smart personalized computer assistant (ACPCA) is established which is a context aware system. Based on user habits, moods, and other routines/situational reactions the system will manage various services and suggestions at appropriate times including what schedule to follow, what to watch, what software to be used, what should be deleted etc. This system will utilize a hybrid fuzzyNeural model to predict what the user will do next and support his actions. This will be done by establishing fuzzy sets of user activities, choices, preferences etc. and utilizing their combinations to predict his moods and immediate preferences. Various application of context aware systems exist separately e.g. on certain websites for music or multimedia suggestions but a personalized autonomous system that could adapt to user’s personality does not exist at present. Due to the novelty and massiveness of this concept, this paper will primarily focus on the problem establishment, product features and its functionality; however a small mini case is also implemented on MATLAB to demonstrate some of the aspects of ACPCA. The mini case involves prediction of user moods, activity, routine and food preference using a hybrid fuzzy-Neural soft computing technique.

Keywords: context aware systems, APCPCA, soft computing techniques, artificial intelligence, fuzzy logic, neural network, mood detection, face detection, activity detection

Procedia PDF Downloads 456
769 Context-Aware Alert Method in Hajj Pilgrim Location-Based Tracking System

Authors: Syarif Hidayat

Abstract:

As millions of people with different backgrounds perform hajj every year in Saudi Arabia, it brings out several problems. Missing people is among many crucial problems need to be encountered. Some people might have had insufficient knowledge of using tracking system equipment. Other might become a victim of an accident, lose consciousness, or even died, prohibiting them to perform certain activity. For those reasons, people could not send proper SOS message. The major contribution of this paper is the application of the diverse alert method in pilgrims tracking system. It offers a simple yet robust solution to send SOS message by pilgrims during Hajj. Knowledge of context aware computing is assumed herein. This study presents four methods that could be utilized by pilgrims to send SOS. The first method is simple mobile application contains only a button. The second method is based on behavior analysis based off GPS location movement anomaly. The third method is by introducing pressing pattern to smartwatch physical button as a panic button. The fourth method is by identifying certain accelerometer pattern recognition as a sign of emergency situations. Presented method in this paper would be an important part of pilgrims tracking system. The discussion provided here includes easy to use design whilst maintaining tracking accuracy, privacy, and security of its users.

Keywords: context aware computing, emergency alert system, GPS, hajj pilgrim tracking, location-based services

Procedia PDF Downloads 207
768 Educational Knowledge Transfer in Indigenous Mexican Areas Using Cloud Computing

Authors: L. R. Valencia Pérez, J. M. Peña Aguilar, A. Lamadrid Álvarez, A. Pastrana Palma, H. F. Valencia Pérez, M. Vivanco Vargas

Abstract:

This work proposes a Cooperation-Competitive (Coopetitive) approach that allows coordinated work among the Secretary of Public Education (SEP), the Autonomous University of Querétaro (UAQ) and government funds from National Council for Science and Technology (CONACYT) or some other international organizations. To work on an overall knowledge transfer strategy with e-learning over the Cloud, where experts in junior high and high school education, working in multidisciplinary teams, perform analysis, evaluation, design, production, validation and knowledge transfer at large scale using a Cloud Computing platform. Allowing teachers and students to have all the information required to ensure a homologated nationally knowledge of topics such as mathematics, statistics, chemistry, history, ethics, civism, etc. This work will start with a pilot test in Spanish and initially in two regional dialects Otomí and Náhuatl. Otomí has more than 285,000 speaking indigenes in Queretaro and Mexico´s central region. Náhuatl is number one indigenous dialect spoken in Mexico with more than 1,550,000 indigenes. The phase one of the project takes into account negotiations with indigenous tribes from different regions, and the Information and Communication technologies to deliver the knowledge to the indigenous schools in their native dialect. The methodology includes the following main milestones: Identification of the indigenous areas where Otomí and Náhuatl are the spoken dialects, research with the SEP the location of actual indigenous schools, analysis and inventory or current schools conditions, negotiation with tribe chiefs, analysis of the technological communication requirements to reach the indigenous communities, identification and inventory of local teachers technology knowledge, selection of a pilot topic, analysis of actual student competence with traditional education system, identification of local translators, design of the e-learning platform, design of the multimedia resources and storage strategy for “Cloud Computing”, translation of the topic to both dialects, Indigenous teachers training, pilot test, course release, project follow up, analysis of student requirements for the new technological platform, definition of a new and improved proposal with greater reach in topics and regions. Importance of phase one of the project is multiple, it includes the proposal of a working technological scheme, focusing in the cultural impact in Mexico so that indigenous tribes can improve their knowledge about new forms of crop improvement, home storage technologies, proven home remedies for common diseases, ways of preparing foods containing major nutrients, disclose strengths and weaknesses of each region, communicating through cloud computing platforms offering regional products and opening communication spaces for inter-indigenous cultural exchange.

Keywords: Mexicans indigenous tribes, education, knowledge transfer, cloud computing, otomi, Náhuatl, language

Procedia PDF Downloads 392
767 Power Iteration Clustering Based on Deflation Technique on Large Scale Graphs

Authors: Taysir Soliman

Abstract:

One of the current popular clustering techniques is Spectral Clustering (SC) because of its advantages over conventional approaches such as hierarchical clustering, k-means, etc. and other techniques as well. However, one of the disadvantages of SC is the time consuming process because it requires computing the eigenvectors. In the past to overcome this disadvantage, a number of attempts have been proposed such as the Power Iteration Clustering (PIC) technique, which is one of versions from SC; some of PIC advantages are: 1) its scalability and efficiency, 2) finding one pseudo-eigenvectors instead of computing eigenvectors, and 3) linear combination of the eigenvectors in linear time. However, its worst disadvantage is an inter-class collision problem because it used only one pseudo-eigenvectors which is not enough. Previous researchers developed Deflation-based Power Iteration Clustering (DPIC) to overcome problems of PIC technique on inter-class collision with the same efficiency of PIC. In this paper, we developed Parallel DPIC (PDPIC) to improve the time and memory complexity which is run on apache spark framework using sparse matrix. To test the performance of PDPIC, we compared it to SC, ESCG, ESCALG algorithms on four small graph benchmark datasets and nine large graph benchmark datasets, where PDPIC proved higher accuracy and better time consuming than other compared algorithms.

Keywords: spectral clustering, power iteration clustering, deflation-based power iteration clustering, Apache spark, large graph

Procedia PDF Downloads 178
766 IT Perspective of Service-Oriented e-Government Enterprise

Authors: Anu Paul, Varghese Paul

Abstract:

The focal aspire of e-Government (eGovt) is to offer citizen-centered service delivery. Accordingly, the citizenry consumes services from multiple government agencies through national portal. Thus, eGovt is an enterprise with the primary business motive of transparent, efficient and effective public services to its citizenry and its logical structure is the eGovernment Enterprise Architecture (eGEA). Since eGovt is IT oriented multifaceted service-centric system, EA doesn’t do much on an automated enterprise other than the business artifacts. Service-Oriented Architecture (SOA) manifestation led some governments to pertain this in their eGovts, but it limits the source of business artifacts. The concurrent use of EA and SOA in eGovt executes interoperability and integration and leads to Service-Oriented e-Government Enterprise (SOeGE). Consequently, agile eGovt system becomes a reality. As an IT perspective eGovt comprises of centralized public service artifacts with the existing application logics belong to various departments at central, state and local level. The eGovt is renovating to SOeGE by apply the Service-Orientation (SO) principles in the entire system. This paper explores IT perspective of SOeGE in India which encompasses the public service models and illustrated with a case study the Passport service of India.

Keywords: enterprise architecture, service-oriented e-Government enterprise, service interface layer, service model

Procedia PDF Downloads 502
765 Isolation of Different Brachyspira spp. from Laying Hens in North-East of Iran

Authors: Ahdieh Alijani, Mina Zarrabi, Abdollah Jamshidi, Jamshid Razmyar

Abstract:

Avian intestinal spirochetosis (AIS) is caused by spiral-shaped Gram-negative Brachyspira spp. in poultry and is known as a cause of diarrhea, low egg production and increased the occurrence of dirty eggs in layer hens. In this study the presence of some Brachyspira spp. was investigated in laying hens. A total of 100 cloacal swab samples were individually collected from 20 laying hen flocks showing fecal egg staining in northeastern Iran. By culture and morphologic examination, 41 samples (41%) from 20 flocks were positive but by using genus–specific PCR only 37 (37%) samples were confirmed as Brachyspira spp. Using species-specific primers, single colonization was identified in 18 samples associated with B. pilosicoli (48.6%) while single colonization with B. intermedia was found in only two samples (5.4%). Simultaneous colonization by B. intermedia and B. murdochii was detected in 3 samples (8.1%). B. pilosicoli was the most prevalent species in concurrent colonization in 11 cases (29.7%). Finally, co-colonization by B. intermedia and B. innocens was identified in 3 samples (8.1%). The results of this study show the colonization of different species of Brachyspira with the dominance of B. pilosicoli in layer hens. In simultaneous colonization with pathogenic and non-pathogenic species the symptoms of intestinal spirochetosis were reduced, suggesting a competitive role in preventing and reducing the colonization of pathogenic species.

Keywords: intestinal spirochetosis, Brachyspira, laying hen, PCR

Procedia PDF Downloads 531
764 A Methodology for Optimisation of Water Containment Systems

Authors: Amir Hedjripour

Abstract:

The required dewatering configuration for a contaminated sediment dam is discussed to meet no-spill criteria for a defined Average Recurrence Interval (ARI). There is an option for the sediment dam to pump the contaminated water to another storage facility before its capacity is exceeded. The system is subjected to a range of storm durations belonging to the design ARI with concurrent dewatering to the other storage facility. The model is set up in 1-minute time intervals and temporal patterns of storm events are used to de-segregate the total storm depth into partial durations. By running the model for selected storm durations, the maximum water volume in the dam is recorded as the critical volume, which indicates the required storage capacity for that storm duration. Runoff from upstream catchment and the direct rainfall over the dam open area are calculated by taking into account the time of concentration for the catchment. Total 99 different storm durations from 5 minutes to 72 hours were modelled together with five dewatering scenarios from 50 l/s to 500 l/s. The optimised dam/pump configuration is selected by plotting critical points for all cases and storage-dewatering envelopes. A simple economic analysis is also presented in the paper using Present-Value (PV) analysis to assist with the financial evaluation of each configuration and selection of the best alternative.

Keywords: contaminated water, optimisation, pump, sediment dam

Procedia PDF Downloads 356
763 Challenge of Baseline Hydrology Estimation at Large-Scale Watersheds

Authors: Can Liu, Graham Markowitz, John Balay, Ben Pratt

Abstract:

Baseline or natural hydrology is commonly employed for hydrologic modeling and quantification of hydrologic alteration due to manmade activities. It can inform planning and policy related efforts for various state and federal water resource agencies to restore natural streamflow flow regimes. A common challenge faced by hydrologists is how to replicate unaltered streamflow conditions, particularly in large watershed settings prone to development and regulation. Three different methods were employed to estimate baseline streamflow conditions for 6 major subbasins the Susquehanna River Basin; those being: 1) incorporation of consumptive water use and reservoir operations back into regulated gaged records; 2) using a map correlation method and flow duration (exceedance probability) regression equations; 3) extending the pre-regulation streamflow records based on the relationship between concurrent streamflows at unregulated and regulated gage locations. Parallel analyses were perform among the three methods and limitations associated with each are presented. Results from these analyses indicate that generating baseline streamflow records at large-scale watersheds remain challenging, even with long-term continuous stream gage records available.

Keywords: baseline hydrology, streamflow gage, subbasin, regression

Procedia PDF Downloads 315
762 An Adiabatic Quantum Optimization Approach for the Mixed Integer Nonlinear Programming Problem

Authors: Maxwell Henderson, Tristan Cook, Justin Chan Jin Le, Mark Hodson, YoungJung Chang, John Novak, Daniel Padilha, Nishan Kulatilaka, Ansu Bagchi, Sanjoy Ray, John Kelly

Abstract:

We present a method of using adiabatic quantum optimization (AQO) to solve a mixed integer nonlinear programming (MINLP) problem instance. The MINLP problem is a general form of a set of NP-hard optimization problems that are critical to many business applications. It requires optimizing a set of discrete and continuous variables with nonlinear and potentially nonconvex constraints. Obtaining an exact, optimal solution for MINLP problem instances of non-trivial size using classical computation methods is currently intractable. Current leading algorithms leverage heuristic and divide-and-conquer methods to determine approximate solutions. Creating more accurate and efficient algorithms is an active area of research. Quantum computing (QC) has several theoretical benefits compared to classical computing, through which QC algorithms could obtain MINLP solutions that are superior to current algorithms. AQO is a particular form of QC that could offer more near-term benefits compared to other forms of QC, as hardware development is in a more mature state and devices are currently commercially available from D-Wave Systems Inc. It is also designed for optimization problems: it uses an effect called quantum tunneling to explore all lowest points of an energy landscape where classical approaches could become stuck in local minima. Our work used a novel algorithm formulated for AQO to solve a special type of MINLP problem. The research focused on determining: 1) if the problem is possible to solve using AQO, 2) if it can be solved by current hardware, 3) what the currently achievable performance is, 4) what the performance will be on projected future hardware, and 5) when AQO is likely to provide a benefit over classical computing methods. Two different methods, integer range and 1-hot encoding, were investigated for transforming the MINLP problem instance constraints into a mathematical structure that can be embedded directly onto the current D-Wave architecture. For testing and validation a D-Wave 2X device was used, as well as QxBranch’s QxLib software library, which includes a QC simulator based on simulated annealing. Our results indicate that it is mathematically possible to formulate the MINLP problem for AQO, but that currently available hardware is unable to solve problems of useful size. Classical general-purpose simulated annealing is currently able to solve larger problem sizes, but does not scale well and such methods would likely be outperformed in the future by improved AQO hardware with higher qubit connectivity and lower temperatures. If larger AQO devices are able to show improvements that trend in this direction, commercially viable solutions to the MINLP for particular applications could be implemented on hardware projected to be available in 5-10 years. Continued investigation into optimal AQO hardware architectures and novel methods for embedding MINLP problem constraints on to those architectures is needed to realize those commercial benefits.

Keywords: adiabatic quantum optimization, mixed integer nonlinear programming, quantum computing, NP-hard

Procedia PDF Downloads 514
761 Reliability of Swine Estrous Detector Probe in Dairy Cattle Breeding

Authors: O. O. Leigh, L. C. Agbugba, A. O. Oyewunmi, A. E. Ibiam, A. Hassan

Abstract:

Accuracy of insemination timing is a key determinant of high pregnancy rates in livestock breeding stations. The estrous detector probes are a recent introduction into the Nigerian livestock farming sector. Many of these probes are species-labeled and they measure changes in the vaginal mucus resistivity (VMR) during the stages of the estrous cycle. With respect to size and shaft conformation, the Draminski® swine estrous detector probe (sEDP) is quite similar to the bovine estrous detector probe. We investigated the reliability of the sEDP at insemination time on two farms designated as FM A and FM B. Cows (Bunaji, n=20 per farm) were evaluated for VMR at 16th h post standard OvSynch protocol, with concurrent insemination on FM B only. The difference in the mean VMR between FM A (221 ± 24.36) Ohms and FM B (254 ± 35.59) Ohms was not significant (p > 0.05). Sixteen cows (80%) at FM B were later (day 70) confirmed pregnant via rectal palpation and calved at term. These findings suggest consistency in VMR evaluated with sEDP at insemination as well as a high predictability for VMR associated with good pregnancy rates in dairy cattle. We conclude that Draminski® swine estrous detector probe is reliable in determining time of insemination in cattle breeding stations.

Keywords: dairy cattle, insemination, swine estrous probe, vaginal mucus resistivity

Procedia PDF Downloads 116
760 Harnessing the Power of Feedback to Assist Progress: A Process-Based Approach of Providing Feedback to L2 Composition Students in the United Arab Emirates

Authors: Brad Curabba

Abstract:

Utilising active, process-based learning methods to improve critical thinking and writing skills of second language (L2) writers brings unique challenges. To comprehensively satisfy different learners' needs, when commenting on student work, instructors can embed multiple feedback methods so that the capstone of their abilities as writers can be achieved. This research project assesses faculty and student perceptions regarding the effectiveness of various feedback practices used in process-based writing classrooms with L2 students at the American University of Sharjah (AUS). In addition, the research explores the challenges encountered by faculty during the provision of feedback practices. The quantitative research findings are based on two concurrent electronically distributed anonymous surveys; one aimed at students who have just completed a process-based writing course, and the other at instructors who delivered these courses. The student sample is drawn from multiple sections of Academic Writing I and II, and the faculty survey was distributed among the Department of Writing Studies (DWS) faculty. Our findings strongly suggest that all methods of feedback are deemed equally important by both students and faculty. Students, in particular, find process writing and its feedback practices to have greatly contributed to their writing proficiency.

Keywords: process writing, feedback, formative feedback, composition, reflection

Procedia PDF Downloads 121
759 Challenging Human Trade in Sub-Saharan Africa and Beyond: A Foresight Approach to Contextualizing and Understanding the Consequences of Sub-Saharan Africa’s Demographic Emergence

Authors: Ricardo Schnug

Abstract:

This paper puts the transnational crime of human trafficking in the context of Sub-Saharan Africa and its quickly growing youth bulge. By mapping recent and concurrent trends and emerging issues, it explores the implications that it has not only for the region itself but also for the greater global dynamics of the issue. Through the application of Causal Layered Analysis to various alternative future scenarios as well as the identification of the core narrative surrounding the international discourse, it is possible to understand more deeply the forces that underlie future trafficking and what change becomes possible. With the provision of a reconstructed narrative that avoids the current blind spots, this research points out the need for a new and organic leadership paradigm that allows for a more holistic and future-oriented inquiry about socio-economic and political change and what it entails for a transnational crime such as human trafficking. 'Ubuntu' as a social and leadership philosophy then, provides the principles needed for creating this path towards a truly preferred future. Furthermore, this paper inspires follow-up research and the continuous monitoring and transdisciplinary research of this region’s demographic emergence as well as its possible consequences that have been explored in this inquiry.

Keywords: causal layered analysis, emerging issues, human trafficking, scenarios, sub-Saharan Africa

Procedia PDF Downloads 183
758 Comparative Study and Parallel Implementation of Stochastic Models for Pricing of European Options Portfolios using Monte Carlo Methods

Authors: Vinayak Bassi, Rajpreet Singh

Abstract:

Over the years, with the emergence of sophisticated computers and algorithms, finance has been quantified using computational prowess. Asset valuation has been one of the key components of quantitative finance. In fact, it has become one of the embryonic steps in determining risk related to a portfolio, the main goal of quantitative finance. This study comprises a drawing comparison between valuation output generated by two stochastic dynamic models, namely Black-Scholes and Dupire’s bi-dimensionality model. Both of these models are formulated for computing the valuation function for a portfolio of European options using Monte Carlo simulation methods. Although Monte Carlo algorithms have a slower convergence rate than calculus-based simulation techniques (like FDM), they work quite effectively over high-dimensional dynamic models. A fidelity gap is analyzed between the static (historical) and stochastic inputs for a sample portfolio of underlying assets. In order to enhance the performance efficiency of the model, the study emphasized the use of variable reduction methods and customizing random number generators to implement parallelization. An attempt has been made to further implement the Dupire’s model on a GPU to achieve higher computational performance. Furthermore, ideas have been discussed around the performance enhancement and bottleneck identification related to the implementation of options-pricing models on GPUs.

Keywords: monte carlo, stochastic models, computational finance, parallel programming, scientific computing

Procedia PDF Downloads 147
757 Enabling Cloud Adoption Based Secured Mobile Banking through Backend as a Service

Authors: P. S. Jagadeesh Kumar, S. Meenakshi Sundaram

Abstract:

With the increase of prevailing non-traditional rivalry, mobile banking experiences an ever changing commercial backdrop. Substantial customer demands have established to be more intricate as customers request more expediency and superintend over their banking services. To enterprise advance and modernization in mobile banking applications, it is gradually obligatory to deeply leapfrog the scuffle using business model transformation. The dramaturgical vicissitudes taking place in mobile banking entail advanced traditions to exploit security. By reforming and transforming older back office into integrated mobile banking applications, banks can engender a supple and nimble banking environment that can rapidly respond to new business requirements over cloud computing. Cloud computing is transfiguring ecosystems in numerous industries, and mobile banking is no exemption providing services innovation, greater flexibility to respond to improved security and enhanced business intelligence with less cost. Cloud technology offer secure deployment possibilities that can provision banks in developing new customer experiences, empower operative relationship and advance speed to efficient banking transaction. Cloud adoption is escalating quickly since it can be made secured for commercial mobile banking transaction through backend as a service in scrutinizing the security strategies of the cloud service provider along with the antiquity of transaction details and their security related practices.

Keywords: cloud adoption, backend as a service, business intelligence, secured mobile banking

Procedia PDF Downloads 241
756 Enhancing Internet of Things Security: A Blockchain-Based Approach for Preventing Spoofing Attacks

Authors: Salha Abdullah Ali Al-Shamrani, Maha Muhammad Dhaher Aljuhani, Eman Ali Ahmed Aldhaheri

Abstract:

With the proliferation of Internet of Things (IoT) devices in various industries, there has been a concurrent rise in security vulnerabilities, particularly spoofing attacks. This study explores the potential of blockchain technology in enhancing the security of IoT systems and mitigating these attacks. Blockchain's decentralized and immutable ledger offers significant promise for improving data integrity, transaction transparency, and tamper-proofing. This research develops and implements a blockchain-based IoT architecture and a reference network to simulate real-world scenarios and evaluate a blockchain-integrated intrusion detection system. Performance measures including time delay, security, and resource utilization are used to assess the system's effectiveness, comparing it to conventional IoT networks without blockchain. The results provide valuable insights into the practicality and efficacy of employing blockchain as a security mechanism, shedding light on the trade-offs between speed and security in blockchain deployment for IoT. The study concludes that despite minor increases in time consumption, the security benefits of incorporating blockchain technology into IoT systems outweigh potential drawbacks, demonstrating a significant potential for blockchain in bolstering IoT security.

Keywords: internet of things, spoofing, IoT, access control, blockchain, raspberry pi

Procedia PDF Downloads 58
755 A Framework of Virtualized Software Controller for Smart Manufacturing

Authors: Pin Xiu Chen, Shang Liang Chen

Abstract:

A virtualized software controller is developed in this research to replace traditional hardware control units. This virtualized software controller transfers motion interpolation calculations from the motion control units of end devices to edge computing platforms, thereby reducing the end devices' computational load and hardware requirements and making maintenance and updates easier. The study also applies the concept of microservices, dividing the control system into several small functional modules and then deploy into a cloud data server. This reduces the interdependency among modules and enhances the overall system's flexibility and scalability. Finally, with containerization technology, the system can be deployed and started in a matter of seconds, which is more efficient than traditional virtual machine deployment methods. Furthermore, this virtualized software controller communicates with end control devices via wireless networks, making the placement of production equipment or the redesign of processes more flexible and no longer limited by physical wiring. To handle the large data flow and maintain low-latency transmission, this study integrates 5G technology, fully utilizing its high speed, wide bandwidth, and low latency features to achieve rapid and stable remote machine control. An experimental setup is designed to verify the feasibility and test the performance of this framework. This study designs a smart manufacturing site with a 5G communication architecture, serving as a field for experimental data collection and performance testing. The smart manufacturing site includes one robotic arm, three Computer Numerical Control machine tools, several Input/Output ports, and an edge computing architecture. All machinery information is uploaded to edge computing servers and cloud servers via 5G communication and the Internet of Things framework. After analysis and computation, this information is converted into motion control commands, which are transmitted back to the relevant machinery for motion control through 5G communication. The communication time intervals at each stage are calculated using the C++ chrono library to measure the time difference for each command transmission. The relevant test results will be organized and displayed in the full-text.

Keywords: 5G, MEC, microservices, virtualized software controller, smart manufacturing

Procedia PDF Downloads 66
754 Security Design of Root of Trust Based on RISC-V

Authors: Kang Huang, Wanting Zhou, Shiwei Yuan, Lei Li

Abstract:

Since information technology develops rapidly, the security issue has become an increasingly critical for computer system. In particular, as cloud computing and the Internet of Things (IoT) continue to gain widespread adoption, computer systems need to new security threats and attacks. The Root of Trust (RoT) is the foundation for providing basic trusted computing, which is used to verify the security and trustworthiness of other components. Design a reliable Root of Trust and guarantee its own security are essential for improving the overall security and credibility of computer systems. In this paper, we discuss the implementation of self-security technology based on the RISC-V Root of Trust at the hardware level. To effectively safeguard the security of the Root of Trust, researches on security safeguard technology on the Root of Trust have been studied. At first, a lightweight and secure boot framework is proposed as a secure mechanism. Secondly, two kinds of memory protection mechanism are built to against memory attacks. Moreover, hardware implementation of proposed method has been also investigated. A series of experiments and tests have been carried on to verify to effectiveness of the proposed method. The experimental results demonstrated that the proposed approach is effective in verifying the integrity of the Root of Trust’s own boot rom, user instructions, and data, ensuring authenticity and enabling the secure boot of the Root of Trust’s own system. Additionally, our approach provides memory protection against certain types of memory attacks, such as cache leaks and tampering, and ensures the security of root-of-trust sensitive information, including keys.

Keywords: root of trust, secure boot, memory protection, hardware security

Procedia PDF Downloads 185
753 Application of Griddization Management to Construction Hazard Management

Authors: Lingzhi Li, Jiankun Zhang, Tiantian Gu

Abstract:

Hazard management that can prevent fatal accidents and property losses is a fundamental process during the buildings’ construction stage. However, due to lack of safety supervision resources and operational pressures, the conduction of hazard management is poor and ineffective in China. In order to improve the quality of construction safety management, it is critical to explore the use of information technologies to ensure that the process of hazard management is efficient and effective. After exploring the existing problems of construction hazard management in China, this paper develops the griddization management model for construction hazard management. First, following the knowledge grid infrastructure, the griddization computing infrastructure for construction hazards management is designed which includes five layers: resource entity layer, information management layer, task management layer, knowledge transformation layer and application layer. This infrastructure will be as the technical support for realizing grid management. Second, this study divides the construction hazards into grids through city level, district level and construction site level according to grid principles. Last, a griddization management process including hazard identification, assessment and control is developed. Meanwhile, all stakeholders of construction safety management, such as owners, contractors, supervision organizations and government departments, should take the corresponding responsibilities in this process. Finally, a case study based on actual construction hazard identification, assessment and control is used to validate the effectiveness and efficiency of the proposed griddization management model. The advantage of this designed model is to realize information sharing and cooperative management between various safety management departments.

Keywords: construction hazard, griddization computing, grid management, process

Procedia PDF Downloads 263
752 EECS: Reimagining the Future of Technology Education through Electrical Engineering and Computer Science Integration

Authors: Yousef Sharrab, Dimah Al-Fraihat, Monther Tarawneh, Aysh Alhroob, Ala’ Khalifeh, Nabil Sarhan

Abstract:

This paper explores the evolution of Electrical Engineering (EE) and Computer Science (CS) education in higher learning, examining the feasibility of unifying them into Electrical Engineering and Computer Science (EECS) for the technology industry. It delves into the historical reasons for their separation and underscores the need for integration. Emerging technologies such as AI, Virtual Reality, IoT, Cloud Computing, and Cybersecurity demand an integrated EE and CS program to enhance students' understanding. The study evaluates curriculum integration models, drawing from prior research and case studies, demonstrating how integration can provide students with a comprehensive knowledge base for industry demands. Successful integration necessitates addressing administrative and pedagogical challenges. For academic institutions considering merging EE and CS programs, the paper offers guidance, advocating for a flexible curriculum encompassing foundational courses and specialized tracks in computer engineering, software engineering, bioinformatics, information systems, data science, AI, robotics, IoT, virtual reality, cybersecurity, and cloud computing. Elective courses are emphasized to keep pace with technological advancements. Implementing this integrated approach can prepare students for success in the technology industry, addressing the challenges of a technologically advanced society reliant on both EE and CS principles. Integrating EE and CS curricula is crucial for preparing students for the future.

Keywords: electrical engineering, computer science, EECS, curriculum integration of EE and CS

Procedia PDF Downloads 47
751 Validating the Arabic Communicative Development Inventory for Assessing the Development of Language in Arabic-Speaking Children

Authors: Alshaimaa Abdelwahab, Allegra Cattani, Caroline Floccia

Abstract:

Assessing children’s language is fundamental for changing their developmental outcome as it gives a chance for a quick and early intervention with the suitable planning and monitoring program. The importance of language assessment lies in helping to find the right test fit for purpose, in addition to achievement and proficiency. This study examines the validity of a new Arabic assessment tool, the Arabic Communicative Development Inventory ‘Arabic CDI’. It assesses the development of language in Arabic children in different Arabic countries, allowing to detect children with language delay. A concurrent validity is set to compare the Arabic CDI to the Arabic Language test. Twenty-three typically developing Egyptian healthy children and their mothers participated in this study. Their age is 24 months (+ or -) two weeks. The sample included 13 males and 10 females. Mothers completed the Arabic CDI either before or after the Arabic Language Test was conducted with the child. The score for comprehension in the Arabic CDI (M= 52.7, SD= 9.7) and words understood in the Arabic Language Test (M= 59.6, SD= 12.5) were strongly and positively correlated (r= .62, p= .002). At the same time, the scores for production in the Arabic CDI (M= 38.4, SD= 14.8) and words expressed in the Arabic Language Test (M= 52.1, SD= 16.3) were also strongly and positively correlated (r= .82, p= .000). The new Arabic CDI is an adequate tool for assessing the development of comprehension and production at Arabic children. In addition, it could be used for detecting children with language impairment. Standardization of the Arabic CDI across 18 different Arabic dialects in children aged 8 to 30 months is underway.

Keywords: Arabic CDI, assessing children, language development, language impairment

Procedia PDF Downloads 448
750 Cloud Resources Utilization and Science Teacher’s Effectiveness in Secondary Schools in Cross River State, Nigeria

Authors: Michael Udey Udam

Abstract:

Background: This study investigated the impact of cloud resources, a component of cloud computing, on science teachers’ effectiveness in secondary schools in Cross River State. Three (3) research questions and three (3) alternative hypotheses guided the study. Method: The descriptive survey design was adopted for the study. The population of the study comprised 1209 science teachers in public secondary schools of Cross River state. Sample: A sample of 487 teachers was drawn from the population using a stratified random sampling technique. The researcher-made structured questionnaire with 18 was used for data collection for the study. Research question one was answered using the Pearson Product Moment Correlation, while research question two and the hypotheses were answered using the Analysis of Variance (ANOVA) statistics in the Statistical Package for Social Sciences (SPSS) at a 0.05 level of significance. Results: The results of the study revealed that there is a positive correlation between the utilization of cloud resources in teaching and teaching effectiveness among science teachers in secondary schools in Cross River state; there is a negative correlation between gender and utilization of cloud resources among science teachers in secondary schools in Cross River state; and that there is a significant correlation between teaching experience and the utilization of cloud resources among science teachers in secondary schools in Cross River state. Conclusion: The study justifies the effectiveness of the Cross River state government policy of introducing cloud computing into the education sector. The study recommends that the policy should be sustained.

Keywords: cloud resources, science teachers, effectiveness, secondary school

Procedia PDF Downloads 59
749 An Adjoint-Based Method to Compute Derivatives with Respect to Bed Boundary Positions in Resistivity Measurements

Authors: Mostafa Shahriari, Theophile Chaumont-Frelet, David Pardo

Abstract:

Resistivity measurements are used to characterize the Earth’s subsurface. They are categorized into two different groups: (a) those acquired on the Earth’s surface, for instance, controlled source electromagnetic (CSEM) and Magnetotellurics (MT), and (b) those recorded with borehole logging instruments such as Logging-While-Drilling (LWD) devices. LWD instruments are mostly used for geo-steering purposes, i.e., to adjust dip and azimuthal angles of a well trajectory to drill along a particular geological target. Modern LWD tools measure all nine components of the magnetic field corresponding to three orthogonal transmitter and receiver orientations. In order to map the Earth’s subsurface and perform geo-steering, we invert measurements using a gradient-based method that utilizes the derivatives of the recorded measurements with respect to the inversion variables. For resistivity measurements, these inversion variables are usually the constant resistivity value of each layer and the bed boundary positions. It is well-known how to compute derivatives with respect to the constant resistivity value of each layer using semi-analytic or numerical methods. However, similar formulas for computing the derivatives with respect to bed boundary positions are unavailable. The main contribution of this work is to provide an adjoint-based formulation for computing derivatives with respect to the bed boundary positions. The key idea to obtain the aforementioned adjoint state formulations for the derivatives is to separate the tangential and normal components of the field and treat them differently. This formulation allows us to compute the derivatives faster and more accurately than with traditional finite differences approximations. In the presentation, we shall first derive a formula for computing the derivatives with respect to the bed boundary positions for the potential equation. Then, we shall extend our formulation to 3D Maxwell’s equations. Finally, by considering a 1D domain and reducing the dimensionality of the problem, which is a common practice in the inversion of resistivity measurements, we shall derive a formulation to compute the derivatives of the measurements with respect to the bed boundary positions using a 1.5D variational formulation. Then, we shall illustrate the accuracy and convergence properties of our formulations by comparing numerical results with the analytical derivatives for the potential equation. For the 1.5D Maxwell’s system, we shall compare our numerical results based on the proposed adjoint-based formulation vs those obtained with a traditional finite difference approach. Numerical results shall show that our proposed adjoint-based technique produces enhanced accuracy solutions while its cost is negligible, as opposed to the finite difference approach that requires the solution of one additional problem per derivative.

Keywords: inverse problem, bed boundary positions, electromagnetism, potential equation

Procedia PDF Downloads 169
748 Phosphorus Recovery Optimization in Microbial Fuel Cell

Authors: Abdullah Almatouq

Abstract:

Understanding the impact of key operational variables on concurrent energy generation and phosphorus recovery in microbial fuel cell is required to improve the process and reduce the operational cost. In this study, full factorial design (FFD) and central composite designs (CCD) were employed to identify the effect of influent COD concentration and cathode aeration flow rate on energy generation and phosphorus (P) recovery and to optimise MFC power density and P recovery. Results showed that influent chemical oxygen demand (COD) concentration and cathode aeration flow rate had a significant effect on power density, coulombic efficiency, phosphorus precipitation efficiency and phosphorus precipitation rate at the cathode. P precipitation was negatively affected by the generated current during the batch duration. The generated energy was reduced due to struvite being precipitated on the cathode surface, which might obstruct the mass transfer of ions and oxygen. Response surface mathematical model was used to predict the optimum operating conditions that resulted in a maximum power density and phosphorus precipitation efficiency of 184 mW/m² and 84%, and this corresponds to COD= 1700 mg/L and aeration flow rate=210 mL/min. The findings highlight the importance of the operational conditions of energy generation and phosphorus recovery.

Keywords: energy, microbial fuel cell, phosphorus, struvite

Procedia PDF Downloads 145
747 Analyzing the Nutritional Challenges in Old People with Diabetes

Authors: Maedeh Gharazi

Abstract:

Adults with age 50 and older will include more than 70% of the diabetic populace by the year 2025. More established patients with diabetes are more inclined to have concurrent ceaseless conditions like hypertension, dyslipidemia, and cardiovascular sickness that may affect their nutritious necessities. The issue of achievement and support of an ideal body weight in elderly diabetic persons may not be as direct as in other age gatherings, and the risk-benefit ratio may be diverse too. Albeit expanded predominance of overweight and weight in the elderly adds to insulin resistance and hyperglycemia, more seasoned tenants of long haul care offices who experience the ill effects of diabetes have a tendency to be underweight. Both may mean insufficient nutritional status and lead to expanded grimness and mortality. The attendant problems of appetite changes, palatability of food, dietary restrictions, loneliness, and depression may influence the sort and amount of food devoured by elderly persons. Organized screening devices may recognize nutrition related issues that warrant proof based mediations. Despite the fact that glucose control and health concerns are essential calculates diet change in the more established populace, different contemplations incorporate personal satisfaction and individual inclinations. Redoing of nutritious rules to the needs of the more seasoned diabetic patient bodes well.

Keywords: diabetes, nutritious necessities, insulin resistance, glucose control

Procedia PDF Downloads 322
746 Optimizing Data Transfer and Processing in Multi-Cloud Environments for Big Data Workloads

Authors: Gaurav Kumar Sinha

Abstract:

In an era defined by the proliferation of data and the utilization of cloud computing environments, the efficient transfer and processing of big data workloads across multi-cloud platforms have emerged as critical challenges. This research paper embarks on a comprehensive exploration of the complexities associated with managing and optimizing big data in a multi-cloud ecosystem.The foundation of this study is rooted in the recognition that modern enterprises increasingly rely on multiple cloud providers to meet diverse business needs, enhance redundancy, and reduce vendor lock-in. As a consequence, managing data across these heterogeneous cloud environments has become intricate, necessitating innovative approaches to ensure data integrity, security, and performance.The primary objective of this research is to investigate strategies and techniques for enhancing the efficiency of data transfer and processing in multi-cloud scenarios. It recognizes that big data workloads are characterized by their sheer volume, variety, velocity, and complexity, making traditional data management solutions insufficient for harnessing the full potential of multi-cloud architectures.The study commences by elucidating the challenges posed by multi-cloud environments in the context of big data. These challenges encompass data fragmentation, latency, security concerns, and cost optimization. To address these challenges, the research explores a range of methodologies and solutions. One of the key areas of focus is data transfer optimization. The paper delves into techniques for minimizing data movement latency, optimizing bandwidth utilization, and ensuring secure data transmission between different cloud providers. It evaluates the applicability of dedicated data transfer protocols, intelligent data routing algorithms, and edge computing approaches in reducing transfer times.Furthermore, the study examines strategies for efficient data processing across multi-cloud environments. It acknowledges that big data processing requires distributed and parallel computing capabilities that span across cloud boundaries. The research investigates containerization and orchestration technologies, serverless computing models, and interoperability standards that facilitate seamless data processing workflows.Security and data governance are paramount concerns in multi-cloud environments. The paper explores methods for ensuring data security, access control, and compliance with regulatory frameworks. It considers encryption techniques, identity and access management, and auditing mechanisms as essential components of a robust multi-cloud data security strategy.The research also evaluates cost optimization strategies, recognizing that the dynamic nature of multi-cloud pricing models can impact the overall cost of data transfer and processing. It examines approaches for workload placement, resource allocation, and predictive cost modeling to minimize operational expenses while maximizing performance.Moreover, this study provides insights into real-world case studies and best practices adopted by organizations that have successfully navigated the challenges of multi-cloud big data management. It presents a comparative analysis of various multi-cloud management platforms and tools available in the market.

Keywords: multi-cloud environments, big data workloads, data transfer optimization, data processing strategies

Procedia PDF Downloads 54