Search results for: execution time
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 17941

Search results for: execution time

17851 Incorporating Priority Round-Robin Scheduler to Sustain Indefinite Blocking Issue and Prioritized Processes in Operating System

Authors: Heng Chia Ying, Charmaine Tan Chai Nie, Burra Venkata Durga Kumar

Abstract:

Process scheduling is the method of process management that determines which process the CPU will proceed with for the next task and how long it takes. Some issues were found in process management, particularly for Priority Scheduling (PS) and Round Robin Scheduling (RR). The proposed recommendations made for IPRRS are to combine the strengths of both into a combining algorithm while they draw on others to compensate for each weakness. A significant improvement on the combining technique of scheduler, Incorporating Priority Round-Robin Scheduler (IPRRS) address an algorithm for both high and low priority task to sustain the indefinite blocking issue faced in the priority scheduling algorithm and minimize the average turnaround time (ATT) and average waiting time (AWT) in RR scheduling algorithm. This paper will delve into the simple rules introduced by IPRRS and enhancements that both PS and RR bring to the execution of processes in the operating system. Furthermore, it incorporates the best aspects of each algorithm to build the optimum algorithm for a certain case in terms of prioritized processes, ATT, and AWT.

Keywords: round Robin scheduling, priority scheduling, indefinite blocking, process management, sustain, turnaround time

Procedia PDF Downloads 99
17850 A Multidimensional Genetic Algorithm Applicable for Our VRP Variant Dealing with the Problems of Infrastructure Defaults SVRDP-CMTW: “Safety Vehicle Routing Diagnosis Problem with Control and Modified Time Windows”

Authors: Ben Mansour Mouin, Elloumi Abdelkarim

Abstract:

We will discuss the problem of routing a fleet of different vehicles from a central depot to different types of infrastructure-defaults with dynamic maintenance requests, modified time windows, and control of default maintained. For this reason, we propose a modified metaheuristicto to solve our mathematical model. SVRDP-CMTW is a variant VRP of an optimal vehicle plan that facilitates the maintenance task of different types of infrastructure-defaults. This task will be monitored after the maintenance, based on its priorities, the degree of danger associated with each default, and the neighborhood at the black-spots. We will present, in this paper, a multidimensional genetic algorithm “MGA” by detailing its characteristics, proposed mechanisms, and roles in our work. The coding of this algorithm represents the necessary parameters that characterize each infrastructure-default with the objective of minimizing a combination of cost, distance and maintenance times while satisfying the priority levels of the most urgent defaults. The developed algorithm will allow the dynamic integration of newly detected defaults at the execution time. This result will be displayed in our programmed interactive system at the routing time. This multidimensional genetic algorithm replaces N genetic algorithm to solve P different type problems of infrastructure defaults (instead of N algorithm for P problem we can solve in one multidimensional algorithm simultaneously who can solve all these problemsatonce).

Keywords: mathematical model, VRP, multidimensional genetic algorithm, metaheuristics

Procedia PDF Downloads 160
17849 Assessing the Impact of Industry 4.0 Implementation on Carbon Neutrality in industries

Authors: Sepinoud Hamedi

Abstract:

The ponder points to observationally look at the impact of carbon-neutrality approaches on the key assets required for Industry 4.0 driven savvy fabricating and how these assets can give a economical competitive advantage. The hypothetical system is coordinates with the regulation hypothesis and the resource-based see (RBV). The observational strategy is utilized for collecting information through studies and assist covariance-based auxiliary condition modeling is utilized to test the theories. Discoveries demonstrate that carbon–neutral-based government arrangements have a more grounded impact on unmistakable assets and human aptitudes than intangible assets related to Industry 4.0 driven shrewd fabricating. Moment, carbon–neutral arrangement arrangement with the firm’s maintainability destinations plays a directing impact on the relationship between carbon–neutral-based government arrangements and assets (substantial, intangible assets and human abilities) for Industry 4.0 driven shrewd fabricating. Finally, the three assets (substantial, intangible assets and human abilities) for Industry 4.0 driven savvy fabricating play a basic part in creating firms’ carbon–neutral capability and assist improving operational execution. Administrative suggestions incorporate venture in progressed advanced innovations, creating a solid mentality among workers and supply chain partners, and planning preparing programs for upgrading shrewd fabricating execution to create carbon-neutrality capability. This think about proposes a crossover hypothesis within the setting of carbon nonpartisanship by coordination institutional theory and RBV. Typically the primary think about that looks at the impact of carbon neutrality-based government arrangements on crucial Industry 4.0-driven savvy fabricating assets and the circuitous impact on carbon nonpartisanship capability and operational execution.

Keywords: carbon, industry 4.0, neutrality, RBV, nonpartisanship

Procedia PDF Downloads 37
17848 Factors Constraining the Utilization of Risk Management Strategies in the Execution of Public Construction Projects in North East Nigeria

Authors: S. U. Kunya, S. A. Mohammad

Abstract:

Construction projects in Nigeria are characterized with risks emanating from delays and accompanying cost-overruns. The aim of the study was to identify and assess factors constraining the utilization of risk management strategies in the execution of public construction project in North-East Nigeria. Data was collected with the aid of a well-structured questionnaire administered to three identified projects in the North-east. Data collected were analysed using the severity index. Findings revealed political involvement, selection of inexperienced contractors and lack of coordinated public sector strategy as the most severe factors constraining the utilization of risk management strategies. The study recommended that: formulation of laws to prevent negative political meddling in construction projects; selection of experienced, risk-informed contractors; and comprehensive risk assessment and planning on all public construction projects.

Keywords: factors, Nigeria, north-east, public projects, risk management, strategies, utilization

Procedia PDF Downloads 488
17847 Deployed Confidence: The Testing in Production

Authors: Shreya Asthana

Abstract:

Testers know that the feature they tested on stage is working perfectly in production only after release went live. Sometimes something breaks in production and testers get to know through the end user’s bug raised. The panic mode starts when your staging test results do not reflect current production behavior. And you started doubting your testing skills when finally the user reported a bug to you. Testers can deploy their confidence on release day by testing on production. Once you start doing testing in production, you will see test result accuracy because it will be running on real time data and execution will be a little faster as compared to staging one due to elimination of bad data. Feature flagging, canary releases, and data cleanup can help to achieve this technique of testing. By this paper it will be easier to understand the steps to achieve production testing before making your feature live, and to modify IT company’s testing procedure, so testers can provide the bug free experience to the end users. This study is beneficial because too many people think that testing should be done in staging but not in production and now this is high time to pull out people from their old mindset of testing into a new testing world. At the end of the day, it all just matters if the features are working in production or not.

Keywords: bug free production, new testing mindset, testing strategy, testing approach

Procedia PDF Downloads 36
17846 Proactive Disk Defragmentation through User's File-Access Patterns

Authors: Gordon Wong

Abstract:

This paper shows how the task of disk defragmentation can be handled by modern operating systems in a transparent, automated, efficient, and confined way through user's file-access patterns. Since files tend to gradually fragment from time to time through file creation, deletion, growth, and shrinking, the problem gets even worse when a disk becomes so fragmented that file accesses cannot be made reasonably efficient without performing the operation of defragmentation for the "entire" disk, which is done manually by the user by launching the disk defragmentation utility program normally bundled with the operating system. In this paper, we argue that the disk defragmentation problem described can be solved without having to manually use the utility program to defragment the entire disk. The argument is based on the observation that system users tend to access certain files in a particular time interval like the way observed for programs exhibiting temporal locality of memory references during their execution. The task of disk defragmentation can be initiated and acted upon for those files contained in the current file-access locality detected and identified by the operating system. The paper also discusses how to use the locality of file references approach to quantitatively measure and determine the locality of user's file access patterns on which the task of disk defragmentation is based.

Keywords: operating systems, disk defragmentation, locality of file accesses, system performance

Procedia PDF Downloads 28
17845 Adaptive Multiple Transforms Hardware Architecture for Versatile Video Coding

Authors: T. Damak, S. Houidi, M. A. Ben Ayed, N. Masmoudi

Abstract:

The Versatile Video Coding standard (VVC) is actually under development by the Joint Video Exploration Team (or JVET). An Adaptive Multiple Transforms (AMT) approach was announced. It is based on different transform modules that provided an efficient coding. However, the AMT solution raises several issues especially regarding the complexity of the selected set of transforms. This can be an important issue, particularly for a future industrial adoption. This paper proposed an efficient hardware implementation of the most used transform in AMT approach: the DCT II. The developed circuit is adapted to different block sizes and can reach a minimum frequency of 192 MHz allowing an optimized execution time.

Keywords: adaptive multiple transforms, AMT, DCT II, hardware, transform, versatile video coding, VVC

Procedia PDF Downloads 121
17844 Application of Natural Language Processing in Education

Authors: Khaled M. Alhawiti

Abstract:

Reading capability is a major segment of language competency. On the other hand, discovering topical writings at a fitting level for outside and second language learners is a test for educators. We address this issue utilizing natural language preparing innovation to survey reading level and streamline content. In the connection of outside and second-language learning, existing measures of reading level are not appropriate to this errand. Related work has demonstrated the profit of utilizing measurable language preparing procedures; we expand these thoughts and incorporate other potential peculiarities to measure intelligibility. In the first piece of this examination, we join characteristics from measurable language models, customary reading level measures and other language preparing apparatuses to deliver a finer technique for recognizing reading level. We examine the execution of human annotators and assess results for our finders concerning human appraisals. A key commitment is that our identifiers are trainable; with preparing and test information from the same space, our finders beat more general reading level instruments (Flesch-Kincaid and Lexile). Trainability will permit execution to be tuned to address the needs of specific gatherings or understudies.

Keywords: natural language processing, trainability, syntactic simplification tools, education

Procedia PDF Downloads 461
17843 Four-dimensional (4D) Decoding Information Presented in Reports of Project Progress in Developing Countries

Authors: Vahid Khadjeh Anvary, Hamideh Karimi Yazdi

Abstract:

Generally, the tool of comparison between performance of each stage in the life of a project, is the number of project progress during that period, which in most cases is only determined as one-dimensional with referring to one of three factors (physical, time, and financial). In many projects in developing countries there are controversies on accuracy and the way of analyzing progress report of projects that hinders getting definitive and engineering conclusions on the status of project.Identifying weakness points of this kind of one-dimensional look on project and determining a reliable and engineering approach for multi-dimensional decoding information receivable from project is of great importance in project management.This can be a tool to help identification of hidden diseases of project before appearing irreversible symptoms that are usually delays or increased costs of execution. The method used in this paper is defining and evaluating a hypothetical project as an example analyzing different scenarios and numerical comparison of them along with related graphs and tables. Finally, by analyzing different possible scenarios in the project, possibility or impossibility of predicting their occurrence is examine through the evidence.

Keywords: physical progress, time progress, financial progress, delays, critical path

Procedia PDF Downloads 348
17842 Identification of Training Topics for the Improvement of the Relevant Cognitive Skills of Technical Operators in the Railway Domain

Authors: Giulio Nisoli, Jonas Brüngger, Karin Hostettler, Nicole Stoller, Katrin Fischer

Abstract:

Technical operators in the railway domain are experts responsible for the supervisory control of the railway power grid as well as of the railway tunnels. The technical systems used to master these demanding tasks are constantly increasing in their degree of automation. It becomes therefore difficult for technical operators to maintain the control over the technical systems and the processes of their job. In particular, the operators must have the necessary experience and knowledge in dealing with a malfunction situation or unexpected event. For this reason, it is of growing importance that the skills relevant for the execution of the job are maintained and further developed beyond the basic training they receive, where they are educated in respect of technical knowledge and the work with guidelines. Training methods aimed at improving the cognitive skills needed by technical operators are still missing and must be developed. Goals of the present study were to identify which are the relevant cognitive skills of technical operators in the railway domain and to define which topics should be addressed by the training of these skills. Observational interviews were conducted in order to identify the main tasks and the organization of the work of technical operators as well as the technical systems used for the execution of their job. Based on this analysis, the most demanding tasks of technical operators could be identified and described. The cognitive skills involved in the execution of these tasks are those, which need to be trained. In order to identify and analyze these cognitive skills a cognitive task analysis (CTA) was developed. CTA specifically aims at identifying the cognitive skills that employees implement when performing their own tasks. The identified cognitive skills of technical operators were summarized and grouped in training topics. For every training topic, specific goals were defined. The goals regard the three main categories; knowledge, skills and attitude to be trained in every training topic. Based on the results of this study, it is possible to develop specific training methods to train the relevant cognitive skills of the technical operators.

Keywords: cognitive skills, cognitive task analysis, technical operators in the railway domain, training topics

Procedia PDF Downloads 122
17841 Conception of Increasing the Efficiency of Excavation Shoring by Prestressing Diaphragm Walls

Authors: Mateusz Frydrych

Abstract:

The construction of diaphragm walls as excavation shoring as well as part of deep foundations is widely used in geotechnical engineering. Today's design challenges lie in the optimal dimensioning of the cross-section, which is demanded by technological considerations. Also in force is the issue of optimization and sustainable use of construction materials, including reduction of carbon footprint, which is currently a relevant challenge for the construction industry. The author presents the concept of an approach to achieving increased efficiency of diaphragm wall excavation shoring by using structural compression technology. The author proposes to implement prestressed tendons in a non-linear manner in the reinforcement cage. As a result bending moment is reduced, which translates into a reduction in the amount of steel needed in the section, a reduction in displacements, and a reduction in the scratching of the casing, including the achievement of better tightness. This task is rarely seen and has not yet been described in a scientific way in the literature. The author has developed a dynamic numerical model that allows the dimensioning of the cross-section of a prestressed shear wall, as well as the study of casing displacements and cross-sectional forces in any defined computational situation. Numerical software from the Sofistik - open source development environment - was used for the study, and models were validated in Plaxis software . This is an interesting idea that allows for optimizing the execution of construction works and reducing the required resources by using fewer materials and saving time. The author presents the possibilities of a prestressed diaphragm wall, among others, using. The example of a diaphragm wall working as a cantilever at the height of two underground floors without additional strutting or stability protection by using ground anchors. This makes the execution of the work more criminal for the contractor and, as a result, cheaper for the investor.

Keywords: prestressed diaphragm wall, Plaxis, Sofistik, innovation, FEM, optimisation

Procedia PDF Downloads 45
17840 Open Forging of Cylindrical Blanks Subjected to Lateral Instability

Authors: A. H. Elkholy, D. M. Almutairi

Abstract:

The successful and efficient execution of a forging process is dependent upon the correct analysis of loading and metal flow of blanks. This paper investigates the Upper Bound Technique (UBT) and its application in the analysis of open forging process when a possibility of blank bulging exists. The UBT is one of the energy rate minimization methods for the solution of metal forming process based on the upper bound theorem. In this regards, the kinematically admissible velocity field is obtained by minimizing the total forging energy rate. A computer program is developed in this research to implement the UBT. The significant advantages of this method is the speed of execution while maintaining a fairly high degree of accuracy and the wide prediction capability. The information from this analysis is useful for the design of forging processes and dies. Results for the prediction of forging loads and stresses, metal flow and surface profiles with the assured benefits in terms of press selection and blank preform design are outlined in some detail. The obtained predictions are ready for comparison with both laboratory and industrial results.

Keywords: forging, upper bound technique, metal forming, forging energy, forging die/platen

Procedia PDF Downloads 263
17839 Implementation of the Recursive Formula for Evaluation of the Strength of Daniels' Bundle

Authors: Vaclav Sadilek, Miroslav Vorechovsky

Abstract:

The paper deals with the classical fiber bundle model of equal load sharing, sometimes referred to as the Daniels' bundle or the democratic bundle. Daniels formulated a multidimensional integral and also a recursive formula for evaluation of the strength cumulative distribution function. This paper describes three algorithms for evaluation of the recursive formula and also their implementations with source codes in high-level programming language Python. A comparison of the algorithms are provided with respect to execution time. Analysis of orders of magnitudes of addends in the recursion is also provided.

Keywords: equal load sharing, mpmath, python, strength of Daniels' bundle

Procedia PDF Downloads 379
17838 Hybrid Control Mode Based on Multi-Sensor Information by Fuzzy Approach for Navigation Task of Autonomous Mobile Robot

Authors: Jonqlan Lin, C. Y. Tasi, K. H. Lin

Abstract:

This paper addresses the issue of the autonomous mobile robot (AMR) navigation task based on the hybrid control modes. The novel hybrid control mode, based on multi-sensors information by using the fuzzy approach, has been presented in this research. The system operates in real time, is robust, enables the robot to operate with imprecise knowledge, and takes into account the physical limitations of the environment in which the robot moves, obtaining satisfactory responses for a large number of different situations. An experiment is simulated and carried out with a pioneer mobile robot. From the experimental results, the effectiveness and usefulness of the proposed AMR obstacle avoidance and navigation scheme are confirmed. The experimental results show the feasibility, and the control system has improved the navigation accuracy. The implementation of the controller is robust, has a low execution time, and allows an easy design and tuning of the fuzzy knowledge base.

Keywords: autonomous mobile robot, obstacle avoidance, MEMS, hybrid control mode, navigation control

Procedia PDF Downloads 440
17837 Weighted Data Replication Strategy for Data Grid Considering Economic Approach

Authors: N. Mansouri, A. Asadi

Abstract:

Data Grid is a geographically distributed environment that deals with data intensive application in scientific and enterprise computing. Data replication is a common method used to achieve efficient and fault-tolerant data access in Grids. In this paper, a dynamic data replication strategy, called Enhanced Latest Access Largest Weight (ELALW) is proposed. This strategy is an enhanced version of Latest Access Largest Weight strategy. However, replication should be used wisely because the storage capacity of each Grid site is limited. Thus, it is important to design an effective strategy for the replication replacement task. ELALW replaces replicas based on the number of requests in future, the size of the replica, and the number of copies of the file. It also improves access latency by selecting the best replica when various sites hold replicas. The proposed replica selection selects the best replica location from among the many replicas based on response time that can be determined by considering the data transfer time, the storage access latency, the replica requests that waiting in the storage queue and the distance between nodes. Simulation results utilizing the OptorSim show our replication strategy achieve better performance overall than other strategies in terms of job execution time, effective network usage and storage resource usage.

Keywords: data grid, data replication, simulation, replica selection, replica placement

Procedia PDF Downloads 237
17836 Genetic Algorithm for In-Theatre Military Logistics Search-and-Delivery Path Planning

Authors: Jean Berger, Mohamed Barkaoui

Abstract:

Discrete search path planning in time-constrained uncertain environment relying upon imperfect sensors is known to be hard, and current problem-solving techniques proposed so far to compute near real-time efficient path plans are mainly bounded to provide a few move solutions. A new information-theoretic –based open-loop decision model explicitly incorporating false alarm sensor readings, to solve a single agent military logistics search-and-delivery path planning problem with anticipated feedback is presented. The decision model consists in minimizing expected entropy considering anticipated possible observation outcomes over a given time horizon. The model captures uncertainty associated with observation events for all possible scenarios. Entropy represents a measure of uncertainty about the searched target location. Feedback information resulting from possible sensor observations outcomes along the projected path plan is exploited to update anticipated unit target occupancy beliefs. For the first time, a compact belief update formulation is generalized to explicitly include false positive observation events that may occur during plan execution. A novel genetic algorithm is then proposed to efficiently solve search path planning, providing near-optimal solutions for practical realistic problem instances. Given the run-time performance of the algorithm, natural extension to a closed-loop environment to progressively integrate real visit outcomes on a rolling time horizon can be easily envisioned. Computational results show the value of the approach in comparison to alternate heuristics.

Keywords: search path planning, false alarm, search-and-delivery, entropy, genetic algorithm

Procedia PDF Downloads 335
17835 Identifying Project Delay Factors in the Australian Construction Industry

Authors: Syed Sohaib Bin Hasib, Hiyam Al-Kilidar

Abstract:

Meeting project deadlines is a major challenge for most construction projects. In this study, perceptions of contractors, clients, and consultants are compared relative to a list of factors derived from the review of the extant literature on project delay. 59 causes (categorized into 8 groups) of project delays were identified from the literature. A survey was devised to get insights and ranking of these factors from clients, consultants & contractors in the Australian construction industry. Findings showed that project delays in the Australian construction industry are mainly the result of skill shortages, interference in execution, and poor coordination and communication between the project stakeholders.

Keywords: construction, delay factors, time delay, australian construction industry

Procedia PDF Downloads 143
17834 Grid Computing for Multi-Objective Optimization Problems

Authors: Aouaouche Elmaouhab, Hassina Beggar

Abstract:

Solving multi-objective discrete optimization applications has always been limited by the resources of one machine: By computing power or by memory, most often both. To speed up the calculations, the grid computing represents a primary solution for the treatment of these applications through the parallelization of these resolution methods. In this work, we are interested in the study of some methods for solving multiple objective integer linear programming problem based on Branch-and-Bound and the study of grid computing technology. This study allowed us to propose an implementation of the method of Abbas and Al on the grid by reducing the execution time. To enhance our contribution, the main results are presented.

Keywords: multi-objective optimization, integer linear programming, grid computing, parallel computing

Procedia PDF Downloads 457
17833 Internal Product Management: The Key to Achieving Digital Maturity and Business Agility for Manufacturing IT Organizations

Authors: Frederick Johnson

Abstract:

Product management has a long and well-established history within the consumer goods industry, despite being one of the most obscure aspects of brand management. Many global manufacturing organizations are now opting for external cloud-based Manufacturing Execution Systems (MES) to replace costly and outdated monolithic MES solutions. Other global manufacturing leaders are restructuring their organizations to support human-centered values, agile methodologies, and fluid operating principles. Still, industry-leading organizations struggle to apply the appropriate framework for managing evolving external MES solutions as internal "digital products." Product management complements these current trends in technology and philosophical thinking in the market. This paper discusses the central problems associated with adopting product management processes by analyzing its traditional theories and characteristics. Considering these ideas, the article then constructs a translated internal digital product management framework by combining new and existing approaches and principles. The report concludes by demonstrating the framework's capabilities and potential effectiveness in achieving digital maturity and business agility within a manufacturing environment.

Keywords: internal product management, digital transformation, manufacturing information technology, manufacturing execution systems

Procedia PDF Downloads 109
17832 Comparative Analysis of Classical and Parallel Inpainting Algorithms Based on Affine Combinations of Projections on Convex Sets

Authors: Irina Maria Artinescu, Costin Radu Boldea, Eduard-Ionut Matei

Abstract:

The paper is a comparative study of two classical variants of parallel projection methods for solving the convex feasibility problem with their equivalents that involve variable weights in the construction of the solutions. We used a graphical representation of these methods for inpainting a convex area of an image in order to investigate their effectiveness in image reconstruction applications. We also presented a numerical analysis of the convergence of these four algorithms in terms of the average number of steps and execution time in classical CPU and, alternatively, in parallel GPU implementation.

Keywords: convex feasibility problem, convergence analysis, inpainting, parallel projection methods

Procedia PDF Downloads 141
17831 Innovation in Lean Thinking to Achieve Rapid Construction

Authors: Muhamad Azani Yahya, Vikneswaran Munikanan, Mohammed Alias Yusof

Abstract:

Lean thinking holds the potential for improving the construction sector, and therefore, it is a concept that should be adopted by construction sector players and academicians in the real industry. Bridging from that, a learning process for construction sector players regarding this matter should be the agenda in gaining the knowledge in preparation for their career. Lean principles offer opportunities for reducing lead times, eliminating non-value adding activities, reducing variability, and are facilitated by methods such as pull scheduling, simplified operations and buffer reduction. Thus, the drive for rapid construction, which is a systematic approach in enhancing efficiency to deliver a project using time reduction, while lean is the continuous process of eliminating waste, meeting or exceeding all customer requirements, focusing on the entire value stream and pursuing perfection in the execution of a constructed project. The methodology presented is shown to be valid through literature, interviews and questionnaire. The results show that the majority of construction sector players unfamiliar with lean thinking and they agreed that it can improve the construction process flow. With this background knowledge established and identified, best practices and recommended action are drawn.

Keywords: construction improvement, rapid construction, time reduction, lean construction

Procedia PDF Downloads 345
17830 A Randomized Comparative Evaluation of Efficacy of Ultrasound Guided Costoclavicular and Supraclavicular Approaches of Brachial Plexus Block for Upper Limb Surgeries

Authors: Anshul, Rajni Kalia, Sachin Kumar

Abstract:

Introduction: The costoclavicular approach, a modification to the infraclavicular approach, has been described for anesthesia for upper limb surgeries. Material And Methods: In this randomized and single-blind study, fourty patients undergoing emergency/elective upper limb surgery were allocated to two groups. Group C and S received ultrasound-guided Costoclavicular block and Supraclavicular block, respectively, with 20 ml 0.5 % ropivacaine with 8 mg dexamethasone under strict asepsis. The primary outcome assessed was the total duration of sensory and motor block in the postoperative period. Secondary outcomes were to compare the time taken to perform the procedure, block characteristics in terms of onset of motor and sensory blockade, the efficacy of analgesia with respect to the time of administration of the first rescue analgesic dose with both the blocks and note the side effects pertaining to either of the blocks. Results: The mean total duration of sensory and motor blockade was longer in group C vs. group S (p=0.002 and 0.024, respectively). The mean duration to perform a block in group S was more than in group C (p=0.012). The mean onset of sensory and motor Blockade Time in group S was more than in group C (p<0.001 and <0.001, respectively). The mean duration to perform a block in group S was more than in group C (p=0.012). Conclusion: The costoclavicular approach is better than supraclavicular in terms of rapid execution, faster onset of sensory-motor blockade, prolonged postoperative analgesia and similar PONV and safety profile.

Keywords: costoclavicular, supraclavicular, ropivacaine, dexamethasone

Procedia PDF Downloads 43
17829 The Parallelization of Algorithm Based on Partition Principle for Association Rules Discovery

Authors: Khadidja Belbachir, Hafida Belbachir

Abstract:

subsequently the expansion of the physical supports storage and the needs ceaseless to accumulate several data, the sequential algorithms of associations’ rules research proved to be ineffective. Thus the introduction of the new parallel versions is imperative. We propose in this paper, a parallel version of a sequential algorithm “Partition”. This last is fundamentally different from the other sequential algorithms, because it scans the data base only twice to generate the significant association rules. By consequence, the parallel approach does not require much communication between the sites. The proposed approach was implemented for an experimental study. The obtained results, shows a great reduction in execution time compared to the sequential version and Count Distributed algorithm.

Keywords: association rules, distributed data mining, partition, parallel algorithms

Procedia PDF Downloads 376
17828 Moving Object Detection Using Histogram of Uniformly Oriented Gradient

Authors: Wei-Jong Yang, Yu-Siang Su, Pau-Choo Chung, Jar-Ferr Yang

Abstract:

Moving object detection (MOD) is an important issue in advanced driver assistance systems (ADAS). There are two important moving objects, pedestrians and scooters in ADAS. In real-world systems, there exist two important challenges for MOD, including the computational complexity and the detection accuracy. The histogram of oriented gradient (HOG) features can easily detect the edge of object without invariance to changes in illumination and shadowing. However, to reduce the execution time for real-time systems, the image size should be down sampled which would lead the outlier influence to increase. For this reason, we propose the histogram of uniformly-oriented gradient (HUG) features to get better accurate description of the contour of human body. In the testing phase, the support vector machine (SVM) with linear kernel function is involved. Experimental results show the correctness and effectiveness of the proposed method. With SVM classifiers, the real testing results show the proposed HUG features achieve better than classification performance than the HOG ones.

Keywords: moving object detection, histogram of oriented gradient, histogram of uniformly-oriented gradient, linear support vector machine

Procedia PDF Downloads 566
17827 Manual to Automated Testing: An Effort-Based Approach for Determining the Priority of Software Test Automation

Authors: Peter Sabev, Katalina Grigorova

Abstract:

Test automation allows performing difficult and time consuming manual software testing tasks efficiently, quickly and repeatedly. However, development and maintenance of automated tests is expensive, so it needs a proper prioritization what to automate first. This paper describes a simple yet efficient approach for such prioritization of test cases based on the effort needed for both manual execution and software test automation. The suggested approach is very flexible because it allows working with a variety of assessment methods, and adding or removing new candidates at any time. The theoretical ideas presented in this article have been successfully applied in real world situations in several software companies by the authors and their colleagues including testing of real estate websites, cryptographic and authentication solutions, OSGi-based middleware framework that has been applied in various systems for smart homes, connected cars, production plants, sensors, home appliances, car head units and engine control units (ECU), vending machines, medical devices, industry equipment and other devices that either contain or are connected to an embedded service gateway.

Keywords: automated testing, manual testing, test automation, software testing, test prioritization

Procedia PDF Downloads 310
17826 Maximum Initial Input Allowed to Iterative Learning Control Set-up Using Singular Values

Authors: Naser Alajmi, Ali Alobaidly, Mubarak Alhajri, Salem Salamah, Muhammad Alsubaie

Abstract:

Iterative Learning Control (ILC) known to be a controlling tool to overcome periodic disturbances for repetitive systems. This technique is required to let the error signal tends to zero as the number of operation increases. The learning process that lies within this context is strongly dependent on the initial input which if selected properly tends to let the learning process be more effective compared to the case where a system starts from blind. ILC uses previous recorded execution data to update the following execution/trial input such that a reference trajectory is followed to a high accuracy. Error convergence in ILC is generally highly dependent on the input applied to a plant for trial $1$, thus a good choice of initial starting input signal would make learning faster and as a consequence the error tends to zero faster as well. In the work presented within, an upper limit based on the Singular Values Principle (SV) is derived for the initial input signal applied at trial $1$ such that the system follow the reference in less number of trials without responding aggressively or exceeding the working envelope where a system is required to move within in a robot arm, for example. Simulation results presented illustrate the theory introduced within this paper.

Keywords: initial input, iterative learning control, maximum input, singular values

Procedia PDF Downloads 214
17825 A Fast GPS Satellites Signals Detection Algorithm Based on Simplified Fast Fourier Transform

Authors: Beldjilali Bilal, Benadda Belkacem, Kahlouche Salem

Abstract:

Due to the Doppler effect caused by the high velocity of satellite and in some case receivers, the frequency of the Global Positioning System (GPS) signals are transformed into a new ones. Several acquisition algorithms frequency of the Global Positioning System (GPS) signals are transformed can be used to estimate the new frequency and phase shifts values. Numerous algorithms are based on the frequencies domain calculation. Our developed algorithm is a new approach dedicated to the Global Positioning System signal acquisition based on the fast Fourier transform. Our proposed new algorithm is easier to implement and has fast execution time compared with elder ones.

Keywords: global positioning system, acquisition, FFT, GPS/L1, software receiver, weak signal

Procedia PDF Downloads 218
17824 Cloud Support for Scientific Workflow Execution: Prototyping Solutions for Remote Sensing Applications

Authors: Sofiane Bendoukha, Daniel Moldt, Hayat Bendoukha

Abstract:

Workflow concepts are essential for the development of remote sensing applications. They can help users to manage and process satellite data and execute scientific experiments on distributed resources. The objective of this paper is to introduce an approach for the specification and the execution of complex scientific workflows in Cloud-like environments. The approach strives to support scientists during the modeling, the deployment and the monitoring of their workflows. This work takes advantage from Petri nets and more pointedly the so-called reference nets formalism, which provides a robust modeling/implementation technique. RENEWGRASS is a tool that we implemented and integrated into the Petri nets editor and simulator RENEW. It provides an easy way to support not experienced scientists during the specification of their workflows. It allows both modeling and enactment of image processing workflows from the remote sensing domain. Our case study is related to the implementation of vegetation indecies. We have implemented the Normalized Differences Vegetation Index (NDVI) workflow. Additionally, we explore the integration possibilities of the Cloud technology as a supplementary layer for the deployment of the current implementation. For this purpose, we discuss migration patterns of data and applications and propose an architecture.

Keywords: cloud computing, scientific workflows, petri nets, RENEWGRASS

Procedia PDF Downloads 424
17823 An Authentic Algorithm for Ciphering and Deciphering Called Latin Djokovic

Authors: Diogen Babuc

Abstract:

The question that is a motivation of writing is how many devote themselves to discovering something in the world of science where much is discerned and revealed, but at the same time, much is unknown. Methods: The insightful elements of this algorithm are the ciphering and deciphering algorithms of Playfair, Caesar, and Vigenère. Only a few of their main properties are taken and modified, with the aim of forming a specific functionality of the algorithm called Latin Djokovic. Specifically, a string is entered as input data. A key k is given, with a random value between the values a and b = a+3. The obtained value is stored in a variable with the aim of being constant during the run of the algorithm. In correlation to the given key, the string is divided into several groups of substrings, and each substring has a length of k characters. The next step involves encoding each substring from the list of existing substrings. Encoding is performed using the basis of Caesar algorithm, i.e., shifting with k characters. However, that k is incremented by 1 when moving to the next substring in that list. When the value of k becomes greater than b+1, it’ll return to its initial value. The algorithm is executed, following the same procedure, until the last substring in the list is traversed. Results: Using this polyalphabetic method, ciphering and deciphering of strings are achieved. The algorithm also works for a 100-character string. The x character isn’t used when the number of characters in a substring is incompatible with the expected length. The algorithm is simple to implement, but it’s questionable if it works better than the other methods from the point of view of execution time and storage space.

Keywords: ciphering, deciphering, authentic, algorithm, polyalphabetic cipher, random key, methods comparison

Procedia PDF Downloads 78
17822 Hardware Implementation for the Contact Force Reconstruction in Tactile Sensor Arrays

Authors: María-Luisa Pinto-Salamanca, Wilson-Javier Pérez-Holguín

Abstract:

Reconstruction of contact forces is a fundamental technique for analyzing the properties of a touched object and is essential for regulating the grip force in slip control loops. This is based on the processing of the distribution, intensity, and direction of the forces during the capture of the sensors. Currently, efficient hardware alternatives have been used more frequently in different fields of application, allowing the implementation of computationally complex algorithms, as is the case with tactile signal processing. The use of hardware for smart tactile sensing systems is a research area that promises to improve the processing time and portability requirements of applications such as artificial skin and robotics, among others. The literature review shows that hardware implementations are present today in almost all stages of smart tactile detection systems except in the force reconstruction process, a stage in which they have been less applied. This work presents a hardware implementation of a model-driven reported in the literature for the contact force reconstruction of flat and rigid tactile sensor arrays from normal stress data. From the analysis of a software implementation of such a model, this implementation proposes the parallelization of tasks that facilitate the execution of matrix operations and a two-dimensional optimization function to obtain a vector force by each taxel in the array. This work seeks to take advantage of the parallel hardware characteristics of Field Programmable Gate Arrays, FPGAs, and the possibility of applying appropriate techniques for algorithms parallelization using as a guide the rules of generalization, efficiency, and scalability in the tactile decoding process and considering the low latency, low power consumption, and real-time execution as the main parameters of design. The results show a maximum estimation error of 32% in the tangential forces and 22% in the normal forces with respect to the simulation by the Finite Element Modeling (FEM) technique of Hertzian and non-Hertzian contact events, over sensor arrays of 10×10 taxels of different sizes. The hardware implementation was carried out on an MPSoC XCZU9EG-2FFVB1156 platform of Xilinx® that allows the reconstruction of force vectors following a scalable approach, from the information captured by means of tactile sensor arrays composed of up to 48 × 48 taxels that use various transduction technologies. The proposed implementation demonstrates a reduction in estimation time of x / 180 compared to software implementations. Despite the relatively high values of the estimation errors, the information provided by this implementation on the tangential and normal tractions and the triaxial reconstruction of forces allows to adequately reconstruct the tactile properties of the touched object, which are similar to those obtained in the software implementation and in the two FEM simulations taken as reference. Although errors could be reduced, the proposed implementation is useful for decoding contact forces for portable tactile sensing systems, thus helping to expand electronic skin applications in robotic and biomedical contexts.

Keywords: contact forces reconstruction, forces estimation, tactile sensor array, hardware implementation

Procedia PDF Downloads 160