Search results for: recursive least square algorithm
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 5076

Search results for: recursive least square algorithm

1386 Key Frame Based Video Summarization via Dependency Optimization

Authors: Janya Sainui

Abstract:

As a rapid growth of digital videos and data communications, video summarization that provides a shorter version of the video for fast video browsing and retrieval is necessary. Key frame extraction is one of the mechanisms to generate video summary. In general, the extracted key frames should both represent the entire video content and contain minimum redundancy. However, most of the existing approaches heuristically select key frames; hence, the selected key frames may not be the most different frames and/or not cover the entire content of a video. In this paper, we propose a method of video summarization which provides the reasonable objective functions for selecting key frames. In particular, we apply a statistical dependency measure called quadratic mutual informaion as our objective functions for maximizing the coverage of the entire video content as well as minimizing the redundancy among selected key frames. The proposed key frame extraction algorithm finds key frames as an optimization problem. Through experiments, we demonstrate the success of the proposed video summarization approach that produces video summary with better coverage of the entire video content while less redundancy among key frames comparing to the state-of-the-art approaches.

Keywords: video summarization, key frame extraction, dependency measure, quadratic mutual information

Procedia PDF Downloads 258
1385 Clinical Comparative Study Comparing Efficacy of Intrathecal Fentanyl and Magnesium as an Adjuvant to Hyperbaric Bupivacaine in Mild Pre-Eclamptic Patients Undergoing Caesarean Section

Authors: Sanchita B. Sarma, M. P. Nath

Abstract:

Adequate analgesia following caesarean section decreases morbidity, hastens ambulation, improves patient outcome and facilitates care of the newborn. Intrathecal magnesium, an NMDA antagonist, has been shown to prolong analgesia without significant side effects in healthy parturients. The aim of this study was to evaluate the onset and duration of sensory and motor block, hemodynamic effect, postoperative analgesia, and adverse effects of magnesium or fentanyl given intrathecally with hyperbaric 0.5% bupivacaine in patients with mild preeclampsia undergoing caesarean section. Sixty women with mild preeclampsia undergoing elective caesarean section were included in a prospective, double blind, controlled trial. Patients were randomly assigned to receive spinal anesthesia with 2 mL 0.5% hyperbaric bupivacaine with 12.5 µg fentanyl (group F) or 0.1 ml of 50% magnesium sulphate (50 mg) (group M) with 0.15ml preservative free distilled water. Onset, duration and recovery of sensory and motor block, time to maximum sensory block, duration of spinal anaesthesia and postoperative analgesic requirements were studied. Statistical comparison was carried out using the Chi-square or Fisher’s exact tests and Independent Student’s t-test where appropriate. The onset of both sensory and motor block was slower in the magnesium group. The duration of spinal anaesthesia (246 vs. 284) and motor block (186.3 vs. 210) were significantly longer in the magnesium group. Total analgesic top up requirement was less in group M. Hemodynamic parameters were similar in both the groups. Intrathecal magnesium caused minimal side effects. Since Fentanyl and other opioid congeners are not available throughout the country easily, magnesium with its easy availability and less side effect profile can be a cost effective alternative to fentanyl in managing pregnancy induced hypertension (PIH) patients given along with Bupivacaine intrathecally in caesarean section.

Keywords: analgesia, magnesium, pre eclampsia, spinal anaesthesia

Procedia PDF Downloads 313
1384 Cost Effective Real-Time Image Processing Based Optical Mark Reader

Authors: Amit Kumar, Himanshu Singal, Arnav Bhavsar

Abstract:

In this modern era of automation, most of the academic exams and competitive exams are Multiple Choice Questions (MCQ). The responses of these MCQ based exams are recorded in the Optical Mark Reader (OMR) sheet. Evaluation of the OMR sheet requires separate specialized machines for scanning and marking. The sheets used by these machines are special and costs more than a normal sheet. Available process is non-economical and dependent on paper thickness, scanning quality, paper orientation, special hardware and customized software. This study tries to tackle the problem of evaluating the OMR sheet without any special hardware and making the whole process economical. We propose an image processing based algorithm which can be used to read and evaluate the scanned OMR sheets with no special hardware required. It will eliminate the use of special OMR sheet. Responses recorded in normal sheet is enough for evaluation. The proposed system takes care of color, brightness, rotation, little imperfections in the OMR sheet images.

Keywords: OMR, image processing, hough circle trans-form, interpolation, detection, binary thresholding

Procedia PDF Downloads 159
1383 Absolute Lymphocyte Count as Predictor of Pneumocystis Pneumonia in Patients With Unknown HIV Status at a Private Tertiary Hospital

Authors: Marja A. Bernardo, Coreena A. Bueser, Cybele Lara R. Abad, Raul V. Destura

Abstract:

Pneumocystis jirovecii pneumonia (PCP) is the most common opportunistic infection among people with HIV. Early consideration of PCP should be made even in patients whose HIV status is unknown as delay in treatment may be fatal. The use of absolute lymphocyte count (ALC) has been suggested as an alternative predictor of PCP especially in resource limited settings where PCR testing is costly or delayed. Objective: To determine whether the absolute lymphocyte count (ALC) can be used as a screening tool to predict Pneumocystis pneumonia in patients with unknown HIV status admitted at a private tertiary hospital. Methods: A retrospective cross-sectional study was conducted at a private tertiary medical center. Inpatient medical records of patients aged 18 years old and above from January 2012 to May 2014, in whom a clinical diagnosis of Pneumocystis jirovecii pneumonia was made were reviewed for inclusion. Demographic data, clinical features, hospital course, PCP PCR and HIV results were recorded. Independent t-test and chi-square analysis was used to determine any statistical difference between PCP-positive and PCP-negative groups. Mann-Whitney U-test was used for comparison of hospital stay. Results: There were no statistically significant differences in baseline characteristics between PCP positive and negative groups. While both the percent lymphocyte count (0.14 ± 0.13 vs 0.21 ± 0.16) and ALC (1160 ± 528.67 vs 1493.70 ± 988.61) were lower for the PCP-positive group, only the percent lymphocyte count reached a statistically significant difference (p= 0.067 vs p= 0.042). Conclusion: A quick determination of the ALC may be useful as an additional parameter to help screen for and diagnose pneumocystis pneumonia. In our study, the ALC of patients with PCP appear to be lower than in patients without PCP. A low ALC (e.g. below 1200) may help with the decision regarding empiric treatment. However, it should be used in conjunction with the patient’s clinical presentation, as well as other diagnostic tests. Larger, prospective studies incorporating the ALC with other clinical predictors are necessary to optimally predict those who would benefit from empiric or expedited management for potential PCP.

Keywords: Pneumocystis carinii pneumonia, Absolute Lymphocyte Count, infection, PCP

Procedia PDF Downloads 346
1382 Kinematic Optimization of Energy Extraction Performances for Flapping Airfoil by Using Radial Basis Function Method and Genetic Algorithm

Authors: M. Maatar, M. Mekadem, M. Medale, B. Hadjed, B. Imine

Abstract:

In this paper, numerical simulations have been carried out to study the performances of a flapping wing used as an energy collector. Metamodeling and genetic algorithms are used to detect the optimal configuration, improving power coefficient and/or efficiency. Radial basis functions and genetic algorithms have been applied to solve this problem. Three optimization factors are controlled, namely dimensionless heave amplitude h₀, pitch amplitude θ₀ and flapping frequency f. ANSYS FLUENT software has been used to solve the principal equations at a Reynolds number of 1100, while the heave and pitch motion of a NACA0015 airfoil has been realized using a developed function (UDF). The results reveal an average power coefficient and efficiency of 0.78 and 0.338 with an inexpensive low-fidelity model and a total relative error of 4.1% versus the simulation. The performances of the simulated optimum RBF-NSGA-II have been improved by 1.2% compared with the validated model.

Keywords: numerical simulation, flapping wing, energy extraction, power coefficient, efficiency, RBF, NSGA-II

Procedia PDF Downloads 27
1381 Evaluate the Effect of Teaching Small Scale Bussiness and Entrepreneurship on Graduates Unemployment in Nigeria: A Case Study of Anambra and Enugu State, South East Nigeria

Authors: Erinma Chibuzo Nwandu

Abstract:

Graduates unemployment has risen astronomically in spite of the emphasis on teaching of small scale business and Entrepreneurship in schools. This study sets out to evaluate the effect of teaching small scale business and Entrepreneurship on graduates’ unemployment in Nigeria. This study adopted the survey research design. Thus the nature of data for this study is primary, sourced by the use of a questionnaire administered to a sample of two thousand and sixty-five (2065) respondents drawn from groups of graduates who are employed, unemployed and self-employed in South East Nigeria. Simple percentages, Chi-square and regression analysis were used to derive useful and meaningful information and test the hypotheses respectively. Findings from the study suggest that Nigeria graduates are ill prepared to embark on small-scale business and entrepreneurship after graduation, and that teaching of small scale business and entrepreneurship in Nigeria tertiary institutions is ineffective on graduate unemployment reduction. Findings also suggest that while a lot of graduates agreed that they have taken a class(s) on small scale or entrepreneurship, they received more theoretical teachings than practical, more so while teachings on small scale business or entrepreneurship motivated graduates to think of self-employment, most of them cannot do a good business plan and hence could not benefit from some kind of Government assisted program for small-scale business and bank loan for the sake of small scale business. Thus, so many graduates are not interested in small scale business or entrepreneurship development as a result of lack of startup capital. The study thus recommends that course content and teaching method of entrepreneurship education needs to be reviewed and re-structured to constitute more practical teachings than theoretical teachings. Also, graduates should be exposed to seminar /workshop for self-employment at least once every semester. There should be practical teaching and practice of developing a business plan that will be viable to attract government or private sponsorship as well for it to be viable to attract financing from financing institutions. Government should provide a fund such as venture capital financing arrangement to empower business startups in Nigeria by graduates’.

Keywords: entrepreneurship, small scale business, startup capital, unemployment

Procedia PDF Downloads 268
1380 Hierarchical Cluster Analysis of Raw Milk Samples Obtained from Organic and Conventional Dairy Farming in Autonomous Province of Vojvodina, Serbia

Authors: Lidija Jevrić, Denis Kučević, Sanja Podunavac-Kuzmanović, Strahinja Kovačević, Milica Karadžić

Abstract:

In the present study, the Hierarchical Cluster Analysis (HCA) was applied in order to determine the differences between the milk samples originating from a conventional dairy farm (CF) and an organic dairy farm (OF) in AP Vojvodina, Republic of Serbia. The clustering was based on the basis of the average values of saturated fatty acids (SFA) content and unsaturated fatty acids (UFA) content obtained for every season. Therefore, the HCA included the annual SFA and UFA content values. The clustering procedure was carried out on the basis of Euclidean distances and Single linkage algorithm. The obtained dendrograms indicated that the clustering of UFA in OF was much more uniform compared to clustering of UFA in CF. In OF, spring stands out from the other months of the year. The same case can be noticed for CF, where winter is separated from the other months. The results could be expected because the composition of fatty acids content is greatly influenced by the season and nutrition of dairy cows during the year.

Keywords: chemometrics, clustering, food engineering, milk quality

Procedia PDF Downloads 270
1379 Indexing and Incremental Approach Using Map Reduce Bipartite Graph (MRBG) for Mining Evolving Big Data

Authors: Adarsh Shroff

Abstract:

Big data is a collection of dataset so large and complex that it becomes difficult to process using data base management tools. To perform operations like search, analysis, visualization on big data by using data mining; which is the process of extraction of patterns or knowledge from large data set. In recent years, the data mining applications become stale and obsolete over time. Incremental processing is a promising approach to refreshing mining results. It utilizes previously saved states to avoid the expense of re-computation from scratch. This project uses i2MapReduce, an incremental processing extension to Map Reduce, the most widely used framework for mining big data. I2MapReduce performs key-value pair level incremental processing rather than task level re-computation, supports not only one-step computation but also more sophisticated iterative computation, which is widely used in data mining applications, and incorporates a set of novel techniques to reduce I/O overhead for accessing preserved fine-grain computation states. To optimize the mining results, evaluate i2MapReduce using a one-step algorithm and three iterative algorithms with diverse computation characteristics for efficient mining.

Keywords: big data, map reduce, incremental processing, iterative computation

Procedia PDF Downloads 339
1378 Artificial Neural Network-Based Short-Term Load Forecasting for Mymensingh Area of Bangladesh

Authors: S. M. Anowarul Haque, Md. Asiful Islam

Abstract:

Electrical load forecasting is considered to be one of the most indispensable parts of a modern-day electrical power system. To ensure a reliable and efficient supply of electric energy, special emphasis should have been put on the predictive feature of electricity supply. Artificial Neural Network-based approaches have emerged to be a significant area of interest for electric load forecasting research. This paper proposed an Artificial Neural Network model based on the particle swarm optimization algorithm for improved electric load forecasting for Mymensingh, Bangladesh. The forecasting model is developed and simulated on the MATLAB environment with a large number of training datasets. The model is trained based on eight input parameters including historical load and weather data. The predicted load data are then compared with an available dataset for validation. The proposed neural network model is proved to be more reliable in terms of day-wise load forecasting for Mymensingh, Bangladesh.

Keywords: load forecasting, artificial neural network, particle swarm optimization

Procedia PDF Downloads 164
1377 Perception of Customers towards Service Quality: A Comparative Analysis of Organized and Unorganised Retail Stores (with Special Reference to Bhopal City)

Authors: Abdul Rashid, Varsha Rokade

Abstract:

Service Quality within retail units is pivotal for satisfying customers and retaining them. This study on customer perception towards Service Quality variables in Retail aims to identify the dimensions and their impact on customers. An analytical study of the different retail service quality variables was done to understand the relationship between them. The study tries exploring the factors that attract the customers towards the organised and unorganised retail stores in the capital city of Madhya Pradesh, India. As organised retailers are seen as offering similar products in the outlets, improving service quality is seen as critical to ensuring competitive advantage over unorganised retailers. Data were collected through a structured questionnaire on a five-point Likert scale from existing walk-in customers of selected organised and unorganised retail stores in Bhopal City of Madhya Pradesh, India. The data was then analysed by factor analysis using (SPSS) Statistical Package for the Social Sciences especially Percentage analysis, ANOVA and Chi-Square. This study tries to find interrelationship between various Retail Service Quality dimensions, which will help the retailers to identify the steps needed to improve the overall quality of service. Thus, the findings of the study prove to be helpful in understanding the service quality variables which should be considered by organised and unorganised retail stores in Capital city of Madhya Pradesh, India.Also, findings of this empirical research reiterate the point of view that dimensions of Service Quality in Retail play an important role in enhancing customer satisfaction – a sector with high growth potential and tremendous opportunities in rapidly growing economies like India’s. With the introduction of FDI in multi-brand retailing, a large number of international retail players are expected to enter the Indian market, this intern will bring more competition in the retail sector. For benchmarking themselves with global standards, the Indian retailers will have to improve their service quality.

Keywords: organized retail, unorganised retail, retail service quality, service quality dimension

Procedia PDF Downloads 223
1376 An Alternative Method for Computing Clothoids

Authors: Gerardo Casal, Miguel E. Vázquez-Méndez

Abstract:

The clothoid (also known as Cornu spiral or Euler spiral) is a curve that is characterized because its curvature is proportional to its length. This property makes that it would be widely used as transition curve for designing the layout of roads and railway tracks. In this work, from the geometrical property characterizing the clothoid, its parametric equations are obtained and two algorithms to compute it are compared. The first (classical), is widely used in Surveying Schools and it is based on the use of explicit formulas obtained from Taylor expansions of sine and cosine functions. The second one (alternative) is a very simple algorithm, based on the numerical solution of the initial value problems giving the clothoid parameterization. Both methods are compared in some typical surveying problems. The alternative method does not use complex formulas and so it is conceptually very simple and easy to apply. It gives good results, even if the classical method goes wrong (if the quotient between length and radius of curvature is high), needs no subsequent translations nor rotations and, consequently, it seems an efficient tool for designing the layout of roads and railway tracks.

Keywords: transition curves, railroad and highway engineering, Runge-Kutta methods

Procedia PDF Downloads 260
1375 Hierarchical Queue-Based Task Scheduling with CloudSim

Authors: Wanqing You, Kai Qian, Ying Qian

Abstract:

The concepts of Cloud Computing provide users with infrastructure, platform and software as service, which make those services more accessible for people via Internet. To better analysis the performance of Cloud Computing provisioning policies as well as resources allocation strategies, a toolkit named CloudSim proposed. With CloudSim, the Cloud Computing environment can be easily constructed by modelling and simulating cloud computing components, such as datacenter, host, and virtual machine. A good scheduling strategy is the key to achieve the load balancing among different machines as well as to improve the utilization of basic resources. Recently, the existing scheduling algorithms may work well in some presumptive cases in a single machine; however they are unable to make the best decision for the unforeseen future. In real world scenario, there would be numbers of tasks as well as several virtual machines working in parallel. Based on the concepts of multi-queue, this paper presents a new scheduling algorithm to schedule tasks with CloudSim by taking into account several parameters, the machines’ capacity, the priority of tasks and the history log.

Keywords: hierarchical queue, load balancing, CloudSim, information technology

Procedia PDF Downloads 412
1374 Data Collection with Bounded-Sized Messages in Wireless Sensor Networks

Authors: Min Kyung An

Abstract:

In this paper, we study the data collection problem in Wireless Sensor Networks (WSNs) adopting the two interference models: The graph model and the more realistic physical interference model known as Signal-to-Interference-Noise-Ratio (SINR). The main issue of the problem is to compute schedules with the minimum number of timeslots, that is, to compute the minimum latency schedules, such that data from every node can be collected without any collision or interference to a sink node. While existing works studied the problem with unit-sized and unbounded-sized message models, we investigate the problem with the bounded-sized message model, and introduce a constant factor approximation algorithm. To the best known of our knowledge, our result is the first result of the data collection problem with bounded-sized model in both interference models.

Keywords: data collection, collision-free, interference-free, physical interference model, SINR, approximation, bounded-sized message model, wireless sensor networks

Procedia PDF Downloads 211
1373 Robust Numerical Scheme for Pricing American Options under Jump Diffusion Models

Authors: Salah Alrabeei, Mohammad Yousuf

Abstract:

The goal of option pricing theory is to help the investors to manage their money, enhance returns and control their financial future by theoretically valuing their options. However, most of the option pricing models have no analytical solution. Furthermore, not all the numerical methods are efficient to solve these models because they have nonsmoothing payoffs or discontinuous derivatives at the exercise price. In this paper, we solve the American option under jump diffusion models by using efficient time-dependent numerical methods. several techniques are integrated to reduced the overcome the computational complexity. Fast Fourier Transform (FFT) algorithm is used as a matrix-vector multiplication solver, which reduces the complexity from O(M2) into O(M logM). Partial fraction decomposition technique is applied to rational approximation schemes to overcome the complexity of inverting polynomial of matrices. The proposed method is easy to implement on serial or parallel versions. Numerical results are presented to prove the accuracy and efficiency of the proposed method.

Keywords: integral differential equations, jump–diffusion model, American options, rational approximation

Procedia PDF Downloads 108
1372 Clinical and Structural Differences in Knee Osteoarthritis with/without Synovial Hypertrophy

Authors: Gi-Young Park, Dong Rak Kwon, Sung Cheol Cho

Abstract:

Objective: The synovium is known to be involved in many pathological characteristic processes. Also, synovitis is common in advanced osteoarthritis. We aimed to evaluate the clinical, radiographic, and ultrasound findings in patients with knee osteoarthritis and to compare the clinical and imaging findings between knee osteoarthritis with and without synovial hypertrophy confirmed by ultrasound. Methods: One hundred knees (54 left, 46 right) in 95 patients (64 women, 31 men; mean age, 65.9 years; range, 43-85 years) with knee osteoarthritis were recruited. The Visual Analogue Scale (VAS) was used to assess the intensity of knee pain. The severity of knee osteoarthritis was classified according to Kellgren and Lawrence's (K-L) grade on a radiograph. Ultrasound examination was performed by a physiatrist who had 24 years of experience in musculoskeletal ultrasound. Ultrasound findings, including the thickness of joint effusion in the suprapatellar pouch, synovial hypertrophy, infrapatellar tendinosis, meniscal tear or extrusion, and Baker cyst, were measured and detected. The thickness of knee joint effusion was measured at the maximal anterior-posterior diameter of fluid collection in the suprapatellar pouch. Synovial hypertrophy was identified as the soft tissue of variable echogenicity, which is poorly compressible and nondisplaceable by compression of an ultrasound transducer. The knees were divided into two groups according to the presence of synovial hypertrophy. The differences in clinical and imaging findings between the two groups were evaluated by independent t-test and chi-square test. Results: Synovial hypertrophy was detected in 48 knees of 100 knees on ultrasound. There were no significant differences in demographic parameters and VAS score except in sex between the two groups (P<0.05). Medial meniscal extrusion and tear were significantly more frequent in knees with synovial hypertrophy than those in knees without synovial hypertrophy. K-L grade and joint effusion thickness were greater in patients with synovial hypertrophy than those in patients without synovial hypertrophy (P<0.05). Conclusion: Synovial hypertrophy in knee osteoarthritis was associated with greater suprapatellar joint effusion and higher K-L grade and maybe a characteristic ultrasound feature of late knee osteoarthritis. These results suggest that synovial hypertrophy on ultrasound can be regarded as a predictor of rapid progression in patients with knee osteoarthritis.

Keywords: knee osteoarthritis, synovial hypertrophy, ultrasound, K-L grade

Procedia PDF Downloads 68
1371 Clustering and Modelling Electricity Conductors from 3D Point Clouds in Complex Real-World Environments

Authors: Rahul Paul, Peter Mctaggart, Luke Skinner

Abstract:

Maintaining public safety and network reliability are the core objectives of all electricity distributors globally. For many electricity distributors, managing vegetation clearances from their above ground assets (poles and conductors) is the most important and costly risk mitigation control employed to meet these objectives. Light Detection And Ranging (LiDAR) is widely used by utilities as a cost-effective method to inspect their spatially-distributed assets at scale, often captured using high powered LiDAR scanners attached to fixed wing or rotary aircraft. The resulting 3D point cloud model is used by these utilities to perform engineering grade measurements that guide the prioritisation of vegetation cutting programs. Advances in computer vision and machine-learning approaches are increasingly applied to increase automation and reduce inspection costs and time; however, real-world LiDAR capture variables (e.g., aircraft speed and height) create complexity, noise, and missing data, reducing the effectiveness of these approaches. This paper proposes a method for identifying each conductor from LiDAR data via clustering methods that can precisely reconstruct conductors in complex real-world configurations in the presence of high levels of noise. It proposes 3D catenary models for individual clusters fitted to the captured LiDAR data points using a least square method. An iterative learning process is used to identify potential conductor models between pole pairs. The proposed method identifies the optimum parameters of the catenary function and then fits the LiDAR points to reconstruct the conductors.

Keywords: point cloud, LİDAR data, machine learning, computer vision, catenary curve, vegetation management, utility industry

Procedia PDF Downloads 90
1370 [Keynote]: No-Trust-Zone Architecture for Securing Supervisory Control and Data Acquisition

Authors: Michael Okeke, Andrew Blyth

Abstract:

Supervisory Control And Data Acquisition (SCADA) as the state of the art Industrial Control Systems (ICS) are used in many different critical infrastructures, from smart home to energy systems and from locomotives train system to planes. Security of SCADA systems is vital since many lives depend on it for daily activities and deviation from normal operation could be disastrous to the environment as well as lives. This paper describes how No-Trust-Zone (NTZ) architecture could be incorporated into SCADA Systems in order to reduce the chances of malicious intent. The architecture is made up of two distinctive parts which are; the field devices such as; sensors, PLCs pumps, and actuators. The second part of the architecture is designed following lambda architecture, which is made up of a detection algorithm based on Particle Swarm Optimization (PSO) and Hadoop framework for data processing and storage. Apache Spark will be a part of the lambda architecture for real-time analysis of packets for anomalies detection.

Keywords: industrial control system (ics, no-trust-zone (ntz), particle swarm optimisation (pso), supervisory control and data acquisition (scada), swarm intelligence (SI)

Procedia PDF Downloads 330
1369 Autonomous Strategic Aircraft Deconfliction in a Multi-Vehicle Low Altitude Urban Environment

Authors: Loyd R. Hook, Maryam Moharek

Abstract:

With the envisioned future growth of low altitude urban aircraft operations for airborne delivery service and advanced air mobility, strategies to coordinate and deconflict aircraft flight paths must be prioritized. Autonomous coordination and planning of flight trajectories is the preferred approach to the future vision in order to increase safety, density, and efficiency over manual methods employed today. Difficulties arise because any conflict resolution must be constrained by all other aircraft, all airspace restrictions, and all ground-based obstacles in the vicinity. These considerations make pair-wise tactical deconfliction difficult at best and unlikely to find a suitable solution for the entire system of vehicles. In addition, more traditional methods which rely on long time scales and large protected zones will artificially limit vehicle density and drastically decrease efficiency. Instead, strategic planning, which is able to respond to highly dynamic conditions and still account for high density operations, will be required to coordinate multiple vehicles in the highly constrained low altitude urban environment. This paper develops and evaluates such a planning algorithm which can be implemented autonomously across multiple aircraft and situations. Data from this evaluation provide promising results with simulations showing up to 10 aircraft deconflicted through a relatively narrow low-altitude urban canyon without any vehicle to vehicle or obstacle conflict. The algorithm achieves this level of coordination beginning with the assumption that each vehicle is controlled to follow an independently constructed flight path, which is itself free of obstacle conflict and restricted airspace. Then, by preferencing speed change deconfliction maneuvers constrained by the vehicles flight envelope, vehicles can remain as close to the original planned path and prevent cascading vehicle to vehicle conflicts. Performing the search for a set of commands which can simultaneously ensure separation for each pair-wise aircraft interaction and optimize the total velocities of all the aircraft is further complicated by the fact that each aircraft's flight plan could contain multiple segments. This means that relative velocities will change when any aircraft achieves a waypoint and changes course. Additionally, the timing of when that aircraft will achieve a waypoint (or, more directly, the order upon which all of the aircraft will achieve their respective waypoints) will change with the commanded speed. Put all together, the continuous relative velocity of each vehicle pair and the discretized change in relative velocity at waypoints resembles a hybrid reachability problem - a form of control reachability. This paper proposes two methods for finding solutions to these multi-body problems. First, an analytical formulation of the continuous problem is developed with an exhaustive search of the combined state space. However, because of computational complexity, this technique is only computable for pairwise interactions. For more complicated scenarios, including the proposed 10 vehicle example, a discretized search space is used, and a depth-first search with early stopping is employed to find the first solution that solves the constraints.

Keywords: strategic planning, autonomous, aircraft, deconfliction

Procedia PDF Downloads 85
1368 Evaluation of the Efficacy of Basic Life Support Teaching in Second and Third Year Medical Students

Authors: Bianca W. O. Silva, Adriana C. M. Andrade, Gustavo C. M. Lucena, Virna M. S. Lima

Abstract:

Introduction: Basic life support (BLS) involves the immediate recognition of cardiopulmonary arrest. Each year, 359.400 and 275.000 individuals with cardiac arrest are attended in emergency departments in USA and Europe. Brazilian data shows that 200.000 cardiac arrests occur every year, and half of them out of the hospital. Medical schools around the world teach BLS in the first years of the course, but studies show that there is a decline of the knowledge as the years go by, affecting the chain of survival. The objective was to analyze the knowledge of medical students about BLS and the retention of this learning throughout the course. Methods: This study included 150 students who were at the second and third year of a medical school in Salvador, Bahia, Brazil. The instrument of data collection was a structured questionnaire composed of 20 questions based on the 2015 American Heart Association guideline. The Pearson Chi-square test was used in order to study the association between previous training, sex and semester with the degree of knowledge of the students. The Kruskal-Wallis test was used to evaluate the different yields obtained between the various semesters. The number of correct answers was described by average and quartiles. Results: Regarding the degree of knowledge, 19.6% of the female students reached the optimal classification, a better outcome than the achieved by the male participants. Of those with previous training, 33.33% were classified as good and optimal, none of the students reached the optimal classification and only 2.2% of them were classified as bad (those who did not have 52.6% of correct answers). The analysis of the degree of knowledge related to each semester revealed that the 5th semester had the highest outcome: 30.5%. However, the acquaintance presented by the semesters was generally unsatisfactory, since 50% of the students, or more, demonstrated knowledge levels classified as bad or regular. When confronting the different semesters and the achieved scores, the value of p was 0.831. Conclusion: It is important to focus on the training of medical professionals that are capable of facing emergency situations, improving the systematization of care, and thereby increasing the victims' possibility of survival.

Keywords: basic life support, cardiopulmonary ressucitacion, education, medical students

Procedia PDF Downloads 171
1367 Discrete Group Search Optimizer for the Travelling Salesman Problem

Authors: Raed Alnajjar, Mohd Zakree, Ahmad Nazri

Abstract:

In this study, we apply Discrete Group Search Optimizer (DGSO) for solving Traveling Salesman Problem (TSP). The DGSO is a nature inspired optimization algorithm that imitates the animal behavior, especially animal searching behavior. The proposed DGSO uses a vector representation and some discrete operators, such as destruction, construction, differential evolution, swap and insert. The TSP is a well-known hard combinatorial optimization problem, which seeks to find the shortest path among numbers of cities. The performance of the proposed DGSO is evaluated and tested on benchmark instances which listed in LIBTSP dataset. The experimental results show that the performance of the proposed DGSO is comparable with the other methods in the state of the art for some instances. The results show that DGSO outperform Ant Colony System (ACS) in some instances whilst outperform other metaheuristic in most instances. In addition to that, the new results obtained a number of optimal solutions and some best known results. DGSO was able to obtain feasible and good quality solution across all dataset.

Keywords: discrete group search optimizer (DGSO); Travelling salesman problem (TSP); Variable neighborhood search(VNS)

Procedia PDF Downloads 315
1366 Dynamic Web-Based 2D Medical Image Visualization and Processing Software

Authors: Abdelhalim. N. Mohammed, Mohammed. Y. Esmail

Abstract:

In the course of recent decades, medical imaging has been dominated by the use of costly film media for review and archival of medical investigation, however due to developments in networks technologies and common acceptance of a standard digital imaging and communication in medicine (DICOM) another approach in light of World Wide Web was produced. Web technologies successfully used in telemedicine applications, the combination of web technologies together with DICOM used to design a web-based and open source DICOM viewer. The Web server allowance to inquiry and recovery of images and the images viewed/manipulated inside a Web browser without need for any preinstalling software. The dynamic site page for medical images visualization and processing created by using JavaScript and HTML5 advancements. The XAMPP ‘apache server’ is used to create a local web server for testing and deployment of the dynamic site. The web-based viewer connected to multiples devices through local area network (LAN) to distribute the images inside healthcare facilities. The system offers a few focal points over ordinary picture archiving and communication systems (PACS): easy to introduce, maintain and independently platforms that allow images to display and manipulated efficiently, the system also user-friendly and easy to integrate with an existing system that have already been making use of web technologies. The wavelet-based image compression technique on which 2-D discrete wavelet transform used to decompose the image then wavelet coefficients are transmitted by entropy encoding after threshold to decrease transmission time, stockpiling cost and capacity. The performance of compression was estimated by using images quality metrics such as mean square error ‘MSE’, peak signal to noise ratio ‘PSNR’ and compression ratio ‘CR’ that achieved (83.86%) when ‘coif3’ wavelet filter is used.

Keywords: DICOM, discrete wavelet transform, PACS, HIS, LAN

Procedia PDF Downloads 151
1365 Role of Male Partners in Postpartum Family Planning

Authors: Stephen Rulisa, Aimee Nyiramahirwe

Abstract:

Background: Strategies to increase the uptake of contraception services have been adopted in Rwanda, but the unmet need for family planning remains high. Women in the postpartum period are at higher risk for unintended pregnancy due to the silent conversion from lactational amenorrhea to reactivation of ovulatory cycles. The purpose of this study was to explore the role of male partners in the uptake of postpartum contraception. Methods: A prospective cross-sectional study was conducted among women who delivered at the University Teaching Hospital of Kigali for a period of 3 months with random sampling. A questionnaire was used to collect socio-demographic and antenatal data, information on male companionship, and intent to use postpartum contraception at admission. Participants were contacted six weeks later to collect data on contraceptive use. The outcome variables were uptake of postpartum contraception and types of contraceptives taken (long-acting vs. short-acting), controlling for male companionship during the antenatal period. A Chi-square test was used and a p-value ≤0.05 was considered significant. Results: A total of 209 women were recruited with a mean age of 30.8±5.2 years. The majority (60.9%) were multigravida, and 66.5% were multiparous. More than half (55%) had male partner companionship, 18.3% had companionship for four antenatal visits, and 28.2% had education on contraception with their male partner. Factors significantly associated with uptake of postpartum contraception were: age above 30 years, owning or heading a business, multigravidity, multiparity, antenatal care at a health center or district hospital, cesarean delivery, and previous utilization of contraception. Male companionship significantly increased the intent to use contraception, uptake of modern contraception in general, and uptake of long active contraceptives but did not predict the uptake of short-acting contraceptives. Conclusions: Our study demonstrates a positive association between male companionship during antenatal care, labor and delivery with the uptake of postpartum family planning. Our study suggests more sensitization to involve the male partners, improving the education on contraception during antenatal care and further research to assess the sustained uptake of contraception beyond the postpartum period.

Keywords: postpartum, family planning, contraception, male partner, uptake

Procedia PDF Downloads 101
1364 A Stokes Optimal Control Model of Determining Cellular Interaction Forces during Gastrulation

Authors: Yuanhao Gao, Ping Lin, Kees Weijer

Abstract:

An optimal control system model is proposed for the cell flow in the process of chick embryo gastrulation in this paper. The target is to determine the cellular interaction forces which are hard to measure. This paper will take an approach to investigate the forces with the idea of the inverse problem. By choosing the forces as the control variable and regarding the cell flow as Stokes fluid, an objective functional will be established to match the numerical result of cell velocity with the experimental data. So that the forces could be determined by minimizing the objective functional. The Lagrange multiplier method is utilized to derive the state and adjoint equations consisting the optimal control system, which specifies the first-order necessary conditions. Finite element method is used to discretize and approximate equations. A conjugate gradient algorithm is given for solving the minimum solution of the system and determine the forces.

Keywords: optimal control model, Stokes equation, conjugate gradient method, finite element method, chick embryo gastrulation

Procedia PDF Downloads 249
1363 Inferring Human Mobility in India Using Machine Learning

Authors: Asra Yousuf, Ajaykumar Tannirkulum

Abstract:

Inferring rural-urban migration trends can help design effective policies that promote better urban planning and rural development. In this paper, we describe how machine learning algorithms can be applied to predict internal migration decisions of people. We consider data collected from household surveys in Tamil Nadu to train our model. To measure the performance of the model, we use data on past migration from National Sample Survey Organisation of India. The factors for training the model include socioeconomic characteristic of each individual like age, gender, place of residence, outstanding loans, strength of the household, etc. and his past migration history. We perform a comparative analysis of the performance of a number of machine learning algorithm to determine their prediction accuracy. Our results show that machine learning algorithms provide a stronger prediction accuracy as compared to statistical models. Our goal through this research is to propose the use of data science techniques in understanding human decisions and behaviour in developing countries.

Keywords: development, migration, internal migration, machine learning, prediction

Procedia PDF Downloads 259
1362 Computational Fluid Dynamicsfd Simulations of Air Pollutant Dispersion: Validation of Fire Dynamic Simulator Against the Cute Experiments of the Cost ES1006 Action

Authors: Virginie Hergault, Siham Chebbah, Bertrand Frere

Abstract:

Following in-house objectives, Central laboratory of Paris police Prefecture conducted a general review on models and Computational Fluid Dynamics (CFD) codes used to simulate pollutant dispersion in the atmosphere. Starting from that review and considering main features of Large Eddy Simulation, Central Laboratory Of Paris Police Prefecture (LCPP) postulates that the Fire Dynamics Simulator (FDS) model, from National Institute of Standards and Technology (NIST), should be well suited for air pollutant dispersion modeling. This paper focuses on the implementation and the evaluation of FDS in the frame of the European COST ES1006 Action. This action aimed at quantifying the performance of modeling approaches. In this paper, the CUTE dataset carried out in the city of Hamburg, and its mock-up has been used. We have performed a comparison of FDS results with wind tunnel measurements from CUTE trials on the one hand, and, on the other, with the models results involved in the COST Action. The most time-consuming part of creating input data for simulations is the transfer of obstacle geometry information to the format required by SDS. Thus, we have developed Python codes to convert automatically building and topographic data to the FDS input file. In order to evaluate the predictions of FDS with observations, statistical performance measures have been used. These metrics include the fractional bias (FB), the normalized mean square error (NMSE) and the fraction of predictions within a factor of two of observations (FAC2). As well as the CFD models tested in the COST Action, FDS results demonstrate a good agreement with measured concentrations. Furthermore, the metrics assessment indicate that FB and NMSE meet the tolerance acceptable.

Keywords: numerical simulations, atmospheric dispersion, cost ES1006 action, CFD model, cute experiments, wind tunnel data, numerical results

Procedia PDF Downloads 125
1361 Building Green Infrastructure Networks Based on Cadastral Parcels Using Network Analysis

Authors: Gon Park

Abstract:

Seoul in South Korea established the 2030 Seoul City Master Plan that contains green-link projects to connect critical green areas within the city. However, the plan does not have detailed analyses for green infrastructure to incorporate land-cover information to many structural classes. This study maps green infrastructure networks of Seoul for complementing their green plans with identifying and raking green areas. Hubs and links of main elements of green infrastructure have been identified from incorporating cadastral data of 967,502 parcels to 135 of land use maps using geographic information system. Network analyses were used to rank hubs and links of a green infrastructure map with applying a force-directed algorithm, weighted values, and binary relationships that has metrics of density, distance, and centrality. The results indicate that network analyses using cadastral parcel data can be used as the framework to identify and rank hubs, links, and networks for the green infrastructure planning under a variable scenarios of green areas in cities.

Keywords: cadastral data, green Infrastructure, network analysis, parcel data

Procedia PDF Downloads 190
1360 San Francisco Public Utilities Commission Headquarters "The Greenest Urban Building in the United States"

Authors: Charu Sharma

Abstract:

San Francisco Public Utilities Commission’s Headquarters was listed in the 2013-American Institute of Architects Committee of the Environment (AIA COTE) Top Ten Green Projects. This 13-story, 277,000-square-foot building, housing more than 900 of the agency’s employees was completed in June 2012. It was designed to achieve LEED Platinum Certification and boasts a plethora of green features to significantly reduce the use of energy and water consumption, and provide a healthy office work environment with high interior air quality and natural daylight. Key sustainability features include on-site clean energy generation through renewable photovoltaic and wind sources providing $118 million in energy cost savings over 75 years; 45 percent daylight harvesting; and the consumption of 55 percent less energy and a 32 percent less electricity demand from the main power grid. It uses 60 percent less water usage than an average 13-story office building as most of that water will be recycled for non-potable uses at the site, running through a system of underground tanks and artificial wetlands that cleans and clarifies whatever is flushed down toilets or washed down drains. This is one of the first buildings in the nation with treatment of gray and black water. The building utilizes an innovative structural system with post tensioned cores that will provide the highest asset preservation for the building. In addition, the building uses a “green” concrete mixture that releases less carbon gases. As a public utility commission this building has set a good example for resource conservation-the building is expected to be cheaper to operate and maintain as time goes on and will have saved rate-payers $500 million in energy and water savings. Within the anticipated 100-year lifespan of the building, our ratepayers will save approximately $3.7 billion through the combination of rental savings, energy efficiencies, and asset ownership.

Keywords: energy efficiency, sustainability, resource conservation, asset ownership, rental savings

Procedia PDF Downloads 429
1359 Routing and Energy Efficiency through Data Coupled Clustering in Large Scale Wireless Sensor Networks (WSNs)

Authors: Jainendra Singh, Zaheeruddin

Abstract:

A typical wireless sensor networks (WSNs) consists of several tiny and low-power sensors which use radio frequency to perform distributed sensing tasks. The longevity of wireless sensor networks (WSNs) is a major issue that impacts the application of such networks. While routing protocols are striving to save energy by acting on sensor nodes, recent studies show that network lifetime can be enhanced by further involving sink mobility. A common approach for energy efficiency is partitioning the network into clusters with correlated data, where the representative nodes simply transmit or average measurements inside the cluster. In this paper, we propose an energy- efficient homogenous clustering (EHC) technique. In this technique, the decision of each sensor is based on their residual energy and an estimate of how many of its neighboring cluster heads (CHs) will benefit from it being a CH. We, also explore the routing algorithm in clustered WSNs. We show that the proposed schemes significantly outperform current approaches in terms of packet delay, hop count and energy consumption of WSNs.

Keywords: wireless sensor network, energy efficiency, clustering, routing

Procedia PDF Downloads 251
1358 Spatiotemporal Neural Network for Video-Based Pose Estimation

Authors: Bin Ji, Kai Xu, Shunyu Yao, Jingjing Liu, Ye Pan

Abstract:

Human pose estimation is a popular research area in computer vision for its important application in human-machine interface. In recent years, 2D human pose estimation based on convolution neural network has got great progress and development. However, in more and more practical applications, people often need to deal with tasks based on video. It’s not far-fetched for us to consider how to combine the spatial and temporal information together to achieve a balance between computing cost and accuracy. To address this issue, this study proposes a new spatiotemporal model, namely Spatiotemporal Net (STNet) to combine both temporal and spatial information more rationally. As a result, the predicted keypoints heatmap is potentially more accurate and spatially more precise. Under the condition of ensuring the recognition accuracy, the algorithm deal with spatiotemporal series in a decoupled way, which greatly reduces the computation of the model, thus reducing the resource consumption. This study demonstrate the effectiveness of our network over the Penn Action Dataset, and the results indicate superior performance of our network over the existing methods.

Keywords: convolutional long short-term memory, deep learning, human pose estimation, spatiotemporal series

Procedia PDF Downloads 137
1357 Optimization of Shear Frame Structures Applying Various Forms of Wavelet Transforms

Authors: Seyed Sadegh Naseralavi, Sohrab Nemati, Ehsan Khojastehfar, Sadegh Balaghi

Abstract:

In the present research, various formulations of wavelet transform are applied on acceleration time history of earthquake. The mentioned transforms decompose the strong ground motion into low and high frequency parts. Since the high frequency portion of strong ground motion has a minor effect on dynamic response of structures, the structure is excited by low frequency part. Consequently, the seismic response of structure is predicted consuming one half of computational time, comparing with conventional time history analysis. Towards reducing the computational effort needed in seismic optimization of structure, seismic optimization of a shear frame structure is conducted by applying various forms of mentioned transformation through genetic algorithm.

Keywords: time history analysis, wavelet transform, optimization, earthquake

Procedia PDF Downloads 222