Search results for: software defined networking (SDN)
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 7636

Search results for: software defined networking (SDN)

6886 Television Sports Exposure and Rape Myth Acceptance: The Mediating Role of Sexual Objectification of Women

Authors: Sofia Mariani, Irene Leo

Abstract:

The objective of the present study is to define the mediating role of attitudes that objectify and devalue women (hostile sexism, benevolent sexism, and sexual objectification of women) in the indirect correlation between exposure to televised sports and acceptance of rape myths. A second goal is to contribute to research on the topic by defining the role of mediators in exposure to different types of sports, following the traditional gender classification of sports. Data collection was carried out by means of an online questionnaire, measuring television sport exposure, sport type, hostile sexism, benevolent sexism, and sexual objectification of women. Data analysis was carried out using IBM SPSS software. The model used was created using Ordinary Least Squares (OLS) regression path analysis. The predictor variable in the model was television sports exposure, the outcome was rape myths acceptance, and the mediators were (1) hostile sexism, (2) benevolent sexism, and (3) sexual objectification of women. Correlation analyses were carried out dividing by sport type and controlling for the participants’ gender. As seen in existing literature, television sports exposure was found to be indirectly and positively related to rape myth acceptance through the mediating role of: (1) hostile sexism, (2) benevolent sexism, and (3) sexual objectification of women. The type of sport watched influenced the role of the mediators: hostile sexism was found to be the common mediator to all sports type, exposure to traditionally considered feminine or neutral sports showed the additional mediation effect of sexual objectification of women. In line with existing literature, controlling for gender showed that the only significant mediators were hostile sexism for male participants and benevolent sexism for female participants. Given the prevalence of men among the viewers of traditionally considered masculine sports, the correlation between television sports exposure and rape myth acceptance through the mediation of hostile sexism is likely due to the gender of the participants. However, this does not apply to the viewers of traditionally considered feminine and neutral sports, as this group is balanced in terms of gender and shows a unique mediation: the correlation between television sports exposure and rape myth acceptance is mediated by both hostile sexism and sexual objectification. Given that hostile sexism is defined as hostility towards women who oppose or fail to conform to traditional gender roles, these findings confirm that sport is perceived as a non-traditional activity for women. Additionally, these results imply that the portrayal of women in traditionally considered feminine and neutral sports - which are defined as such because of their aesthetic characteristics - may have a strong component of sexual objectification of women. The present research contributes to defining the association between sports exposure and rape myth acceptance through the mediation effects of sexist attitudes and sexual objectification of women. The results of this study have practical implications, such as supporting the feminine sports teams who ask for more practical and less revealing uniforms, more similar to their male colleagues and therefore less objectifying.

Keywords: television exposure, sport, rape myths, objectification, sexism

Procedia PDF Downloads 81
6885 A Reading Attempt of the Urban Memory of Jordan University of Science and Technology Campus by Cognitive Mapping

Authors: Bsma Adel Bany Mohammad

Abstract:

The University campuses are a small city containing basic city functions such as educational spaces, accommodations, services and transportation. They are spaces of functional and social life with different activities, different occupants. The campus designed and transformed like cities so both experienced and memorized in same way. Campus memory is the ability of individuals to maintain and reveal the spatial components of designed physical spaces, which form the understandings, experiences, sensations of the environment in all. ‘Cognitive mapping’ is used to decode the physical interaction and emotional relationship between individuals and the city; Cognitive maps are created graphically using geometric and verbal elements on paper by remembering the images of the Urban Environment. In this study, to determine the emotional urban identity belonging to Jordan University of science and technology Campus, architecture students Asked to identify the areas they interact with in the campus by drawing a cognitive map. ‘Campus memory items’ are identified by analyzing the cognitive maps of the campus, then the spatial identity result of such data. The analysis based on the five basic elements of Lynch: paths, districts, edges, nodes, and landmarks. As a result of this analysis, it found that Spatial Identity constructed by the shared elements of the maps. The memory of most students listed the gates structure- which is a large desirable structure, located at the main entrances within the campus defined as major landmarks, then the square spaces defined as nodes, in addition to both stairs and corridors defined as paths. Finally, the districts, edges of educational buildings and service spaces are listed correspondingly in cognitive maps. Findings suggest that the spatial identity of the campus design is related mainly to the gates structures, squares and stairs.

Keywords: cognitive maps, university campus, urban memory, identity

Procedia PDF Downloads 135
6884 Analysis of Various Factors Affecting Hardness and Content of Phases Resulting from 1030 Carbon Steel Heat Treatment Using AC3 Software

Authors: Saeid Shahraki, Mohammad Mahdi Kaekha

Abstract:

1030 steel, a kind of carbon steel used in homogenization, cold-forming, quenching, and tempering conditions, is generally utilized in small parts resisting medium stress, such as connection foundations, hydraulic cylinders, tiny gears, pins, clamps, automotive normal forging parts, camshafts, levers, pundits, and nuts. In this study, AC3 software was used to measure the effect of carbon and manganese percentage, dimensions and geometry of pieces, the type of the cooling fluid, temperature, and time on hardness and the content of 1030 steel phases. Next, the results are compared with the analytical values obtained from the Lumped Capacity Method.

Keywords: 1030Steel, AC3software, heat treatment, lumped capacity method

Procedia PDF Downloads 266
6883 Presenting an Integrated Framework for the Introduction and Evaluation of Social Media in Enterprises

Authors: Gerhard Peter

Abstract:

In this paper, we present an integrated framework that governs the introduction of social media into enterprises and its evaluation. It is argued that the framework should address the following issues: (1) the contribution of social media for increasing efficiency and improving the quality of working life; (2) the level on which this contribution happens (i.e., individual, team, or organisation); (3) a description of the processes for implementing and evaluating social media; and the role of (4) organisational culture and (5) management. We also report the results of a case study where the framework has been employed to introduce a social networking platform at a German enterprise. This paper only considers the internal use of social media.

Keywords: case study, enterprise 2.0, framework, introducing and evaluating social media, social media

Procedia PDF Downloads 343
6882 Comparing the Effect of Group Education and Multimedia Software on Knowledge, Attitude and Self-Efficacy Mothers about of Sexual Health Education to the Boys of between 12-14 Years Old

Authors: Mirzaii Khadigeh

Abstract:

Background and objectives: Sexual health education is an important part of health promotion services. The major role of sex education is on mothers’ shoulders. So, they have to be equipped with enough knowledge, attitude and self-efficacy for teens’ education. The present study aimed to determine the effect of team-learning and multimedia software on mothers’ knowledge, attitudes and self-efficacy in sexual health education to 12-14-year-old sons in Mashhad in 1395. Materials and methods: In this research, two experimental and one control group were employed using random sampling, which was done on 132 mothers of high school pupils. They were randomly assigned into experimental and control groups. The data were collected using demographic information and a researcher-constructed questionnaire to investigate the mothers’ knowledge, attitude, and self-efficacy and DASS21(The Depression, Anxiety and Stress Scale). They were run after confirming their reliability and validity. Intervention for the multimedia group was in the form of four CDs- each for 45 minutes- that were given to the mothers each week. At the end of the fourth week, a question-and-answer session was administered for 60 minutes. The team-learning group received intervention once a week (totally four weeks). Two weeks later, the data were collected and analyzed via Chi-square, Fisher, Kruskal-Wallis and ANOVA. Findings: Knowledge, attitude and self-efficacy of mothers in sexual health before the intervention did not have any significant differences (p >0.05). At the end of the study, the difference between the scores of the knowledge, attitude and self-efficacy in the three groups was meaningfully different (p < 0/001), but the difference between the two groups of multimedia and team-learning was not significant (p> 0.05 ). Discussion and conclusion: The result reported the efficacy of both team-leaning and multimedia software, which implies that the multimedia software training method was as effective as team-learning training one on the knowledge, attitude and self-efficacy of mothers. But, the multimedia training method is highly advised due to its availability, flexibility, and interest.

Keywords: training one on the knowledge, attitude, self-efficacy of mothers, boys

Procedia PDF Downloads 155
6881 The Evaluation of Signal Timing Optimization and Implement of Transit Signal Priority in Intersections and Their Effect on Delay Reduction

Authors: Mohammad Reza Ramezani, Shahriyar Afandizadeh

Abstract:

Since the intersections play a crucial role in traffic delay, it is significant to evaluate them precisely. In this paper, three critical intersections in Tehran (Capital of Iran) had been simulated. The main purpose of this paper was to optimize the public transit delay. The simulation had three different phase in three intersections of Tehran. The first phase was about the current condition of intersection; the second phase was about optimized signal timing and the last phase was about prioritized public transit access. The Aimsun software was used to simulate all phases, and the Synchro software was used to optimization of signals as well. The result showed that the implement of optimization and prioritizing system would reduce about 50% of delay for public transit.

Keywords: transit signal priority, intersection optimization, public transit, simulation

Procedia PDF Downloads 458
6880 Development of Academic Software for Medial Axis Determination of Porous Media from High-Resolution X-Ray Microtomography Data

Authors: S. Jurado, E. Pazmino

Abstract:

Determination of the medial axis of a porous media sample is a non-trivial problem of interest for several disciplines, e.g., hydrology, fluid dynamics, contaminant transport, filtration, oil extraction, etc. However, the computational tools available for researchers are limited and restricted. The primary aim of this work was to develop a series of algorithms to extract porosity, medial axis structure, and pore-throat size distributions from porous media domains. A complementary objective was to provide the algorithms as free computational software available to the academic community comprising researchers and students interested in 3D data processing. The burn algorithm was tested on porous media data obtained from High-Resolution X-Ray Microtomography (HRXMT) and idealized computer-generated domains. The real data and idealized domains were discretized in voxels domains of 550³ elements and binarized to denote solid and void regions to determine porosity. Subsequently, the algorithm identifies the layer of void voxels next to the solid boundaries. An iterative process removes or 'burns' void voxels in sequence of layer by layer until all the void space is characterized. Multiples strategies were tested to optimize the execution time and use of computer memory, i.e., segmentation of the overall domain in subdomains, vectorization of operations, and extraction of single burn layer data during the iterative process. The medial axis determination was conducted identifying regions where burnt layers collide. The final medial axis structure was refined to avoid concave-grain effects and utilized to determine the pore throat size distribution. A graphic user interface software was developed to encompass all these algorithms, including the generation of idealized porous media domains. The software allows input of HRXMT data to calculate porosity, medial axis, and pore-throat size distribution and provide output in tabular and graphical formats. Preliminary tests of the software developed during this study achieved medial axis, pore-throat size distribution and porosity determination of 100³, 320³ and 550³ voxel porous media domains in 2, 22, and 45 minutes, respectively in a personal computer (Intel i7 processor, 16Gb RAM). These results indicate that the software is a practical and accessible tool in postprocessing HRXMT data for the academic community.

Keywords: medial axis, pore-throat distribution, porosity, porous media

Procedia PDF Downloads 100
6879 Glycan Analyzer: Software to Annotate Glycan Structures from Exoglycosidase Experiments

Authors: Ian Walsh, Terry Nguyen-Khuong, Christopher H. Taron, Pauline M. Rudd

Abstract:

Glycoproteins and their covalently bonded glycans play critical roles in the immune system, cell communication, disease and disease prognosis. Ultra performance liquid chromatography (UPLC) coupled with mass spectrometry is conventionally used to qualitatively and quantitatively characterise glycan structures in a given sample. Exoglycosidases are enzymes that catalyze sequential removal of monosaccharides from the non-reducing end of glycans. They naturally have specificity for a particular type of sugar, its stereochemistry (α or β anomer) and its position of attachment to an adjacent sugar on the glycan. Thus, monitoring the peak movements (both in the UPLC and MS1) after application of exoglycosidases provides a unique and effective way to annotate sugars with high detail - i.e. differentiating positional and linkage isomers. Manual annotation of an exoglycosidase experiment is difficult and time consuming. As such, with increasing sample complexity and the number of exoglycosidases, the analysis could result in manually interpreting hundreds of peak movements. Recently, we have implemented pattern recognition software for automated interpretation of UPLC-MS1 exoglycosidase digestions. In this work, we explain the software, indicate how much time it will save and provide example usage showing the annotation of positional and linkage isomers in Immunoglobulin G, apolipoprotein J, and simple glycan standards.

Keywords: bioinformatics, automated glycan assignment, liquid chromatography, mass spectrometry

Procedia PDF Downloads 180
6878 21st Century Provocation: Modern Slavery, the Implications for Individuals on the Autism Spectrum

Authors: Christina Surmei

Abstract:

Autism Spectrum Disorder (ASD) is defined as a diverse range of developmental conditions that affect an individual’s functionality. ASD is not linear, and individuals can present with deficits in social interaction, communication, and demonstrate limited, repetitive patterns of behaviour, interests, or activities. These characteristics may be observed in a variety of ways and range from mild to severe. ASD may include autism disorder, pervasive developmental disorder not otherwise specified, Asperger’s, or other related pervasive developmental disorders. Modern slavery is defined as 'situations of exploitation that a person cannot refuse or leave because of threats, violence, coercion, and abuse of power or deception'. A review of the literature investigated the prevalence of research regarding ASD and modern slavery. Two universal search engines and five online journals were used as the apparatuses of inquiry. The results revealed two editorials, one study, and one act, totaling four publications attesting to ASD and modern slavery as a joint entity. This is representative of a vast absence of research. However, as individual entities research on autism and modern slavery is in a general high occurrence. This paper has identified a significant gap in research on ASD and modern slavery, and initiates the dialogue to unpack a significant global issue in society today.

Keywords: autism spectrum, education, modern slavery, support

Procedia PDF Downloads 148
6877 CT Doses Pre and Post SAFIRE: Sinogram Affirmed Iterative Reconstruction

Authors: N. Noroozian, M. Halim, B. Holloway

Abstract:

Computed Tomography (CT) has become the largest source of radiation exposure in modern countries however, recent technological advances have created new methods to reduce dose without negatively affecting image quality. SAFIRE has emerged as a new software package which utilizes full raw data projections for iterative reconstruction, thereby allowing for lower CT dose to be used. this audit was performed to compare CT doses in certain examinations before and after the introduction of SAFIRE at our Radiology department which showed CT doses were significantly lower using SAFIRE compared with pre-SAFIRE software at SAFIRE 3 setting for the following studies:CSKUH Unenhanced brain scans (-20.9%), CABPEC Abdomen and pelvis with contrast (-21.5%), CCHAPC Chest with contrast (-24.4%), CCHAPC Abdomen and pelvis with contrast (-16.1%), CCHAPC Total chest, abdomen and pelvis (-18.7%).

Keywords: dose reduction, iterative reconstruction, low dose CT techniques, SAFIRE

Procedia PDF Downloads 266
6876 Using AI to Advance Factory Planning: A Case Study to Identify Success Factors of Implementing an AI-Based Demand Planning Solution

Authors: Ulrike Dowie, Ralph Grothmann

Abstract:

Rational planning decisions are based upon forecasts. Precise forecasting has, therefore, a central role in business. The prediction of customer demand is a prime example. This paper introduces recurrent neural networks to model customer demand and combines the forecast with uncertainty measures to derive decision support of the demand planning department. It identifies and describes the keys to the successful implementation of an AI-based solution: bringing together data with business knowledge, AI methods, and user experience, and applying agile software development practices.

Keywords: agile software development, AI project success factors, deep learning, demand forecasting, forecast uncertainty, neural networks, supply chain management

Procedia PDF Downloads 156
6875 Determining the City Development Based on the Modeling of the Pollutant Emission from Power Plant by Using AERMOD Software

Authors: Abbasi Fakhrossadat, Moharreri Mohammadamir, Shadmanmahani Mohammadjavad

Abstract:

The development of cities can be influenced by various factors, including air pollution. In this study, the focus is on the city of Mashhad, which has four large power plants operating. The emission of pollutants from these power plants can have a significant impact on the quality of life and health of the city's residents. Therefore, modeling and analyzing the emission pattern of pollutants can provide useful information for urban decision-makers and help in estimating the urban development model. The aim of this research is to determine the direction of city development based on the modeling of pollutant emissions (NOX, CO, and PM10) from power plants in Mashhad. By using the AERMOD software, the release of these pollutants will be modeled and analyzed.

Keywords: emission of air pollution, thermal power plant, urban development, AERMOD

Procedia PDF Downloads 56
6874 A Probability Analysis of Construction Project Schedule Using Risk Management Tool

Authors: A. L. Agarwal, D. A. Mahajan

Abstract:

Construction industry tumbled along with other industry/sectors during recent economic crash. Construction business could not regain thereafter and still pass through slowdown phase, resulted many real estate as well as infrastructure projects not completed on schedule and within budget. There are many theories, tools, techniques with software packages available in the market to analyze construction schedule. This study focuses on the construction project schedule and uncertainties associated with construction activities. The infrastructure construction project has been considered for the analysis of uncertainty on project activities affecting project duration and analysis is done using @RISK software. Different simulation results arising from three probability distribution functions are compiled to benefit construction project managers to plan more realistic schedule of various construction activities as well as project completion to document in the contract and avoid compensations or claims arising out of missing the planned schedule.

Keywords: construction project, distributions, project schedule, uncertainty

Procedia PDF Downloads 330
6873 Implementation of a Paraconsistent-Fuzzy Digital PID Controller in a Level Control Process

Authors: H. M. Côrtes, J. I. Da Silva Filho, M. F. Blos, B. S. Zanon

Abstract:

In a modern society the factor corresponding to the increase in the level of quality in industrial production demand new techniques of control and machinery automation. In this context, this work presents the implementation of a Paraconsistent-Fuzzy Digital PID controller. The controller is based on the treatment of inconsistencies both in the Paraconsistent Logic and in the Fuzzy Logic. Paraconsistent analysis is performed on the signals applied to the system inputs using concepts from the Paraconsistent Annotated Logic with annotation of two values (PAL2v). The signals resulting from the paraconsistent analysis are two values defined as Dc - Degree of Certainty and Dct - Degree of Contradiction, which receive a treatment according to the Fuzzy Logic theory, and the resulting output of the logic actions is a single value called the crisp value, which is used to control dynamic system. Through an example, it was demonstrated the application of the proposed model. Initially, the Paraconsistent-Fuzzy Digital PID controller was built and tested in an isolated MATLAB environment and then compared to the equivalent Digital PID function of this software for standard step excitation. After this step, a level control plant was modeled to execute the controller function on a physical model, making the tests closer to the actual. For this, the control parameters (proportional, integral and derivative) were determined for the configuration of the conventional Digital PID controller and of the Paraconsistent-Fuzzy Digital PID, and the control meshes in MATLAB were assembled with the respective transfer function of the plant. Finally, the results of the comparison of the level control process between the Paraconsistent-Fuzzy Digital PID controller and the conventional Digital PID controller were presented.

Keywords: fuzzy logic, paraconsistent annotated logic, level control, digital PID

Procedia PDF Downloads 267
6872 Predicting Daily Patient Hospital Visits Using Machine Learning

Authors: Shreya Goyal

Abstract:

The study aims to build user-friendly software to understand patient arrival patterns and compute the number of potential patients who will visit a particular health facility for a given period by using a machine learning algorithm. The underlying machine learning algorithm used in this study is the Support Vector Machine (SVM). Accurate prediction of patient arrival allows hospitals to operate more effectively, providing timely and efficient care while optimizing resources and improving patient experience. It allows for better allocation of staff, equipment, and other resources. If there's a projected surge in patients, additional staff or resources can be allocated to handle the influx, preventing bottlenecks or delays in care. Understanding patient arrival patterns can also help streamline processes to minimize waiting times for patients and ensure timely access to care for patients in need. Another big advantage of using this software is adhering to strict data protection regulations such as the Health Insurance Portability and Accountability Act (HIPAA) in the United States as the hospital will not have to share the data with any third party or upload it to the cloud because the software can read data locally from the machine. The data needs to be arranged in. a particular format and the software will be able to read the data and provide meaningful output. Using software that operates locally can facilitate compliance with these regulations by minimizing data exposure. Keeping patient data within the hospital's local systems reduces the risk of unauthorized access or breaches associated with transmitting data over networks or storing it in external servers. This can help maintain the confidentiality and integrity of sensitive patient information. Historical patient data is used in this study. The input variables used to train the model include patient age, time of day, day of the week, seasonal variations, and local events. The algorithm uses a Supervised learning method to optimize the objective function and find the global minima. The algorithm stores the values of the local minima after each iteration and at the end compares all the local minima to find the global minima. The strength of this study is the transfer function used to calculate the number of patients. The model has an output accuracy of >95%. The method proposed in this study could be used for better management planning of personnel and medical resources.

Keywords: machine learning, SVM, HIPAA, data

Procedia PDF Downloads 52
6871 Classification Framework of Production Planning and Scheduling Solutions from Supply Chain Management Perspective

Authors: Kwan Hee Han

Abstract:

In today’s business environments, frequent change of customer requirements is a tough challenge to manufacturing company. To cope with these challenges, a production planning and scheduling (PP&S) function might be established to provide accountability for both customer service and operational efficiency. Nowadays, many manufacturing firms have utilized PP&S software solutions to generate a realistic production plan and schedule to adapt to external changes efficiently. However, companies which consider the introduction of PP&S software solution, still have difficulties for selecting adequate solution to meet their specific needs. Since the task of PP&S is the one of major building blocks of SCM (Supply Chain Management) architecture, which deals with short term decision making in the production process of SCM, it is needed that the functionalities of PP&S should be analysed within the whole SCM process. The aim of this paper is to analyse the PP&S functionalities and its system architecture from the SCM perspective by using the criteria of level of planning hierarchy, major 4 SCM processes and problem-solving approaches, and finally propose a classification framework of PP&S solutions to facilitate the comparison among various commercial software solutions. By using proposed framework, several major PP&S solutions are classified and positioned according to their functional characteristics in this paper. By using this framework, practitioners who consider the introduction of computerized PP&S solutions in manufacturing firms can prepare evaluation and benchmarking sheets for selecting the most suitable solution with ease and in less time.

Keywords: production planning, production scheduling, supply chain management, the advanced planning system

Procedia PDF Downloads 185
6870 Least Squares Solution for Linear Quadratic Gaussian Problem with Stochastic Approximation Approach

Authors: Sie Long Kek, Wah June Leong, Kok Lay Teo

Abstract:

Linear quadratic Gaussian model is a standard mathematical model for the stochastic optimal control problem. The combination of the linear quadratic estimation and the linear quadratic regulator allows the state estimation and the optimal control policy to be designed separately. This is known as the separation principle. In this paper, an efficient computational method is proposed to solve the linear quadratic Gaussian problem. In our approach, the Hamiltonian function is defined, and the necessary conditions are derived. In addition to this, the output error is defined and the least-square optimization problem is introduced. By determining the first-order necessary condition, the gradient of the sum squares of output error is established. On this point of view, the stochastic approximation approach is employed such that the optimal control policy is updated. Within a given tolerance, the iteration procedure would be stopped and the optimal solution of the linear-quadratic Gaussian problem is obtained. For illustration, an example of the linear-quadratic Gaussian problem is studied. The result shows the efficiency of the approach proposed. In conclusion, the applicability of the approach proposed for solving the linear quadratic Gaussian problem is highly demonstrated.

Keywords: iteration procedure, least squares solution, linear quadratic Gaussian, output error, stochastic approximation

Procedia PDF Downloads 163
6869 Applications for Accounting of Inherited Object-Oriented Class Members

Authors: Jehad Al Dallal

Abstract:

A class in an Object-Oriented (OO) system is the basic unit of design, and it encapsulates a set of attributes and methods. In OO systems, instead of redefining the attributes and methods that are included in other classes, a class can inherit these attributes and methods and only implement its unique attributes and methods, which results in reducing code redundancy and improving code testability and maintainability. Such mechanism is called Class Inheritance. However, some software engineering applications may require accounting for all the inherited class members (i.e., attributes and methods). This paper explains how to account for inherited class members and discusses the software engineering applications that require such consideration.

Keywords: class flattening, external quality attribute, inheritance, internal quality attribute, object-oriented design

Procedia PDF Downloads 243
6868 Methodology for Temporary Analysis of Production and Logistic Systems on the Basis of Distance Data

Authors: M. Mueller, M. Kuehn, M. Voelker

Abstract:

In small and medium-sized enterprises (SMEs), the challenge is to create a well-grounded and reliable basis for process analysis, optimization and planning due to a lack of data. SMEs have limited access to methods with which they can effectively and efficiently analyse processes and identify cause-and-effect relationships in order to generate the necessary database and derive optimization potential from it. The implementation of digitalization within the framework of Industry 4.0 thus becomes a particular necessity for SMEs. For these reasons, the abstract presents an analysis methodology that is subject to the objective of developing an SME-appropriate methodology for efficient, temporarily feasible data collection and evaluation in flexible production and logistics systems as a basis for process analysis and optimization. The overall methodology focuses on retrospective, event-based tracing and analysis of material flow objects. The technological basis consists of Bluetooth low energy (BLE)-based transmitters, so-called beacons, and smart mobile devices (SMD), e.g. smartphones as receivers, between which distance data can be measured and derived motion profiles. The distance is determined using the Received Signal Strength Indicator (RSSI), which is a measure of signal field strength between transmitter and receiver. The focus is the development of a software-based methodology for interpretation of relative movements of transmitters and receivers based on distance data. The main research is on selection and implementation of pattern recognition methods for automatic process recognition as well as methods for the visualization of relative distance data. Due to an existing categorization of the database regarding process types, classification methods (e.g. Support Vector Machine) from the field of supervised learning are used. The necessary data quality requires selection of suitable methods as well as filters for smoothing occurring signal variations of the RSSI, the integration of methods for determination of correction factors depending on possible signal interference sources (columns, pallets) as well as the configuration of the used technology. The parameter settings on which respective algorithms are based have a further significant influence on result quality of the classification methods, correction models and methods for visualizing the position profiles used. The accuracy of classification algorithms can be improved up to 30% by selected parameter variation; this has already been proven in studies. Similar potentials can be observed with parameter variation of methods and filters for signal smoothing. Thus, there is increased interest in obtaining detailed results on the influence of parameter and factor combinations on data quality in this area. The overall methodology is realized with a modular software architecture consisting of independently modules for data acquisition, data preparation and data storage. The demonstrator for initialization and data acquisition is available as mobile Java-based application. The data preparation, including methods for signal smoothing, are Python-based with the possibility to vary parameter settings and to store them in the database (SQLite). The evaluation is divided into two separate software modules with database connection: the achievement of an automated assignment of defined process classes to distance data using selected classification algorithms and the visualization as well as reporting in terms of a graphical user interface (GUI).

Keywords: event-based tracing, machine learning, process classification, parameter settings, RSSI, signal smoothing

Procedia PDF Downloads 111
6867 Biomechanical Study of a Type II Superior Labral Anterior to Posterior Lesion in the Glenohumeral Joint Using Finite Element Analysis

Authors: Javier A. Maldonado E., Duvert A. Puentes T., Diego F. Villegas B.

Abstract:

The SLAP lesion (Superior Labral Anterior to Posterior) involves the labrum, causing pain and mobility problems in the glenohumeral joint. This injury is common in athletes practicing sports that requires throwing or those who receive traumatic impacts on the shoulder area. This paper determines the biomechanical behavior of soft tissues of the glenohumeral joint when type II SLAP lesion is present. This pathology is characterized for a tear in the superior labrum which is simulated in a 3D model of the shoulder joint. A 3D model of the glenohumeral joint was obtained using the free software Slice. Then, a Finite Element analysis was done using a general purpose software which simulates a compression test with external rotation. First, a validation was done assuming a healthy joint shoulder with a previous study. Once the initial model was validated, a lesion of the labrum built using a CAD software and the same test was done again. The results obtained were stress and strain distribution of the synovial capsule and the injured labrum. ANOVA was done for the healthy and injured glenohumeral joint finding significant differences between them. This study will help orthopedic surgeons to know the biomechanics involving this type of lesion and also the other surrounding structures affected by loading the injured joint.

Keywords: biomechanics, computational model, finite elements, glenohumeral joint, superior labral anterior to posterior lesion

Procedia PDF Downloads 190
6866 Realistic Modeling of the Preclinical Small Animal Using Commercial Software

Authors: Su Chul Han, Seungwoo Park

Abstract:

As the increasing incidence of cancer, the technology and modality of radiotherapy have advanced and the importance of preclinical model is increasing in the cancer research. Furthermore, the small animal dosimetry is an essential part of the evaluation of the relationship between the absorbed dose in preclinical small animal and biological effect in preclinical study. In this study, we carried out realistic modeling of the preclinical small animal phantom possible to verify irradiated dose using commercial software. The small animal phantom was modeling from 4D Digital Mouse whole body phantom. To manipulate Moby phantom in commercial software (Mimics, Materialise, Leuven, Belgium), we converted Moby phantom to DICOM image file of CT by Matlab and two- dimensional of CT images were converted to the three-dimensional image and it is possible to segment and crop CT image in Sagittal, Coronal and axial view). The CT images of small animals were modeling following process. Based on the profile line value, the thresholding was carried out to make a mask that was connection of all the regions of the equal threshold range. Using thresholding method, we segmented into three part (bone, body (tissue). lung), to separate neighboring pixels between lung and body (tissue), we used region growing function of Mimics software. We acquired 3D object by 3D calculation in the segmented images. The generated 3D object was smoothing by remeshing operation and smoothing operation factor was 0.4, iteration value was 5. The edge mode was selected to perform triangle reduction. The parameters were that tolerance (0.1mm), edge angle (15 degrees) and the number of iteration (5). The image processing 3D object file was converted to an STL file to output with 3D printer. We modified 3D small animal file using 3- Matic research (Materialise, Leuven, Belgium) to make space for radiation dosimetry chips. We acquired 3D object of realistic small animal phantom. The width of small animal phantom was 2.631 cm, thickness was 2.361 cm, and length was 10.817. Mimics software supported efficiency about 3D object generation and usability of conversion to STL file for user. The development of small preclinical animal phantom would increase reliability of verification of absorbed dose in small animal for preclinical study.

Keywords: mimics, preclinical small animal, segmentation, 3D printer

Procedia PDF Downloads 350
6865 Studies on Separation of Scandium from Sulfate Environment Using Ion Exchange Technique

Authors: H. Hajmohammadi , A. H. Jafari, M. Eskandari Nasab

Abstract:

The ion exchange method was used to assess the absorption of sulfate media from laboratory-grade materials. The Taguchi method was employed for determining the optimum conditions for scandium adsorption. Results show that optimum conditions for scandium adsorption from sulfate were obtained by Purolite C100 cationic resin in 0.1 g/l sulfuric acid and scandium concentration of 2 g/l at 25 °C. Studies also showed that lowering H₂SO₄ concentration and aqueous phase temperature leads to an increase in Sc adsorption. Visual Minteq software was used to ascertain the various possible cation types and the effect of concentration of scandium ion species on scandium adsorption by cationic resins. The simulation results of the above software show that scandium ion species are often cationic species that are consistent with experimental data.

Keywords: scandium, ion exchange resin, simulation, leach copper

Procedia PDF Downloads 120
6864 Gis Database Creation for Impacts of Domestic Wastewater Disposal on BIDA Town, Niger State Nigeria

Authors: Ejiobih Hyginus Chidozie

Abstract:

Geographic Information System (GIS) is a configuration of computer hardware and software specifically designed to effectively capture, store, update, manipulate, analyse and display and display all forms of spatially referenced information. GIS database is referred to as the heart of GIS. It has location data, attribute data and spatial relationship between the objects and their attributes. Sewage and wastewater management have assumed increased importance lately as a result of general concern expressed worldwide about the problems of pollution of the environment contamination of the atmosphere, rivers, lakes, oceans and ground water. In this research GIS database was created to study the impacts of domestic wastewater disposal methods on Bida town, Niger State as a model for investigating similar impacts on other cities in Nigeria. Results from GIS database are very useful to decision makers and researchers. Bida Town was subdivided into four regions, eight zones, and 24 sectors based on the prevailing natural morphology of the town. GIS receiver and structured questionnaire were used to collect information and attribute data from 240 households of the study area. Domestic wastewater samples were collected from twenty four sectors of the study area for laboratory analysis. ArcView 3.2a GIS software, was used to create the GIS databases for ecological, health and socioeconomic impacts of domestic wastewater disposal methods in Bida town.

Keywords: environment, GIS, pollution, software, wastewater

Procedia PDF Downloads 405
6863 Shoulder Range of Motion Measurements using Computer Vision Compared to Hand-Held Goniometric Measurements

Authors: Lakshmi Sujeesh, Aaron Ramzeen, Ricky Ziming Guo, Abhishek Agrawal

Abstract:

Introduction: Range of motion (ROM) is often measured by physiotherapists using hand-held goniometer as part of mobility assessment for diagnosis. Due to the nature of hand-held goniometer measurement procedure, readings often tend to have some variations depending on the physical therapist taking the measurements (Riddle et al.). This study aims to validate computer vision software readings against goniometric measurements for quick and consistent ROM measurements to be taken by clinicians. The use of this computer vision software hopes to improve the future of musculoskeletal space with more efficient diagnosis from recording of patient’s ROM with minimal human error across different physical therapists. Methods: Using the hand-held long arm goniometer measurements as the “gold-standard”, healthy study participants (n = 20) were made to perform 4 exercises: Front elevation, Abduction, Internal Rotation, and External Rotation, using both arms. Assessment of active ROM using computer vision software at different angles set by goniometer for each exercise was done. Interclass Correlation Coefficient (ICC) using 2-way random effects model, Box-Whisker plots, and Root Mean Square error (RMSE) were used to find the degree of correlation and absolute error measured between set and recorded angles across the repeated trials by the same rater. Results: ICC (2,1) values for all 4 exercises are above 0.9, indicating excellent reliability. Lowest overall RMSE was for external rotation (5.67°) and highest for front elevation (8.00°). Box-whisker plots showed have showed that there is a potential zero error in the measurements done by the computer vision software for abduction, where absolute error for measurements taken at 0 degree are shifted away from the ideal 0 line, with its lowest recorded error being 8°. Conclusion: Our results indicate that the use of computer vision software is valid and reliable to use in clinical settings by physiotherapists for measuring shoulder ROM. Overall, computer vision helps improve accessibility to quality care provided for individual patients, with the ability to assess ROM for their condition at home throughout a full cycle of musculoskeletal care (American Academy of Orthopaedic Surgeons) without the need for a trained therapist.

Keywords: physiotherapy, frozen shoulder, joint range of motion, computer vision

Procedia PDF Downloads 80
6862 Improvements and Implementation Solutions to Reduce the Computational Load for Traffic Situational Awareness with Alerts (TSAA)

Authors: Salvatore Luongo, Carlo Luongo

Abstract:

This paper discusses the implementation solutions to reduce the computational load for the Traffic Situational Awareness with Alerts (TSAA) application, based on Automatic Dependent Surveillance-Broadcast (ADS-B) technology. In 2008, there were 23 total mid-air collisions involving general aviation fixed-wing aircraft, 6 of which were fatal leading to 21 fatalities. These collisions occurred during visual meteorological conditions, indicating the limitations of the see-and-avoid concept for mid-air collision avoidance as defined in the Federal Aviation Administration’s (FAA). The commercial aviation aircraft are already equipped with collision avoidance system called TCAS, which is based on classic transponder technology. This system dramatically reduced the number of mid-air collisions involving air transport aircraft. In general aviation, the same reduction in mid-air collisions has not occurred, so this reduction is the main objective of the TSAA application. The major difference between the original conflict detection application and the TSAA application is that the conflict detection is focused on preventing loss of separation in en-route environments. Instead TSAA is devoted to reducing the probability of mid-air collision in all phases of flight. The TSAA application increases the flight crew traffic situation awareness providing alerts of traffic that are detected in conflict with ownship in support of the see-and-avoid responsibility. The relevant effort has been spent in the design process and the code generation in order to maximize the efficiency and performances in terms of computational load and memory consumption reduction. The TSAA architecture is divided into two high-level systems: the “Threats database” and the “Conflict detector”. The first one receives the traffic data from ADS-B device and provides the memorization of the target’s data history. Conflict detector module estimates ownship and targets trajectories in order to perform the detection of possible future loss of separation between ownship and each target. Finally, the alerts are verified by additional conflict verification logic, in order to prevent possible undesirable behaviors of the alert flag. In order to reduce the computational load, a pre-check evaluation module is used. This pre-check is only a computational optimization, so the performances of the conflict detector system are not modified in terms of number of alerts detected. The pre-check module uses analytical trajectories propagation for both target and ownship. This allows major accuracy and avoids the step-by-step propagation, which requests major computational load. Furthermore, the pre-check permits to exclude the target that is certainly not a threat, using an analytical and efficient geometrical approach, in order to decrease the computational load for the following modules. This software improvement is not suggested by FAA documents, and so it is the main innovation of this work. The efficiency and efficacy of this enhancement are verified using fast-time and real-time simulations and by the execution on a real device in several FAA scenarios. The final implementation also permits the FAA software certification in compliance with DO-178B standard. The computational load reduction allows the installation of TSAA application also on devices with multiple applications and/or low capacity in terms of available memory and computational capabilities

Keywords: traffic situation awareness, general aviation, aircraft conflict detection, computational load reduction, implementation solutions, software certification

Procedia PDF Downloads 263
6861 Investigation of Multiple Dynamic Vibration Absorbers' Performance in Overhead Transmission Lines

Authors: Pedro F. D. Oliveira, Rangel S. Maia, Aline S. Paula

Abstract:

As the electric energy consumption grows, the necessity of energy transmission lines increases. One of the problems caused by an oscillatory response to dynamical loads (such as wind effects) in transmission lines is the cable fatigue. Thus, the dynamical behavior of transmission cables understanding and its control is extremely important. The socioeconomic damage caused by a failure in these cables can be quite significant, from large economic losses to energy supply interruption in large regions. Dynamic Vibration Absorbers (DVA) are oscillatory elements used to mitigate the vibration of a primary system subjected to harmonic excitation. The positioning of Stockbridge (DVA for overhead transmission lines) plays an important role in mitigating oscillations of transmission lines caused by airflows. Nowadays, the positioning is defined by technical standards or commercial software. The aim of this paper is to conduct an analysis of multiple DVAs performances in cable conductors of overhead transmission lines. The cable is analyzed by a finite element method and the model is calibrated by experimental results. DVAs performance is analyzed by evaluating total cable energy, and a study of multiple DVAs positioning is conducted. The results are compared to the existing regulations showing situations where proper positioning, different from the standard, can lead to better performance of the DVA. Results also show situations where the use of multiple DVAs is appropriate.

Keywords: dynamical vibration absorber, finite element method, overhead transmission lines, structural dynamics

Procedia PDF Downloads 110
6860 Software Vulnerability Markets: Discoverers and Buyers

Authors: Abdullah M. Algarni, Yashwant K. Malaiya

Abstract:

Some of the key aspects of vulnerability-discovery, dissemination, and disclosure-have received some attention recently. However, the role of interaction among the vulnerability discoverers and vulnerability acquirers has not yet been adequately addressed. Our study suggests that a major percentage of discoverers, a majority in some cases, are unaffiliated with the software developers and thus are free to disseminate the vulnerabilities they discover in any way they like. As a result, multiple vulnerability markets have emerged. In some of these markets, the exchange is regulated, but in others, there is little or no regulation. In recent vulnerability discovery literature, the vulnerability discoverers have remained anonymous individuals. Although there has been an attempt to model the level of their efforts, information regarding their identities, modes of operation, and what they are doing with the discovered vulnerabilities has not been explored. Reports of buying and selling of the vulnerabilities are now appearing in the press; however, the existence of such markets requires validation, and the natures of the markets need to be analysed. To address this need, we have attempted to collect detailed information. We have identified the most prolific vulnerability discoverers throughout the past decade and examined their motivation and methods. A large percentage of these discoverers are located in Eastern and Western Europe and in the Far East. We have contacted several of them in order to collect first-hand information regarding their techniques, motivations, and involvement in the vulnerability markets. We examine why many of the discoverers appear to retire after a highly successful vulnerability-finding career. The paper identifies the actual vulnerability markets, rather than the hypothetical ideal markets that are often examined. The emergence of worldwide government agencies as vulnerability buyers has significant implications. We discuss potential factors that can impact the risk to society and the need for detailed exploration.

Keywords: risk management, software security, vulnerability discoverers, vulnerability markets

Procedia PDF Downloads 236
6859 Build Information Systems Environment Clean Through the Sms Gateway

Authors: Lutpi Ginanjar

Abstract:

Environmental hygiene is indispensable for people to live healthy, safe and peaceful. In a small environment, the cleanliness of the environment is very easy to overcome, but on the larger environment requires a more complicated management and considerable investments. In general environmental hygiene are managed by the Department of Hygiene and Landscaper. Found a good management, but much less good management. The difficulties that are often encountered on waste management also caused public awareness itself. In addition, communities have difficulty in making a report about the rubbish because not dibangunnyasistem good information. Essai aims to build information systems environment clean especially the handling of waste in the city of Bandung, West Java province. The system was built with PHP software. Expected results obtained after the construction of the information system of environmental hygiene can be demonstrated to the community will be the health of the environment.

Keywords: information systems, SMS gateway, management, software, PHP

Procedia PDF Downloads 461
6858 Determination of Lithology, Porosity and Water Saturation for Mishrif Carbonate Formation

Authors: F. S. Kadhim, A. Samsuri, H. Alwan

Abstract:

Well logging records can help to answer many questions from a wide range of special interested information and basic petrophysical properties to formation evaluation of oil and gas reservoirs. The accurate calculations of porosity in carbonate reservoirs are the most challenging aspects of well log analysis. Many equations have been developed over the years based on known physical principles or on empirically derived relationships, which are used to calculate porosity, estimate lithology and water saturation; however these parameters are calculated from well logs by using modern technique in a current study. Nasiriya (NS) oilfield is one of giant oilfields in the Middle East, and the formation under study is the Mishrif carbonate formation which is the shallowest hydrocarbon bearing zone in the NS oilfield. Neurolog software (V5, 2008) was used to digitize the scanned copies of the available logs. Environmental corrections had been made as per Schlumberger charts 2005, which supplied in the Interactive Petrophysics software (IP, V3.5, 2008). Three saturation models have been used to calculate water saturation of carbonate formations, which are simple Archie equation, Dual water model, and Indonesia model. Results indicate that the Mishrif formation consists mainly of limestone, some dolomite and shale. The porosity interpretation shows that the logging tools have a good quality after making the environmental corrections. The average formation water saturation for Mishrif formation is around 0.4-0.6.This study is provided accurate behavior of petrophysical properties with depth for this formation by using modern software.

Keywords: lithology, porosity, water saturation, carbonate formation, mishrif formation

Procedia PDF Downloads 358
6857 Analyzing the Impact of DCF and PCF on WLAN Network Standards 802.11a, 802.11b, and 802.11g

Authors: Amandeep Singh Dhaliwal

Abstract:

Networking solutions, particularly wireless local area networks have revolutionized the technological advancement. Wireless Local Area Networks (WLANs) have gained a lot of popularity as they provide location-independent network access between computing devices. There are a number of access methods used in Wireless Networks among which DCF and PCF are the fundamental access methods. This paper emphasizes on the impact of DCF and PCF access mechanisms on the performance of the IEEE 802.11a, 802.11b and 802.11g standards. On the basis of various parameters viz. throughput, delay, load etc performance is evaluated between these three standards using above mentioned access mechanisms. Analysis revealed a superior throughput performance with low delays for 802.11g standard as compared to 802.11 a/b standard using both DCF and PCF access methods.

Keywords: DCF, IEEE, PCF, WLAN

Procedia PDF Downloads 409