Search results for: software receiver
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 5072

Search results for: software receiver

4202 Proprioceptive Neuromuscular Facilitation Exercises of Upper Extremities Assessment Using Microsoft Kinect Sensor and Color Marker in a Virtual Reality Environment

Authors: M. Owlia, M. H. Azarsa, M. Khabbazan, A. Mirbagheri

Abstract:

Proprioceptive neuromuscular facilitation exercises are a series of stretching techniques that are commonly used in rehabilitation and exercise therapy. Assessment of these exercises for true maneuvering requires extensive experience in this field and could not be down with patients themselves. In this paper, we developed software that uses Microsoft Kinect sensor, a spherical color marker, and real-time image processing methods to evaluate patient’s performance in generating true patterns of movements. The software also provides the patient with a visual feedback by showing his/her avatar in a Virtual Reality environment along with the correct path of moving hand, wrist and marker. Primary results during PNF exercise therapy of a patient in a room environment shows the ability of the system to identify any deviation of maneuvering path and direction of the hand from the one that has been performed by an expert physician.

Keywords: image processing, Microsoft Kinect, proprioceptive neuromuscular facilitation, upper extremities assessment, virtual reality

Procedia PDF Downloads 273
4201 Analysis of Weather Variability Impact on Yields of Some Crops in Southwest, Nigeria

Authors: Olumuyiwa Idowu Ojo, Oluwatobi Peter Olowo

Abstract:

The study developed a Geographical Information Systems (GIS) database and mapped inter-annual changes in crop yields of cassava, cowpea, maize, rice, melon and yam as a response to inter-annual rainfall and temperature variability in Southwest, Nigeria. The aim of this project is to study the comparative analysis of the weather variability impact of six crops yield (Rice, melon, yam, cassava, Maize and cowpea) in South Western States of Nigeria (Oyo, Osun, Ekiti, Ondo, Ogun and Lagos) from 1991 – 2007. The data was imported and analysed in the Arch GIS 9 – 3 software environment. The various parameters (temperature, rainfall, crop yields) were interpolated using the kriging method. The results generated through interpolation were clipped to the study area. Geographically weighted regression was chosen from the spatial statistics toolbox in Arch GIS 9.3 software to analyse and predict the relationship between temperature, rainfall and the different crops (Cowpea, maize, rice, melon, yam, and cassava).

Keywords: GIS, crop yields, comparative analysis, temperature, rainfall, weather variability

Procedia PDF Downloads 324
4200 Application of Agile Project Management to Construction Projects: Case Study

Authors: Ran Etgar, Sarit Freund

Abstract:

Agile project management (APM) has been developed originally for software development project. Construction projects seemed to be more apt to traditional water-fall approach than to APM. However, Construction project suffers from similar problems that necessitated the invention of APM, mainly the need to break down the project structure to small increments, thus minimizing the needed managerial planning and design. Since the classical structure of APM is not applicable the way it is to construction project, a modified version of APM was devised. This method, nicknamed 'The anchor method', exploits the fundamentals of APM (i.e., iterations, or sprints of short time frames or timeboxes, cross-functional teams, risk reduction and adaptation to changes) and adjust them to the construction world. The projects had to be structured appropriately to proactively and quickly adapt to change. The method aims to encompass human behavior and lean towards adaptivity rather than predictability. To enable smooth application of the method, a special project management software was developed, so as to provide solid administrational help and accurate data. The method is tested on a bunch of construction projects and some key performance indicators (KPIs) are collected. According to preliminary results the method is indeed very advantageous and with proper assimilation can radically change the construction project management paradigm.

Keywords: agile project management, construction, information systems, project management

Procedia PDF Downloads 130
4199 Stabilizing a Failed Slope in Islamabad, Pakistan

Authors: Muhammad Umer Zubair, Kamran Akhtar, Muhammad Arsalan Khan

Abstract:

This paper is based on a research carried out on a failed slope in Defence Housing Authority (DHA) Phase I, Islamabad. The research included determination of Soil parameters, Site Surveying and Cost Estimation. Apart from these, the use of three dimensional (3D) slope stability analysis in conjunction with two dimensional (2D) analysis was used determination of slope conditions. In addition collection of soil reports, a detailed survey was carried out to create a 3D model in Surfer 8 software. 2D cross-sections that needed to be analyzed for stability were generated from 3D model. Slope stability softwares, Rocscience Slide 6.0 and Clara-W were employed for 2D and 3D Analyses respectively which have the ability to solve complex mathematical functions. Results of the analyses were used to confirm site conditions and the threats were identified to recommend suitable remedies.The most effective remedy was suggested for slope stability after analyzing all remedies in software Slide 6 and its feasibility was determined through cost benefit analysis. This paper should be helpful to Geotechnical engineers, design engineers and the organizations working with slope stability.

Keywords: slope stability, Rocscience, Clara W., 2d analysis, 3D analysis, sensitivity analysis

Procedia PDF Downloads 522
4198 Implementing Online Applications to Allow Marketing Personnel to Share Their Experiences

Authors: Ishak Kamal Baskhayroun

Abstract:

This study examines consumer attitudes toward mobile marketing, especially toward SMS marketing. Unlike similar studies, this study does not focus on the young, but includes consumers who are in the 18-70 age range to the field research. According to the results, it has been concluded that most participants think SMS marketing is disturbing. Most important problems with SMS marketing are about getting subscribed to message lists without the permission of the receiver; the high number of messages sent; and the irrelevancy of the message content. The emergence of sponsorship as a new marketing communication tool and a source of competitive advantage in the marketplace has changed the entire marketing communication process. Sponsorship has overtaken other marketing communication tools in terms of growth and expenditure. This paper seeks to evaluate the role of sponsorship in marketing communication tools. The study recommends that proper measures be taken before the company embarks into sponsorship programs. This is necessary because investment in sponsorship does not always guarantee sustainable competitive advantage in the marketplace.Mobile phones are one of the direct marketing tools that can be used to reach today’s hard to reach consumers. Mobile phones are very personal devices and they are always carried with the consumer, where ever they go. This creates an opportunity for marketers to create personalized marketing communications messages and send them on the right time and place.

Keywords: employee organizational performance, internal marketing, internal customer, direct marketing, mobile phones mobile marketing, sms advertising, marketing sponsorship, marketing communication theories, marketing communication tools corporate responsibility

Procedia PDF Downloads 58
4197 Proposed Solutions Based on Affective Computing

Authors: Diego Adrian Cardenas Jorge, Gerardo Mirando Guisado, Alfredo Barrientos Padilla

Abstract:

A system based on Affective Computing can detect and interpret human information like voice, facial expressions and body movement to detect emotions and execute a corresponding response. This data is important due to the fact that a person can communicate more effectively with emotions than can be possible with words. This information can be processed through technological components like Facial Recognition, Gait Recognition or Gesture Recognition. As of now, solutions proposed using this technology only consider one component at a given moment. This research investigation proposes two solutions based on Affective Computing taking into account more than one component for emotion detection. The proposals reflect the levels of dependency between hardware devices and software, as well as the interaction process between the system and the user which implies the development of scenarios where both proposals will be put to the test in a live environment. Both solutions are to be developed in code by software engineers to prove the feasibility. To validate the impact on society and business interest, interviews with stakeholders are conducted with an investment mind set where each solution is labeled on a scale of 1 through 5, being one a minimum possible investment and 5 the maximum.

Keywords: affective computing, emotions, emotion detection, face recognition, gait recognition

Procedia PDF Downloads 368
4196 Incidence and Prevalence of Dry Eye Syndrome in Different Occupational Sector of Society

Authors: Vergeena Varghese, G. Gajalakshmi, Jayarajini Vasanth

Abstract:

The present study deals with the indication of prevalence of dry eye and evaluates environmental risk factors attributed to dry eye in different occupational sectors. 240 subjects above 20 years and below 45 years of age were screened for dry eye. Mcmonnies dry eye questionnaire based history and Schirmer’s test were used to diagnose dry eye. For Schirmer’s test Whatman strip and paracaine drop used as an anesthetic. Subject’s demographics include age, sex, smoking, alcoholism, occupation history and working environment. Out of a total of 240 subjects, 52 subjects were positive for dry eye syndrome (21.7%). The highest prevalence of dry eye syndrome in software sector was 14subjects (26.9%) out of a total of 40 subjects. In the construction sector, the prevalence of dry eye syndrome had 12 subjects (23.1%) out of 40 subjects and 9 subjects (17.3%) out of 40 subjects in agriculture sector. 7 subjects (13.5%) who had dry eye out of 40 subjects in the transport sector and in industrial 6 subjects (11.5%). In a normal sector, this was taken as control group had dry eye in 4 subjects (7.7%) out of 40 subjects. We also found the prevalence of dry eye in OS was higher than OD. Dry eye is a most common ocular condition. The highest prevalence of dry eye syndrome in software sector was 14 members than other sector. There was a significant correlation between environmental and occupational factors to cause dry eye. Excessive exposure to sunlight, wind, high temperature, and air pollution, electromagnetic radiation are the factors affect the tear film and ocular surface causing the dry eye syndrome.

Keywords: DES – dry eye syndrome, Mcmonnies dry eye questionnaire, schirmer’s test, whatman vstrip

Procedia PDF Downloads 467
4195 An Automatic Model Transformation Methodology Based on Semantic and Syntactic Comparisons and the Granularity Issue Involved

Authors: Tiexin Wang, Sebastien Truptil, Frederick Benaben

Abstract:

Model transformation, as a pivotal aspect of Model-driven engineering, attracts more and more attentions both from researchers and practitioners. Many domains (enterprise engineering, software engineering, knowledge engineering, etc.) use model transformation principles and practices to serve to their domain specific problems; furthermore, model transformation could also be used to fulfill the gap between different domains: by sharing and exchanging knowledge. Since model transformation has been widely used, there comes new requirement on it: effectively and efficiently define the transformation process and reduce manual effort that involved in. This paper presents an automatic model transformation methodology based on semantic and syntactic comparisons, and focuses particularly on granularity issue that existed in transformation process. Comparing to the traditional model transformation methodologies, this methodology serves to a general purpose: cross-domain methodology. Semantic and syntactic checking measurements are combined into a refined transformation process, which solves the granularity issue. Moreover, semantic and syntactic comparisons are supported by software tool; manual effort is replaced in this way.

Keywords: automatic model transformation, granularity issue, model-driven engineering, semantic and syntactic comparisons

Procedia PDF Downloads 394
4194 Using AI Based Software as an Assessment Aid for University Engineering Assignments

Authors: Waleed Al-Nuaimy, Luke Anastassiou, Manjinder Kainth

Abstract:

As the process of teaching has evolved with the advent of new technologies over the ages, so has the process of learning. Educators have perpetually found themselves on the lookout for new technology-enhanced methods of teaching in order to increase learning efficiency and decrease ever expanding workloads. Shortly after the invention of the internet, web-based learning started to pick up in the late 1990s and educators quickly found that the process of providing learning material and marking assignments could change thanks to the connectivity offered by the internet. With the creation of early web-based virtual learning environments (VLEs) such as SPIDER and Blackboard, it soon became apparent that VLEs resulted in higher reported computer self-efficacy among students, but at the cost of students being less satisfied with the learning process . It may be argued that the impersonal nature of VLEs, and their limited functionality may have been the leading factors contributing to this reported dissatisfaction. To this day, often faced with the prospects of assigning colossal engineering cohorts their homework and assessments, educators may frequently choose optimally curated assessment formats, such as multiple-choice quizzes and numerical answer input boxes, so that automated grading software embedded in the VLEs can save time and mark student submissions instantaneously. A crucial skill that is meant to be learnt during most science and engineering undergraduate degrees is gaining the confidence in using, solving and deriving mathematical equations. Equations underpin a significant portion of the topics taught in many STEM subjects, and it is in homework assignments and assessments that this understanding is tested. It is not hard to see that this can become challenging if the majority of assignment formats students are engaging with are multiple-choice questions, and educators end up with a reduced perspective of their students’ ability to manipulate equations. Artificial intelligence (AI) has in recent times been shown to be an important consideration for many technologies. In our paper, we explore the use of new AI based software designed to work in conjunction with current VLEs. Using our experience with the software, we discuss its potential to solve a selection of problems ranging from impersonality to the reduction of educator workloads by speeding up the marking process. We examine the software’s potential to increase learning efficiency through its features which claim to allow more customized and higher-quality feedback. We investigate the usability of features allowing students to input equation derivations in a range of different forms, and discuss relevant observations associated with these input methods. Furthermore, we make ethical considerations and discuss potential drawbacks to the software, including the extent to which optical character recognition (OCR) could play a part in the perpetuation of errors and create disagreements between student intent and their submitted assignment answers. It is the intention of the authors that this study will be useful as an example of the implementation of AI in a practical assessment scenario insofar as serving as a springboard for further considerations and studies that utilise AI in the setting and marking of science and engineering assignments.

Keywords: engineering education, assessment, artificial intelligence, optical character recognition (OCR)

Procedia PDF Downloads 122
4193 Assessment of the Contribution of Geographic Information System Technology in Non Revenue Water: Case Study Dar Es Salaam Water and Sewerage Authority Kawe - Mzimuni Street

Authors: Victor Pesco Kassa

Abstract:

This research deals with the assessment of the contribution of GIS Technology in NRW. This research was conducted at Dar, Kawe Mzimuni Street. The data collection was obtained from existing source which is DAWASA HQ. The interpretation of the data was processed by using ArcGIS software. The data collected from the existing source reveals a good coverage of DAWASA’s water network at Mzimuni Street. Most of residents are connected to the DAWASA’s customer service. Also the collected data revealed that by using GIS DAWASA’s customer Geodatabase has been improved. Through GIS we can prepare customer location map purposely for site surveying also this map will be able to show different type of customer that are connected to DAWASA’s water service. This is a perfect contribution of the GIS Technology to address and manage the problem of NRW in DAWASA. Finally, the study recommends that the same study should be conducted in other DAWASA’s zones such as Temeke, Boko and Bagamoyo not only at Kawe Mzimuni Street. Through this study it is observed that ArcGIS software can offer powerful tools for managing and processing information geographically and in water and sanitation authorities such as DAWASA.

Keywords: DAWASA, NRW, Esri, EURA, ArcGIS

Procedia PDF Downloads 83
4192 Design and Development of the Force Plate for the Study of Driving-Point Biodynamic Responses

Authors: Vikas Kumar, V. H. Saran, Arpit Mathur, Avik Kathuria

Abstract:

The evaluation of biodynamic responses of the human body to whole body vibration exposure is necessary to quantify the exposure effects. A force plate model has been designed with the help of CAD software, which was investigated by performing the modal, stress and strain analysis using finite element approach in the software. The results of the modal, stress and strain analysis were under the limits for measurements of biodynamic responses to whole body vibration. The physical model of the force plate was manufactured and fixed to the vibration simulator and further used in the experimentation for the evaluation of apparent mass responses of the ten recruited subjects standing in an erect posture exposed to vertical whole body vibration. The platform was excited with sinusoidal vibration at vibration magnitude: 1.0 and 1.5 m/s2 rms at different frequency of 2, 3, 4, 5, 6, 8, 10, 12.5, 16 and 20 Hz. The results of magnitude of normalised apparent mass have shown the trend observed in the many past studies. The peak in the normalised apparent mass has been observed at 4 & 5 Hz frequency of vertical whole body vibration. The nonlinearity with respect to vibration magnitude has been also observed in the normalised apparent mass responses.

Keywords: whole body vibration, apparent mass, modeling, force plate

Procedia PDF Downloads 415
4191 Data and Spatial Analysis for Economy and Education of 28 E.U. Member-States for 2014

Authors: Alexiou Dimitra, Fragkaki Maria

Abstract:

The objective of the paper is the study of geographic, economic and educational variables and their contribution to determine the position of each member-state among the EU-28 countries based on the values of seven variables as given by Eurostat. The Data Analysis methods of Multiple Factorial Correspondence Analysis (MFCA) Principal Component Analysis and Factor Analysis have been used. The cross tabulation tables of data consist of the values of seven variables for the 28 countries for 2014. The data are manipulated using the CHIC Analysis V 1.1 software package. The results of this program using MFCA and Ascending Hierarchical Classification are given in arithmetic and graphical form. For comparison reasons with the same data the Factor procedure of Statistical package IBM SPSS 20 has been used. The numerical and graphical results presented with tables and graphs, demonstrate the agreement between the two methods. The most important result is the study of the relation between the 28 countries and the position of each country in groups or clouds, which are formed according to the values of the corresponding variables.

Keywords: Multiple Factorial Correspondence Analysis, Principal Component Analysis, Factor Analysis, E.U.-28 countries, Statistical package IBM SPSS 20, CHIC Analysis V 1.1 Software, Eurostat.eu Statistics

Procedia PDF Downloads 511
4190 Comparative Study of Accuracy of Land Cover/Land Use Mapping Using Medium Resolution Satellite Imagery: A Case Study

Authors: M. C. Paliwal, A. K. Jain, S. K. Katiyar

Abstract:

Classification of satellite imagery is very important for the assessment of its accuracy. In order to determine the accuracy of the classified image, usually the assumed-true data are derived from ground truth data using Global Positioning System. The data collected from satellite imagery and ground truth data is then compared to find out the accuracy of data and error matrices are prepared. Overall and individual accuracies are calculated using different methods. The study illustrates advanced classification and accuracy assessment of land use/land cover mapping using satellite imagery. IRS-1C-LISS IV data were used for classification of satellite imagery. The satellite image was classified using the software in fourteen classes namely water bodies, agricultural fields, forest land, urban settlement, barren land and unclassified area etc. Classification of satellite imagery and calculation of accuracy was done by using ERDAS-Imagine software to find out the best method. This study is based on the data collected for Bhopal city boundaries of Madhya Pradesh State of India.

Keywords: resolution, accuracy assessment, land use mapping, satellite imagery, ground truth data, error matrices

Procedia PDF Downloads 507
4189 System and Method for Providing Web-Based Remote Application Service

Authors: Shuen-Tai Wang, Yu-Ching Lin, Hsi-Ya Chang

Abstract:

With the development of virtualization technologies, a new type of service named cloud computing service is produced. Cloud users usually encounter the problem of how to use the virtualized platform easily over the web without requiring the plug-in or installation of special software. The object of this paper is to develop a system and a method enabling process interfacing within an automation scenario for accessing remote application by using the web browser. To meet this challenge, we have devised a web-based interface that system has allowed to shift the GUI application from the traditional local environment to the cloud platform, which is stored on the remote virtual machine. We designed the sketch of web interface following the cloud virtualization concept that sought to enable communication and collaboration among users. We describe the design requirements of remote application technology and present implementation details of the web application and its associated components. We conclude that this effort has the potential to provide an elastic and resilience environment for several application services. Users no longer have to burden the system maintenances and reduce the overall cost of software licenses and hardware. Moreover, this remote application service represents the next step to the mobile workplace, and it lets user to use the remote application virtually from anywhere.

Keywords: virtualization technology, virtualized platform, web interface, remote application

Procedia PDF Downloads 288
4188 Decision Analysis Module for Excel

Authors: Radomir Perzina, Jaroslav Ramik

Abstract:

The Analytic Hierarchy Process is frequently used approach for solving decision making problems. There exists wide range of software programs utilizing that approach. Their main disadvantage is that they are relatively expensive and missing intermediate calculations. This work introduces a Microsoft Excel add-in called DAME – Decision Analysis Module for Excel. Comparing to other computer programs DAME is free, can work with scenarios or multiple decision makers and displays intermediate calculations. Users can structure their decision models into three levels – scenarios/users, criteria and variants. Items on all levels can be evaluated either by weights or pair-wise comparisons. There are provided three different methods for the evaluation of the weights of criteria, the variants as well as the scenarios – Saaty’s Method, Geometric Mean Method and Fuller’s Triangle Method. Multiplicative and additive syntheses are supported. The proposed software package is demonstrated on couple of illustrating examples of real life decision problems.

Keywords: analytic hierarchy process, multi-criteria decision making, pair-wise comparisons, Microsoft Excel, scenarios

Procedia PDF Downloads 452
4187 An Effective Route to Control of the Safety of Accessing and Storing Data in the Cloud-Based Data Base

Authors: Omid Khodabakhshi, Amir Rozdel

Abstract:

The subject of cloud computing security research has allocated a number of challenges and competitions because the data center is comprised of complex private information and are always faced various risks of information disclosure by hacker attacks or internal enemies. Accordingly, the security of virtual machines in the cloud computing infrastructure layer is very important. So far, there are many software solutions to develop security in virtual machines. But using software alone is not enough to solve security problems. The purpose of this article is to examine the challenges and security requirements for accessing and storing data in an insecure cloud environment. In other words, in this article, a structure is proposed for the implementation of highly isolated security-sensitive codes using secure computing hardware in virtual environments. It also allows remote code validation with inputs and outputs. We provide these security features even in situations where the BIOS, the operating system, and even the super-supervisor are infected. To achieve these goals, we will use the hardware support provided by the new Intel and AMD processors, as well as the TPM security chip. In conclusion, the use of these technologies ultimately creates a root of dynamic trust and reduces TCB to security-sensitive codes.

Keywords: code, cloud computing, security, virtual machines

Procedia PDF Downloads 191
4186 Comparison of Irradiance Decomposition and Energy Production Methods in a Solar Photovoltaic System

Authors: Tisciane Perpetuo e Oliveira, Dante Inga Narvaez, Marcelo Gradella Villalva

Abstract:

Installations of solar photovoltaic systems have increased considerably in the last decade. Therefore, it has been noticed that monitoring of meteorological data (solar irradiance, air temperature, wind velocity, etc.) is important to predict the potential of a given geographical area in solar energy production. In this sense, the present work compares two computational tools that are capable of estimating the energy generation of a photovoltaic system through correlation analyzes of solar radiation data: PVsyst software and an algorithm based on the PVlib package implemented in MATLAB. In order to achieve the objective, it was necessary to obtain solar radiation data (measured and from a solarimetric database), analyze the decomposition of global solar irradiance in direct normal and horizontal diffuse components, as well as analyze the modeling of the devices of a photovoltaic system (solar modules and inverters) for energy production calculations. Simulated results were compared with experimental data in order to evaluate the performance of the studied methods. Errors in estimation of energy production were less than 30% for the MATLAB algorithm and less than 20% for the PVsyst software.

Keywords: energy production, meteorological data, irradiance decomposition, solar photovoltaic system

Procedia PDF Downloads 142
4185 Design and Implementation of Remote Application Virtualization in Cloud Environments

Authors: Shuen-Tai Wang, Ying-Chuan Chen, Hsi-Ya Chang

Abstract:

Cloud computing is a paradigm of computing that shifts the way computing has been done in the past. The users can use cloud resources such as application software or storage space from the cloud without needing to own them. This paper is focused on solutions that are anticipated to introduce IaaS idea to build cloud base services and enable the individual remote user's applications in cloud environments, which appear as if they are running on the end user's local computer. The available features of application delivery solution have been developed based on our previous research on the virtualization technology to offer applications independent of location so that the users can work online, offline, anywhere, with appropriate device and at any time. This proposed effort has the potential to positively provide an efficient, resilience and elastic environment for cloud service. Users no longer need to burden the system managers and drastically reduces the overall cost of hardware and software licenses. Moreover, this flexible remote application virtualization service represents the next significant step to the mobile workplace, and it lets users access their applications remotely through cloud services anywhere. This is also made possible by the low administrative costs as well as relatively inexpensive end-user terminals and reduced energy expenses.

Keywords: cloud computing, IaaS, virtualization, application delivery

Procedia PDF Downloads 281
4184 Assessing the Effect of Underground Tunnel Diameter on Structure-Foundation-Soil Performance under the Kobe Earthquake

Authors: Masoud Mahdavi

Abstract:

Today, developed and industrial cities have all kinds of sewage and water transfer canals, subway tunnels, infrastructure facilities, etc., which have caused underground cavities to be created under the buildings. The presence of these cavities causes behavioral changes in the structural behavior that must be fully evaluated. In the present study, using Abaqus finite element software, the effect of cavities with 0.5 and 1.5 meters in diameter at a depth of 2.5 meters from the earth's surface (with a circular cross-section) on the performance of the foundation and the ground (soil) has been evaluated. For this purpose, the Kobe earthquake was applied to the models for 10 seconds. Also, pore water pressure and weight were considered on the models to get complete results. The results showed that by creating and increasing the diameter of circular cavities in the soil, three indicators; 1) von Mises stress, 2) displacement and 3) plastic strain have had oscillating, ascending and ascending processes, respectively, which shows the relationship between increasing the diameter index of underground cavities and structural indicators of structure-foundation-soil.

Keywords: underground excavations, foundation, structural substrates, Abaqus software, Kobe earthquake, time history analysis

Procedia PDF Downloads 121
4183 Communication Aesthetics of Techno-Scenery and Lighting in Bolanle Austen-Peters Queen Moremi the Musical

Authors: Badeji Adebayo John

Abstract:

Technology has immense contribution in every aspect of human endeavor; it has not only made work easier but also provided exhilarating impression in the mind of the people. Theatre is not exempted from the multifaceted influence of technology on phenomenon. Therefore, theatre performances have experienced the excellence of technology in the contemporary era such that audiences have unforgettable experiences after seeing theatre performances. Some of these technological advancements that have amplified the aesthetics of performances in the theatre are techno-scenery (3D mapping) and lighting. In view of this, the objective of this study is to explore how techno-scenery and lighting technologies were used to communicate messages in the performance of Queen Moremi the Musical. In so doing, Participant-Observation Method and Content Analysis are adopted. Berlo’s model of communication is also employed to explain the communicative aesthetics of these theatre technologies in the performance. Techno-scenery and lighting are communication media modifier that facilitates audiences’ comprehension of the messages in the performance of Queen Moremi the Musical. They also create clear motion pictures of the setting which the performers cannot communicate in their acting, dances and singing, to ease the audiences’ decoding of messages that the performers are sending to the audience. Therefore, consistent incorporation of these technologies to theatre performances will facilitate easy flow of communication in-between the performers who are the sender, the message which is the performance and the audience who are the receiver.

Keywords: communication, aesthetics, techno-scenery, lighting, musical

Procedia PDF Downloads 85
4182 Specification of Requirements to Ensure Proper Implementation of Security Policies in Cloud-Based Multi-Tenant Systems

Authors: Rebecca Zahra, Joseph G. Vella, Ernest Cachia

Abstract:

The notion of cloud computing is rapidly gaining ground in the IT industry and is appealing mostly due to making computing more adaptable and expedient whilst diminishing the total cost of ownership. This paper focuses on the software as a service (SaaS) architecture of cloud computing which is used for the outsourcing of databases with their associated business processes. One approach for offering SaaS is basing the system’s architecture on multi-tenancy. Multi-tenancy allows multiple tenants (users) to make use of the same single application instance. Their requests and configurations might then differ according to specific requirements met through tenant customisation through the software. Despite the known advantages, companies still feel uneasy to opt for the multi-tenancy with data security being a principle concern. The fact that multiple tenants, possibly competitors, would have their data located on the same server process and share the same database tables heighten the fear of unauthorised access. Security is a vital aspect which needs to be considered by application developers, database administrators, data owners and end users. This is further complicated in cloud-based multi-tenant system where boundaries must be established between tenants and additional access control models must be in place to prevent unauthorised cross-tenant access to data. Moreover, when altering the database state, the transactions need to strictly adhere to the tenant’s known business processes. This paper focuses on the fact that security in cloud databases should not be considered as an isolated issue. Rather it should be included in the initial phases of the database design and monitored continuously throughout the whole development process. This paper aims to identify a number of the most common security risks and threats specifically in the area of multi-tenant cloud systems. Issues and bottlenecks relating to security risks in cloud databases are surveyed. Some techniques which might be utilised to overcome them are then listed and evaluated. After a description and evaluation of the main security threats, this paper produces a list of software requirements to ensure that proper security policies are implemented by a software development team when designing and implementing a multi-tenant based SaaS. This would then assist the cloud service providers to define, implement, and manage security policies as per tenant customisation requirements whilst assuring security for the customers’ data.

Keywords: cloud computing, data management, multi-tenancy, requirements, security

Procedia PDF Downloads 156
4181 Assertion-Driven Test Repair Based on Priority Criteria

Authors: Ruilian Zhao, Shukai Zhang, Yan Wang, Weiwei Wang

Abstract:

Repairing broken test cases is an expensive and challenging task in evolving software systems. Although an automated repair technique with intent preservation has been proposed, but it does not take into account the association between test repairs and assertions, leading to a large number of irrelevant candidates and decreasing the repair capability. This paper proposes an assertion-driven test repair approach. Furthermore, an intent-oriented priority criterion is raised to guide the repair candidate generation, making the repairs closer to the intent of the test. In more detail, repair targets are determined through post-dominance relations between assertions and the methods that directly cause compilation errors. Then, test repairs are generated from the target in a bottom-up way, guided by the intent-oriented priority criteria. Finally, the generated repair candidates are prioritized to match the original test intent. The approach is implemented and evaluated on the benchmark of 4 open-source programs and 91 broken test cases. The result shows that the approach can fix 89% (81/91) of broken test cases, which is more effective than the existing intentpreserved test repair approach, and our intent-oriented priority criteria work well.

Keywords: test repair, test intent, software test, test case evolution

Procedia PDF Downloads 129
4180 The Ontological Memory in Bergson as a Conceptual Tool for the Analysis of the Digital Conjuncture

Authors: Douglas Rossi Ramos

Abstract:

The current digital conjuncture, called by some authors as 'Internet of Things' (IoT), 'Web 2.0' or even 'Web 3.0', consists of a network that encompasses any communication of objects and entities, such as data, information, technologies, and people. At this juncture, especially characterized by an "object socialization," communication can no longer be represented as a simple informational flow of messages from a sender, crossing a channel or medium, reaching a receiver. The idea of communication must, therefore, be thought of more broadly in which it is possible to analyze the process communicative from interactions between humans and nonhumans. To think about this complexity, a communicative process that encompasses both humans and other beings or entities communicating (objects and things), it is necessary to constitute a new epistemology of communication to rethink concepts and notions commonly attributed to humans such as 'memory.' This research aims to contribute to this epistemological constitution from the discussion about the notion of memory according to the complex ontology of Henri Bergson. Among the results (the notion of memory in Bergson presents itself as a conceptual tool for the analysis of posthumanism and the anthropomorphic conjuncture of the new advent of digital), there was the need to think about an ontological memory, analyzed as a being itself (being itself of memory), as a strategy for understanding the forms of interaction and communication that constitute the new digital conjuncture, in which communicating beings or entities tend to interact with each other. Rethinking the idea of communication beyond the dimension of transmission in informative sequences paves the way for an ecological perspective of the digital dwelling condition.

Keywords: communication, digital, Henri Bergson, memory

Procedia PDF Downloads 164
4179 R Statistical Software Applied in Reliability Analysis: Case Study of Diesel Generator Fans

Authors: Jelena Vucicevic

Abstract:

Reliability analysis represents a very important task in different areas of work. In any industry, this is crucial for maintenance, efficiency, safety and monetary costs. There are ways to calculate reliability, unreliability, failure density and failure rate. This paper will try to introduce another way of calculating reliability by using R statistical software. R is a free software environment for statistical computing and graphics. It compiles and runs on a wide variety of UNIX platforms, Windows and MacOS. The R programming environment is a widely used open source system for statistical analysis and statistical programming. It includes thousands of functions for the implementation of both standard and new statistical methods. R does not limit user only to operation related only to these functions. This program has many benefits over other similar programs: it is free and, as an open source, constantly updated; it has built-in help system; the R language is easy to extend with user-written functions. The significance of the work is calculation of time to failure or reliability in a new way, using statistic. Another advantage of this calculation is that there is no need for technical details and it can be implemented in any part for which we need to know time to fail in order to have appropriate maintenance, but also to maximize usage and minimize costs. In this case, calculations have been made on diesel generator fans but the same principle can be applied to any other part. The data for this paper came from a field engineering study of the time to failure of diesel generator fans. The ultimate goal was to decide whether or not to replace the working fans with a higher quality fan to prevent future failures. Seventy generators were studied. For each one, the number of hours of running time from its first being put into service until fan failure or until the end of the study (whichever came first) was recorded. Dataset consists of two variables: hours and status. Hours show the time of each fan working and status shows the event: 1- failed, 0- censored data. Censored data represent cases when we cannot track the specific case, so it could fail or success. Gaining the result by using R was easy and quick. The program will take into consideration censored data and include this into the results. This is not so easy in hand calculation. For the purpose of the paper results from R program have been compared to hand calculations in two different cases: censored data taken as a failure and censored data taken as a success. In all three cases, results are significantly different. If user decides to use the R for further calculations, it will give more precise results with work on censored data than the hand calculation.

Keywords: censored data, R statistical software, reliability analysis, time to failure

Procedia PDF Downloads 401
4178 Optimization of Fin Type and Fin per Inch on Heat Transfer and Pressure Drop of an Air Cooler

Authors: A. Falavand Jozaei, A. Ghafouri

Abstract:

Operation enhancement in an air cooler (heat exchanger) depends on the rate of heat transfer, and pressure drop. In this paper, for a given heat duty, study of the effects of FPI (fin per inch) and fin type (circular and hexagonal fins) on two parameters mentioned above is considered in an air cooler in Iran, Arvand petrochemical. A program in EES (Engineering Equations Solver) software moreover, Aspen B-JAC and HTFS+ software are used for this purpose to solve governing equations. At first the simulated results obtained from this program is compared to the experimental data for two cases of FPI. The effects of FPI from 3 to 15 over heat transfer (Q) to pressure drop ratio (Q/Δp ratio). This ratio is one of the main parameters in design, rating, and simulation heat exchangers. The results show that heat transfer (Q) and pressure drop increase with increasing FPI (fin per inch) steadily, and the Q/Δp ratio increases to FPI = 12 (for circular fins about 47% and for hexagonal fins about 69%) and then decreased gradually to FPI = 15 (for circular fins about 5% and for hexagonal fins about 8%), and Q/Δp ratio is maximum at FPI = 12. The FPI value selection between 8 and 12 obtained as a result to optimum heat transfer to pressure drop ratio. Also by contrast, between circular and hexagonal fins results, the Q/Δp ratio of hexagonal fins more than Q/Δp ratio of circular fins for FPI between 8 and 12 (optimum FPI).

Keywords: air cooler, circular and hexagonal fins, fin per inch, heat transfer and pressure drop

Procedia PDF Downloads 454
4177 Comparative Diagnostic Performance of Diffusion-Weighted Imaging Combined With Microcalcifications on Mammography for Discriminating Malignant From Benign Bi-rads 4 Lesions With the Kaiser Score

Authors: Wangxu Xia

Abstract:

BACKGROUND BI-RADS 4 lesions raise the possibility of malignancy that warrant further clinical and radiologic work-up. This study aimed to evaluate the predictive performance of diffusion-weighted imaging(DWI) and microcalcifications on mammography for predicting malignancy of BI-RADS 4 lesions. In addition, the predictive performance of DWI combined with microcalcifications was alsocompared with the Kaiser score. METHODS During January 2021 and June 2023, 144 patients with 178 BI-RADS 4 lesions underwent conventional MRI, DWI, and mammography were included. The lesions were dichotomized intobenign or malignant according to the pathological results from core needle biopsy or surgical mastectomy. DWI was performed with a b value of 0 and 800s/mm2 and analyzed using theapparent diffusion coefficient, and a Kaiser score > 4 was considered to suggest malignancy. Thediagnostic performances for various diagnostic tests were evaluated with the receiver-operatingcharacteristic (ROC) curve. RESULTS The area under the curve (AUC) for DWI was significantly higher than that of the of mammography (0.86 vs 0.71, P<0.001), but was comparable with that of the Kaiser score (0.86 vs 0.84, P=0.58). However, the AUC for DWI combined with mammography was significantly highthan that of the Kaiser score (0.93 vs 0.84, P=0.007). The sensitivity for discriminating malignant from benign BI-RADS 4 lesions was highest at 89% for Kaiser score, but the highest specificity of 83% can be achieved with DWI combined with mammography. CONCLUSION DWI combined with microcalcifications on mammography could discriminate malignant BI-RADS4 lesions from benign ones with a high AUC and specificity. However, Kaiser score had a better sensitivity for discrimination.

Keywords: MRI, DWI, mammography, breast disease

Procedia PDF Downloads 59
4176 Risk Based Building Information Modeling (BIM) for Urban Infrastructure Transportation Project

Authors: Debasis Sarkar

Abstract:

Building Information Modeling (BIM) is a holistic documentation process for operational visualization, design coordination, estimation and project scheduling. BIM software defines objects parametrically and it is a tool for virtual reality. Primary advantage of implementing BIM is the visual coordination of the building structure and systems such as Mechanical, Electrical and Plumbing (MEP) and it also identifies the possible conflicts between the building systems. This paper is an attempt to develop a risk based BIM model which would highlight the primary advantages of application of BIM pertaining to urban infrastructure transportation project. It has been observed that about 40% of the Architecture, Engineering and Construction (AEC) companies use BIM but primarily for their outsourced projects. Also, 65% of the respondents agree that BIM would be used quiet strongly for future construction projects in India. The 3D models developed with Revit 2015 software would reduce co-ordination problems amongst the architects, structural engineers, contractors and building service providers (MEP). Integration of risk management along with BIM would provide enhanced co-ordination, collaboration and high probability of successful completion of the complex infrastructure transportation project within stipulated time and cost frame.

Keywords: building information modeling (BIM), infrastructure transportation, project risk management, underground metro rail

Procedia PDF Downloads 310
4175 Using Electrical Impedance Tomography to Control a Robot

Authors: Shayan Rezvanigilkolaei, Shayesteh Vefaghnematollahi

Abstract:

Electrical impedance tomography is a non-invasive medical imaging technique suitable for medical applications. This paper describes an electrical impedance tomography device with the ability to navigate a robotic arm to manipulate a target object. The design of the device includes various hardware and software sections to perform medical imaging and control the robotic arm. In its hardware section an image is formed by 16 electrodes which are located around a container. This image is used to navigate a 3DOF robotic arm to reach the exact location of the target object. The data set to form the impedance imaging is obtained by having repeated current injections and voltage measurements between all electrode pairs. After performing the necessary calculations to obtain the impedance, information is transmitted to the computer. This data is fed and then executed in MATLAB which is interfaced with EIDORS (Electrical Impedance Tomography Reconstruction Software) to reconstruct the image based on the acquired data. In the next step, the coordinates of the center of the target object are calculated by image processing toolbox of MATLAB (IPT). Finally, these coordinates are used to calculate the angles of each joint of the robotic arm. The robotic arm moves to the desired tissue with the user command.

Keywords: electrical impedance tomography, EIT, surgeon robot, image processing of electrical impedance tomography

Procedia PDF Downloads 272
4174 Phylogenetic Studies of Six Egyptian Sheep Breeds Using Cytochrome B

Authors: Othman Elmahdy Othman, Agnés Germot, Daniel Petit, Muhammad Khodary, Abderrahman Maftah

Abstract:

Recently, the control (D-loop) and cytochrome b (Cyt b) regions of mtDNA have received more attention due to their role in the genetic diversity and phylogenetic studies in different livestock which give important knowledge towards the genetic resource conservation. Studies based on sequencing of sheep mitochondrial DNA showed that there are five maternal lineages in the world for domestic sheep breeds; A, B, C, D and E. By using cytochrome B sequencing, we aimed to clarify the genetic affinities and phylogeny of six Egyptian sheep breeds. Blood samples were collected from 111 animals belonging to six Egyptian sheep breeds; Barki, Rahmani, Ossimi, Saidi, Sohagi and Fallahi. The total DNA was extracted and the specific primers were used for conventional PCR amplification of the cytochrome B region of mtDNA. PCR amplified products were purified and sequenced. The alignment of sequences was done using BioEdit software and DnaSP 5.00 software was used to identify the sequence variation and polymorphic sites in the aligned sequences. The result showed that the presence of 39 polymorphic sites leading to the formation of 29 haplotypes. The haplotype diversity in six tested breeds ranged from 0.643 in Rahmani breed to 0.871 in Barki breed. The lowest genetic distance was observed between Rahmani and Saidi (D: 1.436 and Dxy: 0.00127) while the highest distance was observed between Ossimi and Sohagi (D: 6.050 and Dxy: 0.00534). Neighbour-joining (Phylogeny) tree was constructed using Mega 5.0 software. The sequences of 111 analyzed samples were aligned with references sequences of different haplogroups; A, B, C, D and E. The phylogeny result showed the presence of four haplogroups; HapA, HapB, HapC and HapE in the examined samples whereas the haplogroup D was not found. The result showed that 88 out of 111 tested animals cluster with haplogroup B (79.28%), whereas 12 tested animals cluster with haplogroup A (10.81%), 10 animals cluster with haplogroup C (9.01%) and one animal belongs to haplogroup E (0.90%).

Keywords: phylogeny, genetic biodiversity, MtDNA, cytochrome B, Egyptian sheep

Procedia PDF Downloads 347
4173 The Design of Intelligent Passenger Organization System for Metro Stations Based on Anylogic

Authors: Cheng Zeng, Xia Luo

Abstract:

Passenger organization has always been an essential part of China's metro operation and management. Facing the massive passenger flow, stations need to improve their intelligence and automation degree by an appropriate integrated system. Based on the existing integrated supervisory control system (ISCS) and simulation software (Anylogic), this paper designs an intelligent passenger organization system (IPOS) for metro stations. Its primary function includes passenger information acquisition, data processing and computing, visualization management, decision recommendations, and decision response based on interlocking equipment. For this purpose, the logical structure and intelligent algorithms employed are particularly devised. Besides, the structure diagram of information acquisition and application module, the application of Anylogic, the case library's function process are all given by this research. Based on the secondary development of Anylogic and existing technologies like video recognition, the IPOS is supposed to improve the response speed and address capacity in the face of emergent passenger flow of metro stations.

Keywords: anylogic software, decision-making support system, intellectualization, ISCS, passenger organization

Procedia PDF Downloads 176