Search results for: Finnegan neonatal abstinence syndrome tool
3366 Urban Growth Prediction Using Artificial Neural Networks in Athens, Greece
Authors: Dimitrios Triantakonstantis, Demetris Stathakis
Abstract:
Urban areas have been expanded throughout the globe. Monitoring and modeling urban growth have become a necessity for a sustainable urban planning and decision making. Urban prediction models are important tools for analyzing the causes and consequences of urban land use dynamics. The objective of this research paper is to analyze and model the urban change, which has been occurred from 1990 to 2000 using CORINE land cover maps. The model was developed using drivers of urban changes (such as road distance, slope, etc.) under an Artificial Neural Network modeling approach. Validation was achieved using a prediction map for 2006 which was compared with a real map of Urban Atlas of 2006. The accuracy produced a Kappa index of agreement of 0,639 and a value of Cramer's V of 0,648. These encouraging results indicate the importance of the developed urban growth prediction model which using a set of available common biophysical drivers could serve as a management tool for the assessment of urban change.Keywords: artificial neural networks, CORINE, urban atlas, urban growth prediction
Procedia PDF Downloads 5293365 Application All Digits Number Benford Law in Financial Statement
Authors: Teguh Sugiarto
Abstract:
Background: The research aims to explore if there is fraud in a financial statement, use the Act stated that Benford's distribution all digits must compare the number will follow the trend of lower number. Research methods: This research uses all the analysis number being in Benford's law. After receiving the results of the analysis of all the digits, the author makes a distinction between implementation using the scale above and below 5%, the rate of occurrence of difference. With the number which have differences in the range of 5%, then can do the follow-up and the detection of the onset of fraud against the financial statements. The findings: From the research that has been done can be drawn the conclusion that the average of all numbers appear in the financial statements, and compare the rates of occurrence of numbers according to the characteristics of Benford's law. About the existence of errors and fraud in the financial statements of PT medco Energy Tbk did not occur. Conclusions: The study concludes that Benford's law can serve as indicator tool in detecting the possibility of in financial statements to case studies of PT Medco Energy Tbk for the fiscal year 2000-2010.Keywords: Benford law, first digits, all digits number Benford law, financial statement
Procedia PDF Downloads 2393364 Medical Advances in Diagnosing Neurological and Genetic Disorders
Authors: Simon B. N. Thompson
Abstract:
Retinoblastoma is a rare type of childhood genetic cancer that affects children worldwide. The diagnosis is often missed due to lack of education and difficulty in presentation of the tumor. Frequently, the tumor on the retina is noticed by photography when the red-eye flash, commonly seen in normal eyes, is not produced. Instead, a yellow or white colored patch is seen or the child has a noticeable strabismus. Early detection can be life-saving though often results in removal of the affected eye. Remaining functioning in the healthy eye when the child is young has resulted in super-vision and high or above-average intelligence. Technological advancement of cameras has helped in early detection. Brain imaging has also made possible early detection of neurological diseases and, together with the monitoring of cortisol levels and yawning frequency, promises to be the next new early diagnostic tool for the detection of neurological diseases where cortisol insufficiency is particularly salient, such as multiple sclerosis and Cushing’s disease.Keywords: cortisol, neurological disease, retinoblastoma, Thompson cortisol hypothesis, yawning
Procedia PDF Downloads 3863363 Optimizing Human Diet Problem Using Linear Programming Approach: A Case Study
Authors: P. Priyanka, S. Shruthi, N. Guruprasad
Abstract:
Health is a common theme in most cultures. In fact all communities have their concepts of health, as part of their culture. Health continues to be a neglected entity. Planning of Human diet should be done very careful by selecting the food items or groups of food items also the composition involved. Low price and good taste of foods are regarded as two major factors for optimal human nutrition. Linear programming techniques have been extensively used for human diet formulation for quiet good number of years. Through the process, we mainly apply “The Simplex Method” which is a very useful statistical tool based on the theorem of Elementary Row Operation from Linear Algebra and also incorporate some other necessary rules set by the Simplex Method to help solve the problem. The study done by us is an attempt to develop a programming model for optimal planning and best use of nutrient ingredients.Keywords: diet formulation, linear programming, nutrient ingredients, optimization, simplex method
Procedia PDF Downloads 5583362 The End Is Just the Beginning: The Importance of Project Post-Implementation Reviews
Authors: Catalin-Teodor Dogaru, Ana-Maria Dogaru
Abstract:
Success means different things to different people. For us, project managers, it becomes even harder to find a definition. Many factors have to be included in the evaluation. Moreover, literature is not very helpful, lacking consensus and neutrality. Post-implementation reviews (PIR) can be an efficient tool in evaluating how things worked on a certain project. Despite the visible progress, PIR is not a very detailed subject yet and there is not a common understanding in this matter. This may be the reason that some organizations include it in the projects’ lifecycle and some do not. Through this paper, we point out the reasons why all project managers should pay proper attention to this important step and to the elements, which can be assessed, beside the already famous triple constraints: cost, budget, and time. It is essential to take notice that PIR is not a checklist. It brings the edge in eliminating subjectivity and judging projects based on actual proof. Based on our experience, our success indicator model, presented in this paper, contributes to the success of the project! In the same time, it increases trust among customers who will perceive success more objectively.Keywords: project, post implementation, review, success, indicators
Procedia PDF Downloads 3713361 DURAFILE: A Collaborative Tool for Preserving Digital Media Files
Authors: Santiago Macho, Miquel Montaner, Raivo Ruusalepp, Ferran Candela, Xavier Tarres, Rando Rostok
Abstract:
During our lives, we generate a lot of personal information such as photos, music, text documents and videos that link us with our past. This data that used to be tangible is now digital information stored in our computers, which implies a software dependence to make them accessible in the future. Technology, however, constantly evolves and goes through regular shifts, quickly rendering various file formats obsolete. The need for accessing data in the future affects not only personal users but also organizations. In a digital environment, a reliable preservation plan and the ability to adapt to fast changing technology are essential for maintaining data collections in the long term. We present in this paper the European FP7 project called DURAFILE that provides the technology to preserve media files for personal users and organizations while maintaining their quality.Keywords: artificial intelligence, digital preservation, social search, digital preservation plans
Procedia PDF Downloads 4453360 A Study on the Development of Social Participation Activity Scale for the Elderly
Authors: Young-Kwang Lee, Eun-Gu Ji, Min-Joo Kim, Seung-Jae Oh
Abstract:
The purpose of this study is to develop a social participation activity scale for the elderly. As a result of exploratory factor analysis, confirmatory factor analysis was conducted using maximum likelihood method using bundled items. In conclusion, thirteen items of social participation activity scale seemed appropriate. Finally, convergent validity and discriminant validity were verified on the scale with the fit. The convergent validity was based on the variance extracted value. In other words, the hypothesis that the variables are the same is rejected and the validity is confirmed. This study extensively considered the measurement items of the social participation activity scale used to measure social participation activities of the elderly. In the future, it will be meaningful that it can be used as a tool to verify the effectiveness of services in organizations that provide social welfare services to elderly people such as comprehensive social welfare centers and the elderly comprehensive social welfare centers.Keywords: elderly, social participation, scale development, validity
Procedia PDF Downloads 1893359 Integration of Internet-Accessible Resources in the Field of Mobile Robots
Authors: B. Madhevan, R. Sakkaravarthi, R. Diya
Abstract:
The number and variety of mobile robot applications are increasing day by day, both in an industry and in our daily lives. First developed as a tool, nowadays mobile robots can be integrated as an entity in Internet-accessible resources. The present work is organized around four potential resources such as cloud computing, Internet of things, Big data analysis and Co-simulation. Further, the focus relies on integrating, analyzing and discussing the need for integrating Internet-accessible resources and the challenges deriving from such integration, and how these issues have been tackled. Hence, the research work investigates the concepts of the Internet-accessible resources from the aspect of the autonomous mobile robots with an overview of the performances of the currently available database systems. IaR is a world-wide network of interconnected objects, can be considered an evolutionary process in mobile robots. IaR constitutes an integral part of future Internet with data analysis, consisting of both physical and virtual things.Keywords: internet-accessible resources, cloud computing, big data analysis, internet of things, mobile robot
Procedia PDF Downloads 3893358 Developing Digital Competencies in Aboriginal Students through University-College Partnerships
Authors: W. S. Barber, S. L. King
Abstract:
This paper reports on a pilot project to develop a collaborative partnership between a community college in rural northern Ontario, Canada, and an urban university in the greater Toronto area in Oshawa, Canada. Partner institutions will collaborate to address learning needs of university applicants whose goals are to attain an undergraduate university BA in Educational Studies and Digital Technology degree, but who may not live in a geographical location that would facilitate this pathways process. The UOIT BA degree is attained through a 2+2 program, where students with a 2 year college diploma or equivalent can attain a four year undergraduate degree. The goals reported on the project are as: 1. Our aim is to expand the BA program to include an additional stream which includes serious educational games, simulations and virtual environments, 2. Develop fully (using both synchronous and asynchronous technologies) online learning modules for use by university applicants who otherwise are not geographically located close to a physical university site, 3. Assess the digital competencies of all students, including members of local, distance and Indigenous communities using a validated tool developed and tested by UOIT across numerous populations. This tool, the General Technical Competency Use and Scale (GTCU) will provide the collaborating institutions with data that will allow for analyzing how well students are prepared to succeed in fully online learning communities. Philosophically, the UOIT BA program is based on a fully online learning communities model (FOLC) that can be accessed from anywhere in the world through digital learning environments via audio video conferencing tools such as Adobe Connect. It also follows models of adult learning and mobile learning, and makes a university degree accessible to the increasing demographic of adult learners who may use mobile devices to learn anywhere anytime. The program is based on key principles of Problem Based Learning, allowing students to build their own understandings through the co-design of the learning environment in collaboration with the instructors and their peers. In this way, this degree allows students to personalize and individualize the learning based on their own culture, background and professional/personal experiences. Using modified flipped classroom strategies, students are able to interrogate video modules on their own time in preparation for one hour discussions occurring in video conferencing sessions. As a consequence of the program flexibility, students may continue to work full or part time. All of the partner institutions will co-develop four new modules, administer the GTCU and share data, while creating a new stream of the UOIT BA degree. This will increase accessibility for students to bridge from community colleges to university through a fully digital environment. We aim to work collaboratively with Indigenous elders, community members and distance education instructors to increase opportunities for more students to attain a university education.Keywords: aboriginal, college, competencies, digital, universities
Procedia PDF Downloads 2153357 Vision and Challenges of Developing VR-Based Digital Anatomy Learning Platforms and a Solution Set for 3D Model Marking
Authors: Gizem Kayar, Ramazan Bakir, M. Ilkay Koşar, Ceren U. Gencer, Alperen Ayyildiz
Abstract:
Anatomy classes are crucial for general education of medical students, whereas learning anatomy is quite challenging and requires memorization of thousands of structures. In traditional teaching methods, learning materials are still based on books, anatomy mannequins, or videos. This results in forgetting many important structures after several years. However, more interactive teaching methods like virtual reality, augmented reality, gamification, and motion sensors are becoming more popular since such methods ease the way we learn and keep the data in mind for longer terms. During our study, we designed a virtual reality based digital head anatomy platform to investigate whether a fully interactive anatomy platform is effective to learn anatomy and to understand the level of teaching and learning optimization. The Head is one of the most complicated human anatomy structures, with thousands of tiny, unique structures. This makes the head anatomy one of the most difficult parts to understand during class sessions. Therefore, we developed a fully interactive digital tool with 3D model marking, quiz structures, 2D/3D puzzle structures, and VR support so as to integrate the power of VR and gamification. The project has been developed in Unity game engine with HTC Vive Cosmos VR headset. The head anatomy 3D model has been selected with full skeletal, muscular, integumentary, head, teeth, lymph, and vein system. The biggest issue during the development was the complexity of our model and the marking of it in the 3D world system. 3D model marking requires to access to each unique structure in the counted subsystems which means hundreds of marking needs to be done. Some parts of our 3D head model were monolithic. This is why we worked on dividing such parts to subparts which is very time-consuming. In order to subdivide monolithic parts, one must use an external modeling tool. However, such tools generally come with high learning curves, and seamless division is not ensured. Second option was to integrate tiny colliders to all unique items for mouse interaction. However, outside colliders which cover inner trigger colliders cause overlapping, and these colliders repel each other. Third option is using raycasting. However, due to its own view-based nature, raycasting has some inherent problems. As the model rotate, view direction changes very frequently, and directional computations become even harder. This is why, finally, we studied on the local coordinate system. By taking the pivot point of the model into consideration (back of the nose), each sub-structure is marked with its own local coordinate with respect to the pivot. After converting the mouse position to the world position and checking its relation with the corresponding structure’s local coordinate, we were able to mark all points correctly. The advantage of this method is its applicability and accuracy for all types of monolithic anatomical structures.Keywords: anatomy, e-learning, virtual reality, 3D model marking
Procedia PDF Downloads 1003356 An Embedded High Speed Adder for Arithmetic Computations
Authors: Kala Bharathan, R. Seshasayanan
Abstract:
In this paper, a 1-bit Embedded Logic Full Adder (EFA) circuit in transistor level is proposed, which reduces logic complexity, gives low power and high speed. The design is further extended till 64 bits. To evaluate the performance of EFA, a 16, 32, 64-bit both Linear and Square root Carry Select Adder/Subtractor (CSLAS) Structure is also proposed. Realistic testing of proposed circuits is done on 8 X 8 Modified Booth multiplier and comparison in terms of power and delay is done. The EFA is implemented for different multiplier architectures for performance parameter comparison. Overall delay for CSLAS is reduced to 78% when compared to conventional one. The circuit implementations are done on TSMC 28nm CMOS technology using Cadence Virtuoso tool. The EFA has power savings of up to 14% when compared to the conventional adder. The present implementation was found to offer significant improvement in terms of power and speed in comparison to other full adder circuits.Keywords: embedded logic, full adder, pdp, xor gate
Procedia PDF Downloads 4483355 A New Tool for Global Optimization Problems: Cuttlefish Algorithm
Authors: Adel Sabry Eesa, Adnan Mohsin Abdulazeez Brifcani, Zeynep Orman
Abstract:
This paper presents a new meta-heuristic bio-inspired optimization algorithm which is called Cuttlefish Algorithm (CFA). The algorithm mimics the mechanism of color changing behavior of the cuttlefish to solve numerical global optimization problems. The colors and patterns of the cuttlefish are produced by reflected light from three different layers of cells. The proposed algorithm considers mainly two processes: reflection and visibility. Reflection process simulates light reflection mechanism used by these layers, while visibility process simulates visibility of matching patterns of the cuttlefish. To show the effectiveness of the algorithm, it is tested with some other popular bio-inspired optimization algorithms such as Genetic Algorithms (GA), Particle Swarm Optimization (PSO) and Bees Algorithm (BA) that have been previously proposed in the literature. Simulations and obtained results indicate that the proposed CFA is superior when compared with these algorithms.Keywords: Cuttlefish Algorithm, bio-inspired algorithms, optimization, global optimization problems
Procedia PDF Downloads 5643354 The Influences of Marketplace Knowledge, General Product Class Knowledge, and Knowledge in Meat Product with Traceability on Trust in Meat Traceability
Authors: Kawpong Polyorat
Abstract:
Since the outbreak of mad cow disease and bird flu, consumers have become more concerned with meat quality and safety. As a result, meat traceability is adopted as one approach to handle consumers’ concern in this issue. Nevertheless, in Thailand, meat traceability is rarely used as a marketing tool to persuade consumers. As a consequence, the present study attempts to understand consumer trust in the meat traceability system by conducting a study in this country to examine the impact of three types of consumer knowledge on this trust. The study results reveal that out of three types of consumer knowledge, marketplace knowledge was the sole predictor of consumer trust in meat traceability and it has a positive influence. General product class knowledge and knowledge in meat products with traceability, however, did not significantly influence consumer trust. The research results provide several implications and directions for future study.Keywords: consumer knowledge, marketing, product knowledge, traceability
Procedia PDF Downloads 3253353 Active Learning Based on Science Experiments to Improve Scientific Literacy
Authors: Kunihiro Kamataki
Abstract:
In this study, active learning based on simple science experiments was developed in a university class of the freshman, in order to improve their scientific literacy. Through the active learning based on simple experiments of generation of cloud in a plastic bottle, students increased the interest in the global atmospheric problem and were able to discuss and find solutions about this problem positively from various viewpoints of the science technology, the politics, the economy, the diplomacy and the relations among nations. The results of their questionnaires and free descriptions of this class indicate that they improve the scientific literacy and motivations of other classroom lectures to acquire knowledge. It is thus suggested that the science experiment is strong tool to improve their intellectual curiosity rapidly and the connections that link the impression of science experiment and their interest of the social problem is very important to enhance their learning effect in this education.Keywords: active learning, scientific literacy, simple scientific experiment, university education
Procedia PDF Downloads 2613352 Mean Shift-Based Preprocessing Methodology for Improved 3D Buildings Reconstruction
Authors: Nikolaos Vassilas, Theocharis Tsenoglou, Djamchid Ghazanfarpour
Abstract:
In this work we explore the capability of the mean shift algorithm as a powerful preprocessing tool for improving the quality of spatial data, acquired from airborne scanners, from densely built urban areas. On one hand, high resolution image data corrupted by noise caused by lossy compression techniques are appropriately smoothed while at the same time preserving the optical edges and, on the other, low resolution LiDAR data in the form of normalized Digital Surface Map (nDSM) is upsampled through the joint mean shift algorithm. Experiments on both the edge-preserving smoothing and upsampling capabilities using synthetic RGB-z data show that the mean shift algorithm is superior to bilateral filtering as well as to other classical smoothing and upsampling algorithms. Application of the proposed methodology for 3D reconstruction of buildings of a pilot region of Athens, Greece results in a significant visual improvement of the 3D building block model.Keywords: 3D buildings reconstruction, data fusion, data upsampling, mean shift
Procedia PDF Downloads 3153351 Implementation of Cloud Customer Relationship Management in Banking Sector: Strategies, Benefits and Challenges
Authors: Ngoc Dang Khoa Nguyen, Imran Ali
Abstract:
The cloud customer relationship management (CRM) has emerged as an innovative tool to augment the customer satisfaction and performance of banking systems. Cloud CRM allows to collect, analyze and utilize customer-associated information and update the systems, thereby offer superior customer service. Cloud technologies have invaluable potential to ensure innovative customer experiences, successful collaboration, enhanced speed to marketplace and IT effectiveness. As such, many leading banks have been attracted towards adoption of such innovative and customer-driver solutions to revolutionize their existing business models. Chief Information Officers (CIOs) are already implemented or in the process of implementation of cloud CRM. However, many organizations are still reluctant to take such initiative due to the lack of information on the factors influencing its implementation. This paper, therefore, aims to delve into the strategies, benefits and challenges intertwined in the implementation of Cloud CRM in banking sector and provide reliable solutions.Keywords: banking sector, cloud computing, cloud CRM, strategy
Procedia PDF Downloads 1663350 Sand Production Modelled with Darcy Fluid Flow Using Discrete Element Method
Authors: M. N. Nwodo, Y. P. Cheng, N. H. Minh
Abstract:
In the process of recovering oil in weak sandstone formations, the strength of sandstones around the wellbore is weakened due to the increase of effective stress/load from the completion activities around the cavity. The weakened and de-bonded sandstone may be eroded away by the produced fluid, which is termed sand production. It is one of the major trending subjects in the petroleum industry because of its significant negative impacts, as well as some observed positive impacts. For efficient sand management therefore, there has been need for a reliable study tool to understand the mechanism of sanding. One method of studying sand production is the use of the widely recognized Discrete Element Method (DEM), Particle Flow Code (PFC3D) which represents sands as granular individual elements bonded together at contact points. However, there is limited knowledge of the particle-scale behavior of the weak sandstone, and the parameters that affect sanding. This paper aims to investigate the reliability of using PFC3D and a simple Darcy flow in understanding the sand production behavior of a weak sandstone. An isotropic tri-axial test on a weak oil sandstone sample was first simulated at a confining stress of 1MPa to calibrate and validate the parallel bond models of PFC3D using a 10m height and 10m diameter solid cylindrical model. The effect of the confining stress on the number of bonds failure was studied using this cylindrical model. With the calibrated data and sample material properties obtained from the tri-axial test, simulations without and with fluid flow were carried out to check on the effect of Darcy flow on bonds failure using the same model geometry. The fluid flow network comprised of every four particles connected with tetrahedral flow pipes with a central pore or flow domain. Parametric studies included the effects of confining stress, and fluid pressure; as well as validating flow rate – permeability relationship to verify Darcy’s fluid flow law. The effect of model size scaling on sanding was also investigated using 4m height, 2m diameter model. The parallel bond model successfully calibrated the sample’s strength of 4.4MPa, showing a sharp peak strength before strain-softening, similar to the behavior of real cemented sandstones. There seems to be an exponential increasing relationship for the bigger model, but a curvilinear shape for the smaller model. The presence of the Darcy flow induced tensile forces and increased the number of broken bonds. For the parametric studies, flow rate has a linear relationship with permeability at constant pressure head. The higher the fluid flow pressure, the higher the number of broken bonds/sanding. The DEM PFC3D is a promising tool to studying the micromechanical behavior of cemented sandstones.Keywords: discrete element method, fluid flow, parametric study, sand production/bonds failure
Procedia PDF Downloads 3233349 Breast Cancer Prediction Using Score-Level Fusion of Machine Learning and Deep Learning Models
Authors: Sam Khozama, Ali M. Mayya
Abstract:
Breast cancer is one of the most common types in women. Early prediction of breast cancer helps physicians detect cancer in its early stages. Big cancer data needs a very powerful tool to analyze and extract predictions. Machine learning and deep learning are two of the most efficient tools for predicting cancer based on textual data. In this study, we developed a fusion model of two machine learning and deep learning models. To obtain the final prediction, Long-Short Term Memory (LSTM) and ensemble learning with hyper parameters optimization are used, and score-level fusion is used. Experiments are done on the Breast Cancer Surveillance Consortium (BCSC) dataset after balancing and grouping the class categories. Five different training scenarios are used, and the tests show that the designed fusion model improved the performance by 3.3% compared to the individual models.Keywords: machine learning, deep learning, cancer prediction, breast cancer, LSTM, fusion
Procedia PDF Downloads 1633348 A Timed and Colored Petri Nets for Modeling and Verify Cloud System Elasticity
Authors: Walid Louhichi, Mouhebeddine Berrima, Narjes Ben Rajed
Abstract:
Elasticity is the essential property of cloud computing. As the name suggests, it constitutes the ability of a cloud system to adjust resource provisioning in relation to fluctuating workload. There are two types of elasticity operations, vertical and horizontal. In this work, we are interested in horizontal scaling, which is ensured by two mechanisms; scaling in and scaling out. Following the sizing of the system, we can adopt scaling in in the event of over-supply and scaling out in the event of under-supply. In this paper, we propose a formal model, based on colored and temporized Petri nets, for the modeling of the duplication and the removal of a virtual machine from a server. This model is based on formal Petri Nets modeling language. The proposed models are edited, verified, and simulated with two examples implemented in CPNtools, which is a modeling tool for colored and timed Petri nets.Keywords: cloud computing, elasticity, elasticity controller, petri nets, scaling in, scaling out
Procedia PDF Downloads 1543347 Application of Response Surface Methodology (RSM) for Optimization of Fluoride Removal by Using Banana Peel
Authors: Pallavi N., Gayatri Jadhav
Abstract:
Good quality water is of prime importance for a healthy living. Fluoride is one such mineral present in water which causes many health problems in humans and specially children. Fluoride is said to be a double edge sword because lesser and higher concentration of fluoride in drinking water can cause both dental and skeletal fluorosis. Fluoride is one of the important mineral usually present at a higher concentration in ground water. There are many researches being carried out for defluoridation method. In the present research, fluoride removal is demonstrated using banana peel which is a biowaste as a biocoagulant. Response Surface Methodology (RSM) is a statistical design tool which is used to design the experiment. Central Composite Design (CCD) was used to determine the influence of the pH and dosage of the coagulant on the optimal removal of fluoride from a simulated water sample. 895 of fluoride removal were obtained in a acidic pH range of 4 – 9 and bio coagulant dosage of dosage of 18 – 20mg/L.Keywords: Fluoride, Response Surface Methodology, Dosage, banana peel
Procedia PDF Downloads 1603346 Human Error Analysis in the USA Marine Accidents Reports
Authors: J. Sánchez-Beaskoetxea
Abstract:
The analysis of accidents, such as marine accidents, is one of the most useful instruments to avoid future accidents. In the case of marine accidents, from a simple collision of a small boat in a port to the wreck of a gigantic tanker ship, the study of the causes of the accidents is the basis of a great part of the marine international legislation. Some countries have official institutions who investigate all the accidents in which a ship with their flag is involved. In the case of the USA, the National Transportation Safety Board (NTSB) is responsible for these researches. The NTSB, after a deep investigation into each accident, publishes a Marine Accident Report with the possible cause of the accident. This paper analyses all the Marine Accident Reports published by the NTBS and focuses its attention especially in the Human Errors that led to reported accidents. In this research, the different Human Errors made by crew members are cataloged in 10 different groups. After a complete analysis of all the reports, the statistical analysis on the Human Errors typology in marine accidents is presented in order to use it as a tool to avoid the same errors in the future.Keywords: human error, marine accidents, ship crew, USA
Procedia PDF Downloads 4193345 Reducing Energy Consumption and GHG Emission by Integration of Flare Gas with Fuel Gas Network in Refinery
Authors: N. Tahouni, M. Gholami, M. H. Panjeshahi
Abstract:
Gas flaring is one of the most GHG emitting sources in the oil and gas industries. It is also a major way for wasting such an energy that could be better utilized and even generates revenue. Minimize flaring is an effective approach for reducing GHG emissions and also conserving energy in flaring systems. Integrating waste and flared gases into the fuel gas networks (FGN) of refineries is an efficient tool. A fuel gas network collects fuel gases from various source streams and mixes them in an optimal manner, and supplies them to different fuel sinks such as furnaces, boilers, turbines, etc. In this article we use fuel gas network model proposed by Hasan et al. as a base model and modify some of its features and add constraints on emission pollution by gas flaring to reduce GHG emissions as possible. Results for a refinery case study showed that integration of flare gas stream with waste and natural gas streams to construct an optimal FGN can significantly reduce total annualized cost and flaring emissions.Keywords: flaring, fuel gas network, GHG emissions, stream
Procedia PDF Downloads 3443344 Estimating Tree Height and Forest Classification from Multi Temporal Risat-1 HH and HV Polarized Satellite Aperture Radar Interferometric Phase Data
Authors: Saurav Kumar Suman, P. Karthigayani
Abstract:
In this paper the height of the tree is estimated and forest types is classified from the multi temporal RISAT-1 Horizontal-Horizontal (HH) and Horizontal-Vertical (HV) Polarised Satellite Aperture Radar (SAR) data. The novelty of the proposed project is combined use of the Back-scattering Coefficients (Sigma Naught) and the Coherence. It uses Water Cloud Model (WCM). The approaches use two main steps. (a) Extraction of the different forest parameter data from the Product.xml, BAND-META file and from Grid-xxx.txt file come with the HH & HV polarized data from the ISRO (Indian Space Research Centre). These file contains the required parameter during height estimation. (b) Calculation of the Vegetation and Ground Backscattering, Coherence and other Forest Parameters. (c) Classification of Forest Types using the ENVI 5.0 Tool and ROI (Region of Interest) calculation.Keywords: RISAT-1, classification, forest, SAR data
Procedia PDF Downloads 4073343 Rehabilitation of the Blind Using Sono-Visualization Tool
Authors: Ashwani Kumar
Abstract:
In human beings, eyes play a vital role. A very less research has been done for rehabilitation of blindness for the blind people. This paper discusses the work that helps blind people for recognizing the basic shapes of the objects like circle, square, triangle, horizontal lines, vertical lines, diagonal lines and the wave forms like sinusoidal, square, triangular etc. This is largely achieved by using a digital camera, which is used to capture the visual information present in front of the blind person and a software program, which achieves the image processing operations, and finally the processed image is converted into sound. After the sound generation process, the generated sound is fed to the blind person through headphones for visualizing the imaginary image of the object. For visualizing the imaginary image of the object, it needs to train the blind person. Various training process methods had been applied for recognizing the object.Keywords: image processing, pixel, pitch, loudness, sound generation, edge detection, brightness
Procedia PDF Downloads 3883342 Analyzing Essential Patents of Mobile Communication Based on Patent Portfolio: Case Study of Long Term Evolution-Advanced
Authors: Kujhin Jeong, Sungjoo Lee
Abstract:
In the past, cross-licensing was made up of various application or commercial patents. Today, cross-licensing is restricted to essential patents, which has emphasized their importance significantly. Literature has shown that patent portfolio provides information for patent protection or strategy decision-making, but little empirical research has found strategic tool of essential patents. This paper will highlight four types of essential patent portfolio and analysis about each strategy in the field of LTE-A. Specifically we collected essential patents of mobile communication company through ETSI (European Telecommunication Standards Institute) and build-up portfolio activity, concentration, diversity, and quality. Using these portfolios, we can understand each company’s strategic character about the technology of LTE-A and comparison analysis of financial results. Essential patents portfolio displays a mobile communication company’s strategy and its strategy’s impact on the performance of a company.Keywords: essential patent, portfolio, patent portfolio, essential patent portfolio
Procedia PDF Downloads 3943341 Efficient Subsurface Mapping: Automatic Integration of Ground Penetrating Radar with Geographic Information Systems
Authors: Rauf R. Hussein, Devon M. Ramey
Abstract:
Integrating Ground Penetrating Radar (GPR) with Geographic Information Systems (GIS) can provide valuable insights for various applications, such as archaeology, transportation, and utility locating. Although there has been progress toward automating the integration of GPR data with GIS, fully automatic integration has not been achieved yet. Additionally, manually integrating GPR data with GIS can be a time-consuming and error-prone process. In this study, actual, real-world GPR applications are presented, and a software named GPR-GIS 10 is created to interactively extract subsurface targets from GPR radargrams and automatically integrate them into GIS. With this software, it is possible to quickly and reliably integrate the two techniques to create informative subsurface maps. The results indicated that automatic integration of GPR with GIS can be an efficient tool to map and view any subsurface targets in their appropriate location in a 3D space with the needed precision. The findings of this study could help GPR-GIS integrators save time and reduce errors in many GPR-GIS applications.Keywords: GPR, GIS, GPR-GIS 10, drone technology, automation
Procedia PDF Downloads 923340 Advances in Food Processing Using Extrusion Technology
Authors: Javeed Akhtar, R. K. Pandey, Z. R. Azaz Ahmad Azad
Abstract:
For the purpose of making different uses of food material for the development of extruded foods are produced using single and twin extruders. Extrusion cooking is a useful and economical tool for processing of novel food. This high temperature, short time processing technology causes chemical and physical changes that alter the nutritional and physical quality of the product. Extrusion processing of food ingredients characteristically depends on associating process conditions that influence the product qualities. The process parameters are optimized for extrusion of food material in order to obtain the maximum nutritive value by inactivating the anti-nutritional factors. The processing conditions such as moisture content, temperature and time are controlled to avoid over heating or under heating which otherwise would result in a product of lower nutritional quality.Keywords: extrusion processing, single and twin extruder, operating condition of extruders and extruded novel foods, food and agricultural engineering
Procedia PDF Downloads 3823339 A Tool for Rational Assessment of Dynamic Trust in Networked Organizations
Authors: Simon Samwel Msanjila
Abstract:
Networked environments which provides platforms and environments for business organizations are configured in different forms depending on many factors including life time, member characteristics, communication structure, and business objectives, among others. With continuing advances in digital technologies the distance has become a less barrier for business minded collaboration among organizations. With the need and ease to make business collaborate nowadays organizations are sometimes forced to co-work with others that are either unknown or less known to them in terms of history and performance. A promising approach for sustaining established collaboration has been establishment of trust relationship among organizations based on assessed trustworthiness for each participating organization. It has been stated in research that trust in organization is dynamic and thus assessment of trust level must address such dynamic nature. This paper assess relevant aspects of trust and applies the concepts to propose a semi-automated system for assessing the Sustainability and Evolution of trust in organizations participating in specific objective in a networked organizations environment.Keywords: trust evolution, trust sustainability, networked organizations, dynamic trust
Procedia PDF Downloads 4313338 Accelerating Personalization Using Digital Tools to Drive Circular Fashion
Authors: Shamini Dhana, G. Subrahmanya VRK Rao
Abstract:
The fashion industry is advancing towards a mindset of zero waste, personalization, creativity, and circularity. The trend of upcycling clothing and materials into personalized fashion is being demanded by the next generation. There is a need for a digital tool to accelerate the process towards mass customization. Dhana’s D/Sphere fashion technology platform uses digital tools to accelerate upcycling. In essence, advanced fashion garments can be designed and developed via reuse, repurposing, recreating activities, and using existing fabric and circulating materials. The D/Sphere platform has the following objectives: to provide (1) An opportunity to develop modern fashion using existing, finished materials and clothing without chemicals or water consumption; (2) The potential for an everyday customer and designer to use the medium of fashion for creative expression; (3) A solution to address the global textile waste generated by pre- and post-consumer fashion; (4) A solution to reduce carbon emissions, water, and energy consumption with the participation of all stakeholders; (5) An opportunity for brands, manufacturers, retailers to work towards zero-waste designs and as an alternative revenue stream. Other benefits of this alternative approach include sustainability metrics, trend prediction, facilitation of disassembly and remanufacture deep learning, and hyperheuristics for high accuracy. A design tool for mass personalization and customization utilizing existing circulating materials and deadstock, targeted to fashion stakeholders will lower environmental costs, increase revenues through up to date upcycled apparel, produce less textile waste during the cut-sew-stitch process, and provide a real design solution for the end customer to be part of circular fashion. The broader impact of this technology will result in a different mindset to circular fashion, increase the value of the product through multiple life cycles, find alternatives towards zero waste, and reduce the textile waste that ends up in landfills. This technology platform will be of interest to brands and companies that have the responsibility to reduce their environmental impact and contribution to climate change as it pertains to the fashion and apparel industry. Today, over 70% of the $3 trillion fashion and apparel industry ends up in landfills. To this extent, the industry needs such alternative techniques to both address global textile waste as well as provide an opportunity to include all stakeholders and drive circular fashion with new personalized products. This type of modern systems thinking is currently being explored around the world by the private sector, organizations, research institutions, and governments. This technological innovation using digital tools has the potential to revolutionize the way we look at communication, capabilities, and collaborative opportunities amongst stakeholders in the development of new personalized and customized products, as well as its positive impacts on society, our environment, and global climate change.Keywords: circular fashion, deep learning, digital technology platform, personalization
Procedia PDF Downloads 663337 American Slavery and the Consciousness of Play
Authors: Janaka B. Lewis
Abstract:
“Narratives of Slavery and the Culture of Play” examines how play is discussed in early African American literature by both men and women to illustrate ways that they negotiated the hierarchy and oppression of enslavement. Reading narratives categorized as “slave narratives,” including those written by Frederick Douglass, Harriet Jacobs, and Olaudah Equiano, through the lens of play theory offers an illuminated analysis of the significance of play culture in these texts. It then reads late nineteenth-century play culture (or absence thereof) portrayed in literature as a lens for more contemporary African American oral and literary culture. These discussions of social constructions through literature bridge analyses of African American-authored texts and create a larger conversation about print media as a tool of activism and resistance. This essay also contributes to a larger body of analysis of nineteenth-century African American culture through literature.Keywords: childhood, slavery, consciousness of play, 19th century African American culture
Procedia PDF Downloads 501