Search results for: archiving
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 33

Search results for: archiving

33 Author Self-Archiving in Open Access Institutional Repositories for Awareness Creation in Universities

Authors: Kwame Kodua-Ntim

Abstract:

The study explored the authors self-archiving to create awareness of open-access institutional repositories in universities. The qualitative approach of the study was informed by the interpretive paradigm as well as the case research design. The target population for the study was all twelve (12) open-access institutional repositories managers and administrators purposively selected from the five (5) universities in Ghana. The universities were chosen since they were the only ones listed in the Directory of Open Access Repositories. Interviews were conducted using a semi-structured interview guide and data were analyzed using thematic analysis. The study revealed that academics had some information about self-archiving in open-access institutional repositories and university libraries with open-access institutional repositories were using DSpace software. Managers and administrators of open-access institutional repositories mediated content uploaded and believed that author self-archiving could improve awareness of open-access institutional repositories. The study recommended that universities should fully implement the author’s self-archiving protocol, and academics should be trained to be able to upload research works onto open-access institutional repositories. Furthermore, the university and university library should provide rigorous policies on author self-archiving and incentives for author self-archiving in the open access institutional repositories.

Keywords: author, awareness, institutional repositories, open access, open archive, self-archiving

Procedia PDF Downloads 43
32 Research Repository System (RRS) for Academics

Authors: Ajayi Olusola Olajide, O. Ojeyinka Taiwo, Adeolara Oluwawemimo Janet, Isheyemi Olufemi Gabriel, Lawal Muideen Adekunle

Abstract:

In an academic world where research work is the tool for promotion and elevation to higher cadres, the quest for a system that secure researchers’ work, monitor as well as alert researchers of pending academic research work, cannot be over-emphasized. This study describes how a research repository system for academics is designed. The invention further relates to a system for archiving any paperwork and journal that comprises of a database for storing all researches. It relates to a method for users to communicate through messages which will also allow reviewing all the messages. To create this research repository system, PHP and MySQL were married together for the system implementation.

Keywords: research, repository, academic, archiving, secure, system, implementation

Procedia PDF Downloads 550
31 Handling, Exporting and Archiving Automated Mineralogy Data Using TESCAN TIMA

Authors: Marek Dosbaba

Abstract:

Within the mining sector, SEM-based Automated Mineralogy (AM) has been the standard application for quickly and efficiently handling mineral processing tasks. Over the last decade, the trend has been to analyze larger numbers of samples, often with a higher level of detail. This has necessitated a shift from interactive sample analysis performed by an operator using a SEM, to an increased reliance on offline processing to analyze and report the data. In response to this trend, TESCAN TIMA Mineral Analyzer is designed to quickly create a virtual copy of the studied samples, thereby preserving all the necessary information. Depending on the selected data acquisition mode, TESCAN TIMA can perform hyperspectral mapping and save an X-ray spectrum for each pixel or segment, respectively. This approach allows the user to browse through elemental distribution maps of all elements detectable by means of energy dispersive spectroscopy. Re-evaluation of the existing data for the presence of previously unconsidered elements is possible without the need to repeat the analysis. Additional tiers of data such as a secondary electron or cathodoluminescence images can also be recorded. To take full advantage of these information-rich datasets, TIMA utilizes a new archiving tool introduced by TESCAN. The dataset size can be reduced for long-term storage and all information can be recovered on-demand in case of renewed interest. TESCAN TIMA is optimized for network storage of its datasets because of the larger data storage capacity of servers compared to local drives, which also allows multiple users to access the data remotely. This goes hand in hand with the support of remote control for the entire data acquisition process. TESCAN also brings a newly extended open-source data format that allows other applications to extract, process and report AM data. This offers the ability to link TIMA data to large databases feeding plant performance dashboards or geometallurgical models. The traditional tabular particle-by-particle or grain-by-grain export process is preserved and can be customized with scripts to include user-defined particle/grain properties.

Keywords: Tescan, electron microscopy, mineralogy, SEM, automated mineralogy, database, TESCAN TIMA, open format, archiving, big data

Procedia PDF Downloads 77
30 Proposal for an Inspection Tool for Damaged Structures after Disasters

Authors: Karim Akkouche, Amine Nekmouche, Leyla Bouzid

Abstract:

This study focuses on the development of a multifunctional Expert System (ES) called post-seismic damage inspection tool (PSDIT), a powerful tool which allows the evaluation, the processing, and the archiving of the collected data stock after earthquakes. PSDIT can be operated by two user types; an ordinary user (ingineer, expert, or architect) for the damage visual inspection and an administrative user for updating the knowledge and / or for adding or removing the ordinary user. The knowledge acquisition is driven by a hierarchical knowledge model, the Information from investigation reports and those acquired through feedback from expert / engineer questionnaires are part.

Keywords: .disaster, damaged structures, damage assessment, expert system

Procedia PDF Downloads 45
29 Proposal of a Damage Inspection Tool After Earthquakes: Case of Algerian Buildings

Authors: Akkouche Karim, Nekmouche Aghiles, Bouzid Leyla

Abstract:

This study focuses on the development of a multifunctional Expert System (ES) called post-seismic damage inspection tool (PSDIT), a powerful tool which allows the evaluation, the processing and the archiving of the collected data stock after earthquakes. PSDIT can be operated by two user types; an ordinary user (engineer, expert or architect) for the damage visual inspection and an administrative user for updating the knowledge and / or for adding or removing the ordinary user. The knowledge acquisition is driven by a hierarchical knowledge model, the Information from investigation reports and those acquired through feedback from expert / engineer questionnaires are part.

Keywords: buildings, earthquake, seismic damage, damage assessment, expert system

Procedia PDF Downloads 42
28 An Improvement of Multi-Label Image Classification Method Based on Histogram of Oriented Gradient

Authors: Ziad Abdallah, Mohamad Oueidat, Ali El-Zaart

Abstract:

Image Multi-label Classification (IMC) assigns a label or a set of labels to an image. The big demand for image annotation and archiving in the web attracts the researchers to develop many algorithms for this application domain. The existing techniques for IMC have two drawbacks: The description of the elementary characteristics from the image and the correlation between labels are not taken into account. In this paper, we present an algorithm (MIML-HOGLPP), which simultaneously handles these limitations. The algorithm uses the histogram of gradients as feature descriptor. It applies the Label Priority Power-set as multi-label transformation to solve the problem of label correlation. The experiment shows that the results of MIML-HOGLPP are better in terms of some of the evaluation metrics comparing with the two existing techniques.

Keywords: data mining, information retrieval system, multi-label, problem transformation, histogram of gradients

Procedia PDF Downloads 344
27 HIS Integration Systems Using Modality Worklist and DICOM

Authors: Kulvinder Singh Mann

Abstract:

The usability and simulation of information systems, known as Hospital Information System (HIS), Radiology Information System (RIS), and Picture Archiving, Communication System, for electronic medical records has shown a good impact for actors in the hospital. The objective is to help and make their work easier; such as for a nurse or administration staff to record the medical records of the patient, and for a patient to check their bill transparently. However, several limitations still exists on such area regarding the type of data being stored in the system, ability for data transfer, storage and protocols to support communication between medical devices and digital images. This paper reports the simulation result of integrating several systems to cope with those limitations by using the Modality Worklist and DICOM standard. It succeeds in documenting the reason of that failure so future research will gain better understanding and be able to integrate those systems.

Keywords: HIS, RIS, PACS, modality worklist, DICOM, digital images

Procedia PDF Downloads 281
26 Digital Technology Relevance in Archival and Digitising Practices in the Republic of South Africa

Authors: Tashinga Matindike

Abstract:

By means of definition, digital artworks encompass an array of artistic productions that are expressed in a technological form as an essential part of a creative process. Examples include illustrations, photos, videos, sculptures, and installations. Within the context of the visual arts, the process of repatriation involves the return of once-appropriated goods. Archiving denotes the preservation of a commodity for storage purposes in order to nurture its continuity. The aforementioned definitions form the foundation of the academic framework and premise of the argument, which is outlined in this paper. This paper aims to define, discuss and decipher the complexities involved in digitising artworks, whilst explaining the benefits of the process, particularly within the South African context, which is rich in tangible and intangible traditional cultural material, objects, and performances. With the internet having been introduced to the African Continent in the early 1990s, this new form of technology, in its own right, initiated a high degree of efficiency, which also resulted in the progressive transformation of computer-generated visual output. Subsequently, this caused a revolutionary influence on the manner in which technological software was developed and uterlised in art-making. Digital technology and the digitisation of creative processes then opened up new avenues of collating and recording information. One of the first visual artists to make use of digital technology software in his creative productions was United States-based artist John Whitney. His inventive work contributed greatly to the onset and development of digital animation. Comparable by technique and originality, South African contemporary visual artists who make digital artworks, both locally and internationally, include David Goldblatt, Katherine Bull, Fritha Langerman, David Masoga, Zinhle Sethebe, Alicia Mcfadzean, Ivan Van Der Walt, Siobhan Twomey, and Fhatuwani Mukheli. In conclusion, the main objective of this paper is to address the following questions: In which ways has the South African art community of visual artists made use of and benefited from technology, in its digital form, as a means to further advance creativity? What are the positive changes that have resulted in art production in South Africa since the onset and use of digital technological software? How has digitisation changed the manner in which we record, interpret, and archive both written and visual information? What is the role of South African art institutions in the development of digital technology and its use in the field of visual art. What role does digitisation play in the process of the repatriation of artworks and artefacts. The methodology in terms of the research process of this paper takes on a multifacted form, inclusive of data analysis of information attained by means of qualitative and quantitative approaches.

Keywords: digital art, digitisation, technology, archiving, transformation and repatriation

Procedia PDF Downloads 19
25 Greyscale: A Tree-Based Taxonomy for Grey Literature Published by Fisheries Agencies

Authors: Tatiana Tunon, Gottfried Pestal

Abstract:

Government agencies responsible for the management of fisheries resources publish many types of grey literature, and these materials are increasingly accessible to the public on agency websites. However, scope and quality vary considerably, and end-users need meta-data about the report series when deciding whether to use the information (e.g. apply the methods, include the results in a systematic review), or when prioritizing materials for archiving (e.g. library holdings, reference databases). A proposed taxonomy for these report series was developed based on a review of 41 report series from 6 government agencies in 4 countries (Canada, New Zealand, Scotland, and United States). Each report series was categorized according to multiple criteria describing peer-review process, content, and purpose. A robust classification tree was then fitted to these descriptions, and the resulting taxonomic groups were used to compare agency output from 4 countries using reports available in their online repositories.

Keywords: classification tree, fisheries, government, grey literature

Procedia PDF Downloads 241
24 A Study on How to Link BIM Services to Cloud Computing Architecture

Authors: Kim Young-Jin, Kim Byung-Kon

Abstract:

Although more efforts to expand the application of BIM (Building Information Modeling) technologies have be pursued in recent years than ever, it’s true that there have been various challenges in doing so, including a lack or absence of relevant institutions, lots of costs required to build BIM-related infrastructure, incompatible processes, etc. This, in turn, has led to a more prolonged delay in the expansion of their application than expected at an early stage. Especially, attempts to save costs for building BIM-related infrastructure and provide various BIM services compatible with domestic processes include studies to link between BIM and cloud computing technologies. Also in this study, the author attempted to develop a cloud BIM service operation model through analyzing the level of BIM applications for the construction sector and deriving relevant service areas, and find how to link BIM services to the cloud operation model, as through archiving BIM data and creating a revenue structure so that the BIM services may grow spontaneously, considering a demand for cloud resources.

Keywords: construction IT, BIM (building information modeling), cloud computing, BIM service based cloud computing

Procedia PDF Downloads 457
23 The Effectiveness of Using MS SharePoint for the Curriculum Repository System

Authors: Misook Ahn

Abstract:

This study examines the Institutional Curriculum Repository (ICR) developed with MS SharePoint. The purpose of using MS SharePoint is to organize, share, and manage the curriculum data. The ICR aims to build a centralized curriculum infrastructure, preserve all curriculum materials, and provide academic service to users (faculty, students, or other agencies). The ICR collection includes core language curriculum materials developed by each language school—foreign language textbooks, language survival kits, and audio files currently in or not in use at the schools. All core curriculum materials with audio and video files have been coded, collected, and preserved at the ICR. All metadata for the collected curriculum materials have been input by language, code, year, book type, level, user, version, and current status (in use/not in use). The qualitative content analysis, including the survey data, is used to evaluate the effectiveness of using MS SharePoint for the repository system. This study explains how to manage and preserve curriculum materials with MS SharePoint, along with challenges and suggestions for further research. This study will be beneficial to other universities or organizations considering archiving or preserving educational materials.

Keywords: digital preservation, ms sharepoint, repository, curriculum materials

Procedia PDF Downloads 61
22 Study on Carbon Nanostructures Influence on Changes in Static Friction Forces

Authors: Rafał Urbaniak, Robert Kłosowiak, Michał Ciałkowski, Jarosław Bartoszewicz

Abstract:

The Chair of Thermal Engineering at Poznan University of Technology has been conducted research works on the possibilities of using carbon nanostructures in energy and mechanics applications for a couple of years. Those studies have provided results in a form of co-operation with foreign research centres, numerous publications and patent applications. Authors of this paper have studied the influence of multi-walled carbon nanostructures on changes in static friction arising when steel surfaces were moved. Tests were made using the original test stand consisting of automatically controlled inclined plane driven by precise stepper motors. Computer program created in the LabView environment was responsible for monitoring of the stand operation, accuracy of measurements and archiving the obtained results. Such a solution enabled to obtain high accuracy and repeatability of all conducted experiments. Tests and analysis of the obtained results allowed us to determine how additional layers of carbon nanostructures influenced on changes of static friction coefficients. At the same time, we analyzed the potential possibilities of applying nanostructures under consideration in mechanics.

Keywords: carbon nanotubes, static friction, dynamic friction

Procedia PDF Downloads 279
21 Smartphone Photography in Urban China

Authors: Wen Zhang

Abstract:

The smartphone plays a significant role in media convergence, and smartphone photography is reconstructing the way we communicate and think. This article aims to explore the smartphone photography practices of urban Chinese smartphone users and images produced by smartphones from a techno-cultural perspective. The analysis consists of two types of data: One is a semi-structured interview of 21 participants, and the other consists of the images created by the participants. The findings are organised in two parts. The first part summarises the current tendencies of capturing, editing, sharing and archiving digital images via smartphones. The second part shows that food and selfie/anti-selfie are the preferred subjects of smartphone photographic images from a technical and multi-purpose perspective and demonstrates that screenshots and image texts are new genres of non-photographic images that are frequently made by smartphones, which contributes to improving operational efficiency, disseminating information and sharing knowledge. The analyses illustrate the positive impacts between smartphones and photography enthusiasm and practices based on the diffusion of innovation theory, which also makes us rethink the value of photographs and the practice of ‘photographic seeing’ from the screen itself.

Keywords: digital photography, image-text, media convergence, photographic- seeing, selfie/anti-selfie, smartphone, technological innovation

Procedia PDF Downloads 318
20 Design of an Instrumentation Setup and Data Acquisition System for a GAS Turbine Engine Using Suitable DAQ Software

Authors: Syed Nauman Bin Asghar Bukhari, Mohtashim Mansoor, Mohammad Nouman

Abstract:

Engine test-Bed system is a fundamental tool to measure dynamic parameters, economic performance, and reliability of an aircraft Engine, and its automation and accuracy directly influences the precision of acquired and analysed data. In this paper, we present the design of digital Data Acquisition (DAQ) system for a vintage aircraft engine test bed that lacks the capability of displaying all the analyzed parameters at one convenient location (one panel-one screen). Recording such measurements in the vintage test bed is not only time consuming but also prone to human errors. Digitizing such measurement system requires a Data Acquisition (DAQ) system capable of recording these parameters and displaying them on one screen-one panel monitor. The challenge in designing upgrade to the vintage systems arises with a need to build and integrate digital measurement system from scratch with a minimal budget and modifications to the existing vintage system. The proposed design not only displays all the key performance / maintenance parameters of the gas turbine engines for operator as well as quality inspector on separate screens but also records the data for further processing / archiving.

Keywords: Gas turbine engine, engine test cell, data acquisition, instrumentation

Procedia PDF Downloads 91
19 Sportband: An Idea for Workout Monitoring in Amateur and Recreational Sports

Authors: Kamila Mazur-Oleszczuk, Rafal Banasiuk, Dawid Krasnowski, Maciej Pek, Marcin Podgorski, Krzysztof Rykaczewski, Sabina Zoledowska, Dawid Nidzworski

Abstract:

Workout safety is one of the most significant challenges of recreational sports. Loss of water and electrolytes is a consequence of thermoregulatory sweating during exercise. The rate of sweat loss and its chemical composition can fluctuate within and among individuals. That is why we propose our sportband 'Flow' as a device for monitoring these parameters. 'Flow' consists of two parts: an intelligent module and a mobile application. The application allows verifying the training progress and data archiving. The sportband intelligent module includes temperature, heart rate and pulse measurement (non-invasive, continuous methods of workout monitoring). Apart from the standard components, the device will consist of a sweat composition analyzer situated in sportband intelligent module. Sweat is a water solution of numerous compounds such as ions (sodium up to 1609 µg/ml, potassium up to 274 µg/ml), lactic acid (skin pH is between 4.5 - 6) and a small amount of glucose. Awareness of sweat composition allows personalizing electrolyte intake after training. A comprehensive workout monitoring (sweat composition, heart rate, blood oxygen level) will provide improvement in the training routine and time management, which is our goal for the development of the sweat composition analyzer.

Keywords: flow, sportband, sweat, workout monitoring

Procedia PDF Downloads 116
18 Visual Template Detection and Compositional Automatic Regular Expression Generation for Business Invoice Extraction

Authors: Anthony Proschka, Deepak Mishra, Merlyn Ramanan, Zurab Baratashvili

Abstract:

Small and medium-sized businesses receive over 160 billion invoices every year. Since these documents exhibit many subtle differences in layout and text, extracting structured fields such as sender name, amount, and VAT rate from them automatically is an open research question. In this paper, existing work in template-based document extraction is extended, and a system is devised that is able to reliably extract all required fields for up to 70% of all documents in the data set, more than any other previously reported method. The approaches are described for 1) detecting through visual features which template a given document belongs to, 2) automatically generating extraction rules for a given new template by composing regular expressions from multiple components, and 3) computing confidence scores that indicate the accuracy of the automatic extractions. The system can generate templates with as little as one training sample and only requires the ground truth field values instead of detailed annotations such as bounding boxes that are hard to obtain. The system is deployed and used inside a commercial accounting software.

Keywords: data mining, information retrieval, business, feature extraction, layout, business data processing, document handling, end-user trained information extraction, document archiving, scanned business documents, automated document processing, F1-measure, commercial accounting software

Procedia PDF Downloads 94
17 Static Analysis of Security Issues of the Python Packages Ecosystem

Authors: Adam Gorine, Faten Spondon

Abstract:

Python is considered the most popular programming language and offers its own ecosystem for archiving and maintaining open-source software packages. This system is called the python package index (PyPI), the repository of this programming language. Unfortunately, one-third of these software packages have vulnerabilities that allow attackers to execute code automatically when a vulnerable or malicious package is installed. This paper contributes to large-scale empirical studies investigating security issues in the python ecosystem by evaluating package vulnerabilities. These provide a series of implications that can help the security of software ecosystems by improving the process of discovering, fixing, and managing package vulnerabilities. The vulnerable dataset is generated using the NVD, the national vulnerability database, and the Snyk vulnerability dataset. In addition, we evaluated 807 vulnerability reports in the NVD and 3900 publicly known security vulnerabilities in Python Package Manager (pip) from the Snyk database from 2002 to 2022. As a result, many Python vulnerabilities appear in high severity, followed by medium severity. The most problematic areas have been improper input validation and denial of service attacks. A hybrid scanning tool that combines the three scanners bandit, snyk and dlint, which provide a clear report of the code vulnerability, is also described.

Keywords: Python vulnerabilities, bandit, Snyk, Dlint, Python package index, ecosystem, static analysis, malicious attacks

Procedia PDF Downloads 92
16 Performance Comparison of Tablet Devices and Medical Diagnostic Display Devices Using Digital Object Patterns in PACS Environment

Authors: Yan-Lin Liu, Cheng-Ting Shih, Jay Wu

Abstract:

Tablet devices have been introduced into the medical environment in recent years. The performance of display can be varied based on the use of different hardware specifications and types of display technologies. Therefore, the differences between tablet devices and medical diagnostic LCDs have to be verified to ensure that image quality is not jeopardized for clinical diagnosis in a picture archiving and communication system (PACS). In this study, a set of randomized object test patterns (ROTPs) were developed, which included randomly located spheres in abdominal CT images. Five radiologists were asked to independently review the CT images on different generations of iPads and a diagnostic monochrome medical LCD monitor. Receiver operating characteristic (ROC) analysis was performed by using a five-point rating scale, and the average area under curve (AUC) and average reading time (ART) were calculated. The AUC values for the second generation iPad, iPad mini, iPad Air, and monochrome medical monitor were 0.712, 0.717, 0.725, and 0.740, respectively. The differences between iPads were not significant. The ARTs were 177 min and 127 min for iPad mini and medical LCD monitor, respectively. A significant difference appeared (p = 0.04). The results show that the iPads were slightly inferior to the monochrome medical LCD monitor. However, tablet devices possess advantages in portability and versatility, which can improve the convenience of rapid diagnosis and teleradiology. With advances in display technology, the applicability of tablet devices and mobile devices may be more diversified in PACS.

Keywords: tablet devices, PACS, receiver operating characteristic, LCD monitor

Procedia PDF Downloads 444
15 A Large Ion Collider Experiment (ALICE) Diffractive Detector Control System for RUN-II at the Large Hadron Collider

Authors: J. C. Cabanillas-Noris, M. I. Martínez-Hernández, I. León-Monzón

Abstract:

The selection of diffractive events in the ALICE experiment during the first data taking period (RUN-I) of the Large Hadron Collider (LHC) was limited by the range over which rapidity gaps occur. It would be possible to achieve better measurements by expanding the range in which the production of particles can be detected. For this purpose, the ALICE Diffractive (AD0) detector has been installed and commissioned for the second phase (RUN-II). Any new detector should be able to take the data synchronously with all other detectors and be operated through the ALICE central systems. One of the key elements that must be developed for the AD0 detector is the Detector Control System (DCS). The DCS must be designed to operate safely and correctly this detector. Furthermore, the DCS must also provide optimum operating conditions for the acquisition and storage of physics data and ensure these are of the highest quality. The operation of AD0 implies the configuration of about 200 parameters, from electronics settings and power supply levels to the archiving of operating conditions data and the generation of safety alerts. It also includes the automation of procedures to get the AD0 detector ready for taking data in the appropriate conditions for the different run types in ALICE. The performance of AD0 detector depends on a certain number of parameters such as the nominal voltages for each photomultiplier tube (PMT), their threshold levels to accept or reject the incoming pulses, the definition of triggers, etc. All these parameters define the efficiency of AD0 and they have to be monitored and controlled through AD0 DCS. Finally, AD0 DCS provides the operator with multiple interfaces to execute these tasks. They are realized as operating panels and scripts running in the background. These features are implemented on a SCADA software platform as a distributed control system which integrates to the global control system of the ALICE experiment.

Keywords: AD0, ALICE, DCS, LHC

Procedia PDF Downloads 271
14 A Medical Vulnerability Scoring System Incorporating Health and Data Sensitivity Metrics

Authors: Nadir A. Carreon, Christa Sonderer, Aakarsh Rao, Roman Lysecky

Abstract:

With the advent of complex software and increased connectivity, the security of life-critical medical devices is becoming an increasing concern, particularly with their direct impact on human safety. Security is essential, but it is impossible to develop completely secure and impenetrable systems at design time. Therefore, it is important to assess the potential impact on the security and safety of exploiting a vulnerability in such critical medical systems. The common vulnerability scoring system (CVSS) calculates the severity of exploitable vulnerabilities. However, for medical devices it does not consider the unique challenges of impacts to human health and privacy. Thus, the scoring of a medical device on which human life depends (e.g., pacemakers, insulin pumps) can score very low, while a system on which human life does not depend (e.g., hospital archiving systems) might score very high. In this paper, we propose a medical vulnerability scoring system (MVSS) that extends CVSS to address the health and privacy concerns of medical devices. We propose incorporating two new parameters, namely health impact, and sensitivity impact. Sensitivity refers to the type of information that can be stolen from the device, and health represents the impact on the safety of the patient if the vulnerability is exploited (e.g., potential harm, life-threatening). We evaluate fifteen different known vulnerabilities in medical devices and compare MVSS against two state-of-the-art medical device-oriented vulnerability scoring systems and the foundational CVSS.

Keywords: common vulnerability system, medical devices, medical device security, vulnerabilities

Procedia PDF Downloads 121
13 A Study on the Importance and Contributions of Transforming from OEM to ODM in Malaysian Furniture Industry

Authors: Nurul Ain Haron, Saiful Hazmi Bachek, Hafez Zainudin

Abstract:

This study is aimed to establish the importance and contribution of Original Design Manufacturing (ODM) in Malaysian Furniture Industry and to close the gap between the players in the industry. The study confirms that today’s furniture industry favor Original Equipment Manufacturing (OEM) basis. Thus, resulting in the local industry lacking the expertise of designing furniture to a state of no contest. This study method used consists of literature reviews, observation, questionnaire and sessions of interviews. The result shows that the public has from minimum to almost no knowledge of the term Original Design Manufacturing (ODM) concept and the impact towards our current future industry. The manufacturers however, prefers Original Equipment Manufacturing (OEM) concept due to its easy and fast investment returns with the need of product designing process, while the interviews carried out with the authorized council had some convincing urges of doing their part promoting the awareness through trainings and seminars. Findings show that, in the rush of archiving ODM status needs extensive cooperation from many parties that are authorized council, furniture manufacturers, designers and also the public perceptions of labeling local made goods as the black goat. The current mind set of OEM manufacturers need to be change for industry’s future. As Malaysia’s living status constantly increases, so will the demands of a better salary. If these current issues are not resolved, OEM international buyers will definitely have to shift to some other lower cost manufacturer like China or Taiwan. When vendors stopped to order, today’s OEM manufacturers will no longer exist in the future.

Keywords: design manufacturing, furniture design, original design manufacturing, original equipment manufacturing

Procedia PDF Downloads 411
12 Methodological Proposal, Archival Thesaurus in Colombian Sign Language

Authors: Pedro A. Medina-Rios, Marly Yolie Quintana-Daza

Abstract:

Having the opportunity to communicate in a social, academic and work context is very relevant for any individual and more for a deaf person when oral language is not their natural language, and written language is their second language. Currently, in Colombia, there is not a specialized dictionary for our best knowledge in sign language archiving. Archival is one of the areas that the deaf community has a greater chance of performing. Nourishing new signs in dictionaries for deaf people extends the possibility that they have the appropriate signs to communicate and improve their performance. The aim of this work was to illustrate the importance of designing pedagogical and technological strategies of knowledge management, for the academic inclusion of deaf people through proposals of lexicon in Colombian sign language (LSC) in the area of archival. As a method, the analytical study was used to identify relevant words in the technical area of the archival and its counterpart with the LSC, 30 deaf people, apprentices - students of the Servicio Nacional de Aprendizaje (SENA) in Documentary or Archival Management programs, were evaluated through direct interviews in LSC. For the analysis tools were maintained to evaluate correlation patterns and linguistic methods of visual, gestural analysis and corpus; besides, methods of linear regression were used. Among the results, significant data were found among the variables socioeconomic stratum, academic level, labor location. The need to generate new signals on the subject of the file to improve communication between the deaf person, listener and the sign language interpreter. It is concluded that the generation of new signs to nourish the LSC dictionary in archival subjects is necessary to improve the labor inclusion of deaf people in Colombia.

Keywords: archival, inclusion, deaf, thesaurus

Procedia PDF Downloads 243
11 Computed Tomography Guided Bone Biopsies: Experience at an Australian Metropolitan Hospital

Authors: K. Hinde, R. Bookun, P. Tran

Abstract:

Percutaneous CT guided biopsies provide a fast, minimally invasive, cost effective and safe method for obtaining tissue for histopathology and culture. Standards for diagnostic yield vary depending on whether the tissue is being obtained for histopathology or culture. We present a retrospective audit from Western Health in Melbourne Australia over a 12-month period which aimed to determine the diagnostic yield, technical success and complication rate for CT guided bone biopsies and identify factors affecting these results. The digital imaging storage program (Synapse Picture Archiving and Communication System – Fujifilm Australia) was analysed with key word searches from October 2015 to October 2016. Nineteen CT guided bone biopsies were performed during this time. The most common referring unit was oncology, work up imaging included CT, MRI, bone scan and PET scan. The complication rate was 0%, overall diagnostic yield was 74% with a technical success of 95%. When performing biopsies for histologic analysis diagnostic yield was 85% and when performing biopsies for bacterial culture diagnostic yield was 60%. There was no significant relationship identified between size of lesion, distance of lesion to skin, lesion appearance on CT, the number of samples taken or gauge of needle to diagnostic yield or technical success. CT guided bone biopsy at Western Health meets the standard reported at other major clinical centres for technical success and safety. It is a useful investigation in identification of primary malignancy in distal bone metastases.

Keywords: bone biopsy, computed tomography, core biopsy, histopathology

Procedia PDF Downloads 170
10 The Role of Pharmacist in The Community: A Study of Methanol Toxicity Disaster in Tripoli Libya During March 2013

Authors: Abdurrauf M. Gusbi, Mahmud H. Arhima, Abdurrahim A. Elouzi, Ebtisam A. Benomran, Salsabeela Elmezwghi, Aram Elhatan, Nafesa Elgusbi

Abstract:

Mass poisonings with methanol are rare but occur regularly both in developed and in non-developing countries. As a result of the tragedy that happened in the city of Tripoli Libya in March during year 2013 a number of patients were admitted to Tripoli Medical Center and Tripoli Central Hospital suffering from poisoning following ingestion of methanol by mistake. Our aims have been formulated to collect Information about those cases as much as we can from the archiving departments from the two hospitals including the number of cases that had been admitted, recovered patients and died victims. This retrospective study was planned to find out the reasons which allow those patients to drink methanol in our Muslim community and also the role of pharmacist to prevent such a disaster that claimed the lives of many people. During this tragedy 291 ospitalized patients their ages between 16-32 years old were admitted to both hospitals, total number of died 189 (121 at Tripoli medical center) and (68 at Tripoli central hospital), demographic data also shows that most of them are male (97%) and (3% female), about 4% of the patients foreigners and 96% were Libyans. There were a lot of obstacles and poor facilities at the time of patient admission as recognized in many cases including lack of first line of treatment. The morbidity was high due to the lack of antidote and availability of dialysis machines at this two main hospitals in Tripoli also according to survey done to the medical staff and also a random number of medical students shows about 28% have no idea about the first aid procedure used for methanol poisoning cases and this due to the absence of continuing education for all medical staff through the establishment of training courses on first aid, rapid diagnosis of poisoning and follow the written procedures to dealing with such cases.

Keywords: ethanol, fomepizole, methanol, poisoning

Procedia PDF Downloads 332
9 Τhe Importance of Previous Examination Results, in Futural Differential Diagnostic Procedures and Especially in the Era of Covid-19

Authors: Angelis P. Barlampas

Abstract:

Purpose or Learning Objective It is well known that previous examinations play a major role in futural diagnosis, thus avoiding unnecessary new exams that cost in time and money both for the patient and the health system. A case is presented in which past patient’s results, in combination with the least needed new tests, give an easy final diagnosis. Methods or Background A middle aged man visited the emergency department complaining of hard controlled, persisting fever for the last few days. Laboratory tests showed an elevated number of white blood cells with neutrophil shift and abnormal CRP. The patient was admitted to hospital a month ago for continuing lungs symptomatology after a recent covid-19 infection. Results or Findings Computed tomography scanning showed a solid mass with spiculating margins in right lower lobe. After intravenous iodine contrast administration, there was mildly peripheral enhancement and eccentric non enhancing area. A pneumonic cancer was suspected. Comparison with the patient’s latest computed tomography revealed no mass in the area of interest but only signs of recent post covid-19 lung parenchyma abnormalities. Any new mass that appears in a month’s time span can not be a cancer but a benign lesion. It was obvious that an abscess was the most suitable explanation. The patient was admitted to hospital, and antibiotic therapy was given, with very good results. After a few days, the patient was afebrile and in good condition. Conclusion In this case , a PET scan or a biopsy was avoided, thanks to the patient’s medical history and the availability of previous examinations. It is worthy encouraging the patients to keep their medical records and organizing more efficiently the health system with the current technology of archiving the medical examinations, too.

Keywords: covid-19, chest ct, cancer, abscess, fever

Procedia PDF Downloads 28
8 Dynamic Contrast-Enhanced Breast MRI Examinations: Clinical Use and Technical Challenges

Authors: Janet Wing-Chong Wai, Alex Chiu-Wing Lee, Hailey Hoi-Ching Tsang, Jeffrey Chiu, Kwok-Wing Tang

Abstract:

Background: Mammography has limited sensitivity and specificity though it is the primary imaging technique for detection of early breast cancer. Ultrasound imaging and contrast-enhanced MRI are useful adjunct tools to mammography. The advantage of breast MRI is high sensitivity for invasive breast cancer. Therefore, indications for and use of breast magnetic resonance imaging have increased over the past decade. Objectives: 1. Cases demonstration on different indications for breast MR imaging. 2. To review of the common artifacts and pitfalls in breast MR imaging. Materials and Methods: This is a retrospective study including all patients underwent dynamic contrast-enhanced breast MRI examination in our centre, performed from Jan 2011 to Dec 2017. The clinical data and radiological images were retrieved from the EPR (electronic patient record), RIS (Radiology Information System) and PACS (Picture Archiving and Communication System). Results and Discussion: Cases including (1) Screening of the contralateral breast in patient with a new breast malignancy (2) Breast augmentation with free injection of unknown foreign materials (3) Finding of axillary adenopathy with an unknown site of primary malignancy (4) Neo-adjuvant chemotherapy: before, during, and after chemotherapy to evaluate treatment response and extent of residual disease prior to operation. Relevant images will be included and illustrated in the presentation. As with other types of MR imaging, there are different artifacts and pitfalls that can potentially limit interpretation of the images. Because of the coils and software specific to breast MR imaging, there are some other technical considerations that are unique to MR imaging of breast regions. Case demonstration images will be available in presentation. Conclusion: Breast MR imaging is a highly sensitive and reasonably specific method for the detection of breast cancer. Adherent to appropriate clinical indications and technical optimization are crucial for achieving satisfactory images for interpretation.

Keywords: MRI, breast, clinical, cancer

Procedia PDF Downloads 201
7 Dynamic Web-Based 2D Medical Image Visualization and Processing Software

Authors: Abdelhalim. N. Mohammed, Mohammed. Y. Esmail

Abstract:

In the course of recent decades, medical imaging has been dominated by the use of costly film media for review and archival of medical investigation, however due to developments in networks technologies and common acceptance of a standard digital imaging and communication in medicine (DICOM) another approach in light of World Wide Web was produced. Web technologies successfully used in telemedicine applications, the combination of web technologies together with DICOM used to design a web-based and open source DICOM viewer. The Web server allowance to inquiry and recovery of images and the images viewed/manipulated inside a Web browser without need for any preinstalling software. The dynamic site page for medical images visualization and processing created by using JavaScript and HTML5 advancements. The XAMPP ‘apache server’ is used to create a local web server for testing and deployment of the dynamic site. The web-based viewer connected to multiples devices through local area network (LAN) to distribute the images inside healthcare facilities. The system offers a few focal points over ordinary picture archiving and communication systems (PACS): easy to introduce, maintain and independently platforms that allow images to display and manipulated efficiently, the system also user-friendly and easy to integrate with an existing system that have already been making use of web technologies. The wavelet-based image compression technique on which 2-D discrete wavelet transform used to decompose the image then wavelet coefficients are transmitted by entropy encoding after threshold to decrease transmission time, stockpiling cost and capacity. The performance of compression was estimated by using images quality metrics such as mean square error ‘MSE’, peak signal to noise ratio ‘PSNR’ and compression ratio ‘CR’ that achieved (83.86%) when ‘coif3’ wavelet filter is used.

Keywords: DICOM, discrete wavelet transform, PACS, HIS, LAN

Procedia PDF Downloads 131
6 A Geo DataBase to Investigate the Maximum Distance Error in Quality of Life Studies

Authors: Paolino Di Felice

Abstract:

The background and significance of this study come from papers already appeared in the literature which measured the impact of public services (e.g., hospitals, schools, ...) on the citizens’ needs satisfaction (one of the dimensions of QOL studies) by calculating the distance between the place where they live and the location on the territory of the services. Those studies assume that the citizens' dwelling coincides with the centroid of the polygon that expresses the boundary of the administrative district, within the city, they belong to. Such an assumption “introduces a maximum measurement error equal to the greatest distance between the centroid and the border of the administrative district.”. The case study, this abstract reports about, investigates the implications descending from the adoption of such an approach but at geographical scales greater than the urban one, namely at the three levels of nesting of the Italian administrative units: the (20) regions, the (110) provinces, and the 8,094 municipalities. To carry out this study, it needs to be decided: a) how to store the huge amount of (spatial and descriptive) input data and b) how to process them. The latter aspect involves: b.1) the design of algorithms to investigate the geometry of the boundary of the Italian administrative units; b.2) their coding in a programming language; b.3) their execution and, eventually, b.4) archiving the results in a permanent support. The IT solution we implemented is centered around a (PostgreSQL/PostGIS) Geo DataBase structured in terms of three tables that fit well to the hierarchy of nesting of the Italian administrative units: municipality(id, name, provinceId, istatCode, regionId, geometry) province(id, name, regionId, geometry) region(id, name, geometry). The adoption of the DBMS technology allows us to implement the steps "a)" and "b)" easily. In particular, step "b)" is simplified dramatically by calling spatial operators and spatial built-in User Defined Functions within SQL queries against the Geo DB. The major findings coming from our experiments can be summarized as follows. The approximation that, on the average, descends from assimilating the residence of the citizens with the centroid of the administrative unit of reference is of few kilometers (4.9) at the municipalities level, while it becomes conspicuous at the other two levels (28.9 and 36.1, respectively). Therefore, studies such as those mentioned above can be extended up to the municipal level without affecting the correctness of the interpretation of the results, but not further. The IT framework implemented to carry out the experiments can be replicated for studies referring to the territory of other countries all over the world.

Keywords: quality of life, distance measurement error, Italian administrative units, spatial database

Procedia PDF Downloads 338
5 Data Quality and Associated Factors on Regular Immunization Programme at Ararso District: Somali Region- Ethiopia

Authors: Eyob Seife, Molla Alemayaehu, Tesfalem Teshome, Bereket Seyoum, Behailu Getachew

Abstract:

Globally, immunization averts between 2 and 3 million deaths yearly, but Vaccine-Preventable Diseases still account for more in Sub-Saharan African countries and takes the majority of under-five deaths yearly, which indicates the need for consistent and on-time information to have evidence-based decision so as to save lives of these vulnerable groups. However, ensuring data of sufficient quality and promoting an information-use culture at the point of collection remains critical and challenging, especially in remote areas where the Ararso district is selected based on a hypothesis of there is a difference in reported and recounted immunization data consistency. Data quality is dependent on different factors where organizational, behavioral, technical and contextual factors are the mentioned ones. A cross-sectional quantitative study was conducted on September 2022 in the Ararso district. The study used the world health organization (WHO) recommended data quality self-assessment (DQS) tools. Immunization tally sheets, registers and reporting documents were reviewed at 4 health facilities (1 health center and 3 health posts) of primary health care units for one fiscal year (12 months) to determine the accuracy ratio, availability and timeliness of reports. The data was collected by trained DQS assessors to explore the quality of monitoring systems at health posts, health centers, and at the district health office. A quality index (QI), availability and timeliness of reports were assessed. Accuracy ratios formulated were: the first and third doses of pentavalent vaccines, fully immunized (FI), TT2+ and the first dose of measles-containing vaccines (MCV). In this study, facility-level results showed poor timeliness at all levels and both over-reporting and under-reporting were observed at all levels when computing the accuracy ratio of registration to health post reports found at health centers for almost all antigens verified. A quality index (QI) of all facilities also showed poor results. Most of the verified immunization data accuracy ratios were found to be relatively better than that of quality index and timeliness of reports. So attention should be given to improving the capacity of staff, timeliness of reports and quality of monitoring system components, namely recording, reporting, archiving, data analysis and using information for decisions at all levels, especially in remote and areas.

Keywords: accuracy ratio, ararso district, quality of monitoring system, regular immunization program, timeliness of reports, Somali region-Ethiopia

Procedia PDF Downloads 28
4 Shark Detection and Classification with Deep Learning

Authors: Jeremy Jenrette, Z. Y. C. Liu, Pranav Chimote, Edward Fox, Trevor Hastie, Francesco Ferretti

Abstract:

Suitable shark conservation depends on well-informed population assessments. Direct methods such as scientific surveys and fisheries monitoring are adequate for defining population statuses, but species-specific indices of abundance and distribution coming from these sources are rare for most shark species. We can rapidly fill these information gaps by boosting media-based remote monitoring efforts with machine learning and automation. We created a database of shark images by sourcing 24,546 images covering 219 species of sharks from the web application spark pulse and the social network Instagram. We used object detection to extract shark features and inflate this database to 53,345 images. We packaged object-detection and image classification models into a Shark Detector bundle. We developed the Shark Detector to recognize and classify sharks from videos and images using transfer learning and convolutional neural networks (CNNs). We applied these models to common data-generation approaches of sharks: boosting training datasets, processing baited remote camera footage and online videos, and data-mining Instagram. We examined the accuracy of each model and tested genus and species prediction correctness as a result of training data quantity. The Shark Detector located sharks in baited remote footage and YouTube videos with an average accuracy of 89\%, and classified located subjects to the species level with 69\% accuracy (n =\ eight species). The Shark Detector sorted heterogeneous datasets of images sourced from Instagram with 91\% accuracy and classified species with 70\% accuracy (n =\ 17 species). Data-mining Instagram can inflate training datasets and increase the Shark Detector’s accuracy as well as facilitate archiving of historical and novel shark observations. Base accuracy of genus prediction was 68\% across 25 genera. The average base accuracy of species prediction within each genus class was 85\%. The Shark Detector can classify 45 species. All data-generation methods were processed without manual interaction. As media-based remote monitoring strives to dominate methods for observing sharks in nature, we developed an open-source Shark Detector to facilitate common identification applications. Prediction accuracy of the software pipeline increases as more images are added to the training dataset. We provide public access to the software on our GitHub page.

Keywords: classification, data mining, Instagram, remote monitoring, sharks

Procedia PDF Downloads 85