Search results for: course files
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 238

Search results for: course files

238 Detecting Venomous Files in IDS Using an Approach Based on Data Mining Algorithm

Authors: Sukhleen Kaur

Abstract:

In security groundwork, Intrusion Detection System (IDS) has become an important component. The IDS has received increasing attention in recent years. IDS is one of the effective way to detect different kinds of attacks and malicious codes in a network and help us to secure the network. Data mining techniques can be implemented to IDS, which analyses the large amount of data and gives better results. Data mining can contribute to improving intrusion detection by adding a level of focus to anomaly detection. So far the study has been carried out on finding the attacks but this paper detects the malicious files. Some intruders do not attack directly, but they hide some harmful code inside the files or may corrupt those file and attack the system. These files are detected according to some defined parameters which will form two lists of files as normal files and harmful files. After that data mining will be performed. In this paper a hybrid classifier has been used via Naive Bayes and Ripper classification methods. The results show how the uploaded file in the database will be tested against the parameters and then it is characterised as either normal or harmful file and after that the mining is performed. Moreover, when a user tries to mine on harmful file it will generate an exception that mining cannot be made on corrupted or harmful files.

Keywords: data mining, association, classification, clustering, decision tree, intrusion detection system, misuse detection, anomaly detection, naive Bayes, ripper

Procedia PDF Downloads 388
237 A Comprehensive Methodology for Voice Segmentation of Large Sets of Speech Files Recorded in Naturalistic Environments

Authors: Ana Londral, Burcu Demiray, Marcus Cheetham

Abstract:

Speech recording is a methodology used in many different studies related to cognitive and behaviour research. Modern advances in digital equipment brought the possibility of continuously recording hours of speech in naturalistic environments and building rich sets of sound files. Speech analysis can then extract from these files multiple features for different scopes of research in Language and Communication. However, tools for analysing a large set of sound files and automatically extract relevant features from these files are often inaccessible to researchers that are not familiar with programming languages. Manual analysis is a common alternative, with a high time and efficiency cost. In the analysis of long sound files, the first step is the voice segmentation, i.e. to detect and label segments containing speech. We present a comprehensive methodology aiming to support researchers on voice segmentation, as the first step for data analysis of a big set of sound files. Praat, an open source software, is suggested as a tool to run a voice detection algorithm, label segments and files and extract other quantitative features on a structure of folders containing a large number of sound files. We present the validation of our methodology with a set of 5000 sound files that were collected in the daily life of a group of voluntary participants with age over 65. A smartphone device was used to collect sound using the Electronically Activated Recorder (EAR): an app programmed to record 30-second sound samples that were randomly distributed throughout the day. Results demonstrated that automatic segmentation and labelling of files containing speech segments was 74% faster when compared to a manual analysis performed with two independent coders. Furthermore, the methodology presented allows manual adjustments of voiced segments with visualisation of the sound signal and the automatic extraction of quantitative information on speech. In conclusion, we propose a comprehensive methodology for voice segmentation, to be used by researchers that have to work with large sets of sound files and are not familiar with programming tools.

Keywords: automatic speech analysis, behavior analysis, naturalistic environments, voice segmentation

Procedia PDF Downloads 257
236 Nickel-Titanium Endodontic Instruments: The Evolution

Authors: Fadwa Chtioui

Abstract:

The field of endodontics has witnessed constant advancements in treatment methods and instrument design, particularly for nickel-titanium (NiTi) files. Despite these developments, it remains crucial for clinicians to have a thorough understanding of their characteristics and behavior to choose the appropriate instruments for different clinical and anatomical situations. Research Aim: The aim of this work is to study and discuss the impact of heat treatment developments on the properties of endodontic NiTi files, with the ultimate goal of providing ways to adapt these files to the anatomical features of dental roots. Methodology: This study involves both clinical cases and extensive bibliographic research. Findings: The study highlights the importance of heat treatment in the design and manufacture of NiTi files, as it significantly affects their physical and mechanical properties. It also provides insights into the ways in which NiTi files can be adapted to the complex geometries of dental roots for more effective endodontic treatments. Theoretical Importance: Theoretical implications of this study include a better understanding of the relationship between heat treatment and the properties of NiTi files, leading to improvements in both their manufacturing methods and clinical applications. Data Collection and Analysis Procedures: The data for this study was collected through clinical cases and an extensive review of relevant literature. Analysis was performed through qualitative and quantitative methods, examining the impact of heat treatment on the physical and mechanical properties of NiTi files. Questions Addressed: This study aims to answer questions concerning the properties of NiTi files and the impact of heat treatment on their behavior. It also seeks to examine ways in which these files can be adapted to complex dental root geometries for more effective endodontic treatments. Conclusion: In conclusion, this study emphasizes the importance of heat treatment in the design and manufacture of NiTi files, as it significantly impacts their physical and mechanical properties. Further research is necessary to explore additional methods for adapting NiTi files to the unique anatomies of dental roots to improve endodontic treatments further. Ultimately, this study provides valuable insights into the continued evolution of endodontic treatment and instrument design.

Keywords: endodontic files, nickel-titanium, tooth anatomy, heat treatment

Procedia PDF Downloads 32
235 Searchable Encryption in Cloud Storage

Authors: Ren Junn Hwang, Chung-Chien Lu, Jain-Shing Wu

Abstract:

Cloud outsource storage is one of important services in cloud computing. Cloud users upload data to cloud servers to reduce the cost of managing data and maintaining hardware and software. To ensure data confidentiality, users can encrypt their files before uploading them to a cloud system. However, retrieving the target file from the encrypted files exactly is difficult for cloud server. This study proposes a protocol for performing multikeyword searches for encrypted cloud data by applying k-nearest neighbor technology. The protocol ranks the relevance scores of encrypted files and keywords, and prevents cloud servers from learning search keywords submitted by a cloud user. To reduce the costs of file transfer communication, the cloud server returns encrypted files in order of relevance. Moreover, when a cloud user inputs an incorrect keyword and the number of wrong alphabet does not exceed a given threshold; the user still can retrieve the target files from cloud server. In addition, the proposed scheme satisfies security requirements for outsourced data storage.

Keywords: fault-tolerance search, multi-keywords search, outsource storage, ranked search, searchable encryption

Procedia PDF Downloads 342
234 Python Implementation for S1000D Applicability Depended Processing Model - SALERNO

Authors: Theresia El Khoury, Georges Badr, Amir Hajjam El Hassani, Stéphane N’Guyen Van Ky

Abstract:

The widespread adoption of machine learning and artificial intelligence across different domains can be attributed to the digitization of data over several decades, resulting in vast amounts of data, types, and structures. Thus, data processing and preparation turn out to be a crucial stage. However, applying these techniques to S1000D standard-based data poses a challenge due to its complexity and the need to preserve logical information. This paper describes SALERNO, an S1000d AppLicability dEpended pRocessiNg mOdel. This python-based model analyzes and converts the XML S1000D-based files into an easier data format that can be used in machine learning techniques while preserving the different logic and relationships in files. The model parses the files in the given folder, filters them, and extracts the required information to be saved in appropriate data frames and Excel sheets. Its main idea is to group the extracted information by applicability. In addition, it extracts the full text by replacing internal and external references while maintaining the relationships between files, as well as the necessary requirements. The resulting files can then be saved in databases and used in different models. Documents in both English and French languages were tested, and special characters were decoded. Updates on the technical manuals were taken into consideration as well. The model was tested on different versions of the S1000D, and the results demonstrated its ability to effectively handle the applicability, requirements, references, and relationships across all files and on different levels.

Keywords: aeronautics, big data, data processing, machine learning, S1000D

Procedia PDF Downloads 82
233 Evaluation of Dynamic Log Files for Different Dose Rates in IMRT Plans

Authors: Saad Bin Saeed, Fayzan Ahmed, Shahbaz Ahmed, Amjad Hussain

Abstract:

The aim of this study is to evaluate dynamic log files (Dynalogs) at different dose rates by dose-volume histograms (DVH) and used as a (QA) procedure of IMRT. Seven patients of phase one head and neck cancer with similar OAR`s are selected randomly. Reference plans of dose rate 300 and 600 MU/Min with prescribed dose of 50Gy in 25 fractions for each patient is made. Dynalogs produced by delivery of reference plans processed by in-house MATLAB program which produces new field files contain actual positions of multi-leaf collimators (MLC`s) instead of planned positions in reference plans. Copies of reference plans are used to import new field files generated by MATLAB program and renamed as Dyn.plan. After dose calculations of Dyn.plans for different dose rates, DVH, and multiple linear regression tools are used to evaluate reference and Dyn.plans. The results indicate good agreement of correlation between different dose rate plans. The maximum dose difference among PTV and OAR`s are found to be less than 5% and 9% respectively. The study indicates the potential of dynalogs to be used as patient-specific QA of IMRT at different dose rate.

Keywords: IMRT, dynalogs, dose rate, DVH

Procedia PDF Downloads 504
232 Unstructured-Data Content Search Based on Optimized EEG Signal Processing and Multi-Objective Feature Extraction

Authors: Qais M. Yousef, Yasmeen A. Alshaer

Abstract:

Over the last few years, the amount of data available on the globe has been increased rapidly. This came up with the emergence of recent concepts, such as the big data and the Internet of Things, which have furnished a suitable solution for the availability of data all over the world. However, managing this massive amount of data remains a challenge due to their large verity of types and distribution. Therefore, locating the required file particularly from the first trial turned to be a not easy task, due to the large similarities of names for different files distributed on the web. Consequently, the accuracy and speed of search have been negatively affected. This work presents a method using Electroencephalography signals to locate the files based on their contents. Giving the concept of natural mind waves processing, this work analyses the mind wave signals of different people, analyzing them and extracting their most appropriate features using multi-objective metaheuristic algorithm, and then classifying them using artificial neural network to distinguish among files with similar names. The aim of this work is to provide the ability to find the files based on their contents using human thoughts only. Implementing this approach and testing it on real people proved its ability to find the desired files accurately within noticeably shorter time and retrieve them as a first choice for the user.

Keywords: artificial intelligence, data contents search, human active memory, mind wave, multi-objective optimization

Procedia PDF Downloads 147
231 A Machine Learning Approach to Detecting Evasive PDF Malware

Authors: Vareesha Masood, Ammara Gul, Nabeeha Areej, Muhammad Asif Masood, Hamna Imran

Abstract:

The universal use of PDF files has prompted hackers to use them for malicious intent by hiding malicious codes in their victim’s PDF machines. Machine learning has proven to be the most efficient in identifying benign files and detecting files with PDF malware. This paper has proposed an approach using a decision tree classifier with parameters. A modern, inclusive dataset CIC-Evasive-PDFMal2022, produced by Lockheed Martin’s Cyber Security wing is used. It is one of the most reliable datasets to use in this field. We designed a PDF malware detection system that achieved 99.2%. Comparing the suggested model to other cutting-edge models in the same study field, it has a great performance in detecting PDF malware. Accordingly, we provide the fastest, most reliable, and most efficient PDF Malware detection approach in this paper.

Keywords: PDF, PDF malware, decision tree classifier, random forest classifier

Procedia PDF Downloads 51
230 Design of Bacterial Pathogens Identification System Based on Scattering of Laser Beam Light and Classification of Binned Plots

Authors: Mubashir Hussain, Mu Lv, Xiaohan Dong, Zhiyang Li, Bin Liu, Nongyue He

Abstract:

Detection and classification of microbes have a vast range of applications in biomedical engineering especially in detection, characterization, and quantification of bacterial contaminants. For identification of pathogens, different techniques are emerging in the field of biomedical engineering. Latest technology uses light scattering, capable of identifying different pathogens without any need for biochemical processing. Bacterial Pathogens Identification System (BPIS) which uses a laser beam, passes through the sample and light scatters off. An assembly of photodetectors surrounded by the sample at different angles to detect the scattering of light. The algorithm of the system consists of two parts: (a) Library files, and (b) Comparator. Library files contain data of known species of bacterial microbes in the form of binned plots, while comparator compares data of unknown sample with library files. Using collected data of unknown bacterial species, highest voltage values stored in the form of peaks and arranged in 3D histograms to find the frequency of occurrence. Resulting data compared with library files of known bacterial species. If sample data matching with any library file of known bacterial species, sample identified as a matched microbe. An experiment performed to identify three different bacteria particles: Enterococcus faecalis, Pseudomonas aeruginosa, and Escherichia coli. By applying algorithm using library files of given samples, results were compromising. This system is potentially applicable to several biomedical areas, especially those related to cell morphology.

Keywords: microbial identification, laser scattering, peak identification, binned plots classification

Procedia PDF Downloads 121
229 A Comparative Study of Malware Detection Techniques Using Machine Learning Methods

Authors: Cristina Vatamanu, Doina Cosovan, Dragos Gavrilut, Henri Luchian

Abstract:

In the past few years, the amount of malicious software increased exponentially and, therefore, machine learning algorithms became instrumental in identifying clean and malware files through semi-automated classification. When working with very large datasets, the major challenge is to reach both a very high malware detection rate and a very low false positive rate. Another challenge is to minimize the time needed for the machine learning algorithm to do so. This paper presents a comparative study between different machine learning techniques such as linear classifiers, ensembles, decision trees or various hybrids thereof. The training dataset consists of approximately 2 million clean files and 200.000 infected files, which is a realistic quantitative mixture. The paper investigates the above mentioned methods with respect to both their performance (detection rate and false positive rate) and their practicability.

Keywords: ensembles, false positives, feature selection, one side class algorithm

Procedia PDF Downloads 260
228 The Integration and Automation of EDA Tools in an Integrated Circuit Design Environment

Authors: Rohaya Abdul Wahab, Raja Mohd Fuad Tengku Aziz, Nazaliza Othman, Sharifah Saleh, Nabihah Razali, Rozaimah Baharim, M. Hanif M. Nasir

Abstract:

This paper will discuss how EDA tools are integrated and automated in an Integrated Circuit Design Environment. Some of the problems face in our current environment is that users need to configure manually on the library paths, start-up files and project directories. Certain manual processes that happen between the users and applications can be automated but they must be transparent to the users. For example, the users can run the applications directly after login without knowing the library paths and start-up files locations. The solution to these problems is to automate the processes using standard configuration files which will benefit the users and EDA support. This paper will discuss how the implementation is done to automate the process using scripting languages such as Perl, Tcl, Scheme and Shell Script. These scripting tools are great assets for design engineers to build a robust and powerful design flow and this technique is widely used to integrate all the tools together.

Keywords: EDA tools, Integrated Circuits, scripting, integration, automation

Procedia PDF Downloads 289
227 Metamorphic Computer Virus Classification Using Hidden Markov Model

Authors: Babak Bashari Rad

Abstract:

A metamorphic computer virus uses different code transformation techniques to mutate its body in duplicated instances. Characteristics and function of new instances are mostly similar to their parents, but they cannot be easily detected by the majority of antivirus in market, as they depend on string signature-based detection techniques. The purpose of this research is to propose a Hidden Markov Model for classification of metamorphic viruses in executable files. In the proposed solution, portable executable files are inspected to extract the instructions opcodes needed for the examination of code. A Hidden Markov Model trained on portable executable files is employed to classify the metamorphic viruses of the same family. The proposed model is able to generate and recognize common statistical features of mutated code. The model has been evaluated by examining the model on a test data set. The performance of the model has been practically tested and evaluated based on False Positive Rate, Detection Rate and Overall Accuracy. The result showed an acceptable performance with high average of 99.7% Detection Rate.

Keywords: malware classification, computer virus classification, metamorphic virus, metamorphic malware, Hidden Markov Model

Procedia PDF Downloads 283
226 Comparative Study of Case Files in the Context of H. P. Grice’s Pragmatic Theory

Authors: Tugce Arslan

Abstract:

For a communicative act to be carried out successfully, the speaker and the listener must consider certain principles in line with the intention–centered “Cooperative Principle” expressed by H. P. Grice. Violation of a communication principle causes the listener to make new inferences called “implicatures”. In this study, focusing on the linguistic use of H. P. Grice’s principles, we aim to find out which principles of conversation are generally followed in case files from different fields and which principles are frequently violated. Three case files were examined, and the violating and the abiding cases of the maxims were classified in terms of four categories (Quality, Quantity, Relevance and Manner). The results of this investigation is reported below (V: Violating, A: Abiding): Quality Quantity Relevance Manner V A V A V A V A Case 1 10 8 5 9 3 15 16 6 Case 2 4 5 11 6 2 11 7 14 Case 3 21 13 7 12 9 14 15 9 Total 35 26 23 27 14 40 38 29 The excerpts were selected from files covering three different areas: the Assize Court, the Family Court and the Commercial Court of First Instance. In this way, the relations between the types of violations and the types of courts are examined. Our main finding is that in the 1st and the 3rd file, as the cases of violation in “Quality” and “Manner” increase, the cases of violation in “Quantity” and “Relevance” decrease. In the second file, on the other hand, as the cases of violation in “Quantity” increase, the cases of violation in “Quality”, “Relevance” and “Manner” decrease. In the talk, we shall compare these results with the results obtained in the study of Tajabadi, Dowlatabadi, and Mehric (2014), which examined various case files in Iran. Our main finding is that in the study conducted in Iran, violations were found only on the principles of “Quantity” and “Relevance”, while violations were found on the principles of “Quality”, “Quantity” and “Manner” in this study. In this case, it shows us that there is a connection between at least two maxims. In both cases, it has been noticed that the “Quantity” maxim is a common denominator. Studies in this field can be enlightening for many areas such as discourse analysis, legal studies, etc. Accordingly, comments will be made about the nature of the violations mentioned in H. P. Grice’s “Cooperation Principle”. We shall also discuss various conversational practices that cannot be analysed with these maxims.

Keywords: comparative analysis, cooperation principle, forensic linguistics, pragmatic.

Procedia PDF Downloads 182
225 Central African Republic Government Recruitment Agency Based on Identity Management and Public Key Encryption

Authors: Koyangbo Guere Monguia Michel Alex Emmanuel

Abstract:

In e-government and especially recruitment, many researches have been conducted to build a trustworthy and reliable online or application system capable to process users or job applicant files. In this research (Government Recruitment Agency), cloud computing, identity management and public key encryption have been used to management domains, access control authorization mechanism and to secure data exchange between entities for reliable procedure of processing files.

Keywords: cloud computing network, identity management systems, public key encryption, access control and authorization

Procedia PDF Downloads 324
224 Enhanced Thai Character Recognition with Histogram Projection Feature Extraction

Authors: Benjawan Rangsikamol, Chutimet Srinilta

Abstract:

This research paper deals with extraction of Thai character features using the proposed histogram projection so as to improve the recognition performance. The process starts with transformation of image files into binary files before thinning. After character thinning, the skeletons are entered into the proposed extraction using histogram projection (horizontal and vertical) to extract unique features which are inputs of the subsequent recognition step. The recognition rate with the proposed extraction technique is as high as 97 percent since the technique works very well with the idiosyncrasies of Thai characters.

Keywords: character recognition, histogram projection, multilayer perceptron, Thai character features extraction

Procedia PDF Downloads 431
223 Practical Challenges of Tunable Parameters in Matlab/Simulink Code Generation

Authors: Ebrahim Shayesteh, Nikolaos Styliaras, Alin George Raducu, Ozan Sahin, Daniel Pombo VáZquez, Jonas Funkquist, Sotirios Thanopoulos

Abstract:

One of the important requirements in many code generation projects is defining some of the model parameters tunable. This helps to update the model parameters without performing the code generation again. This paper studies the concept of embedded code generation by MATLAB/Simulink coder targeting the TwinCAT Simulink system. The generated runtime modules are then tested and deployed to the TwinCAT 3 engineering environment. However, defining the parameters tunable in MATLAB/Simulink code generation targeting TwinCAT is not very straightforward. This paper focuses on this subject and reviews some of the techniques tested here to make the parameters tunable in generated runtime modules. Three techniques are proposed for this purpose, including normal tunable parameters, callback functions, and mask subsystems. Moreover, some test Simulink models are developed and used to evaluate the results of proposed approaches. A brief summary of the study results is presented in the following. First of all, the parameters defined tunable and used in defining the values of other Simulink elements (e.g., gain value of a gain block) could be changed after the code generation and this value updating will affect the values of all elements defined based on the values of the tunable parameter. For instance, if parameter K=1 is defined as a tunable parameter in the code generation process and this parameter is used to gain a gain block in Simulink, the gain value for the gain block is equal to 1 in the gain block TwinCAT environment after the code generation. But, the value of K can be changed to a new value (e.g., K=2) in TwinCAT (without doing any new code generation in MATLAB). Then, the gain value of the gain block will change to 2. Secondly, adding a callback function in the form of “pre-load function,” “post-load function,” “start function,” and will not help to make the parameters tunable without performing a new code generation. This means that any MATLAB files should be run before performing the code generation. The parameters defined/calculated in this file will be used as fixed values in the generated code. Thus, adding these files as callback functions to the Simulink model will not make these parameters flexible since the MATLAB files will not be attached to the generated code. Therefore, to change the parameters defined/calculated in these files, the code generation should be done again. However, adding these files as callback functions forces MATLAB to run them before the code generation, and there is no need to define the parameters mentioned in these files separately. Finally, using a tunable parameter in defining/calculating the values of other parameters through the mask is an efficient method to change the value of the latter parameters after the code generation. For instance, if tunable parameter K is used in calculating the value of two other parameters K1 and K2 and, after the code generation, the value of K is updated in TwinCAT environment, the value of parameters K1 and K2 will also be updated (without any new code generation).

Keywords: code generation, MATLAB, tunable parameters, TwinCAT

Procedia PDF Downloads 197
222 Encryption Image via Mutual Singular Value Decomposition

Authors: Adil Al-Rammahi

Abstract:

Image or document encryption is needed through e- government data base. Really in this paper we introduce two matrices images, one is the public, and the second is the secret (original). The analyses of each matrix is achieved using the transformation of singular values decomposition. So each matrix is transformed or analyzed to three matrices say row orthogonal basis, column orthogonal basis, and spectral diagonal basis. Product of the two row basis is calculated. Similarly the product of the two column basis is achieved. Finally we transform or save the files of public, row product and column product. In decryption stage, the original image is deduced by mutual method of the three public files.

Keywords: image cryptography, singular values decomposition

Procedia PDF Downloads 396
221 Proactive Disk Defragmentation through User's File-Access Patterns

Authors: Gordon Wong

Abstract:

This paper shows how the task of disk defragmentation can be handled by modern operating systems in a transparent, automated, efficient, and confined way through user's file-access patterns. Since files tend to gradually fragment from time to time through file creation, deletion, growth, and shrinking, the problem gets even worse when a disk becomes so fragmented that file accesses cannot be made reasonably efficient without performing the operation of defragmentation for the "entire" disk, which is done manually by the user by launching the disk defragmentation utility program normally bundled with the operating system. In this paper, we argue that the disk defragmentation problem described can be solved without having to manually use the utility program to defragment the entire disk. The argument is based on the observation that system users tend to access certain files in a particular time interval like the way observed for programs exhibiting temporal locality of memory references during their execution. The task of disk defragmentation can be initiated and acted upon for those files contained in the current file-access locality detected and identified by the operating system. The paper also discusses how to use the locality of file references approach to quantitatively measure and determine the locality of user's file access patterns on which the task of disk defragmentation is based.

Keywords: operating systems, disk defragmentation, locality of file accesses, system performance

Procedia PDF Downloads 19
220 Modelling of Passengers Exchange between Trains and Platforms

Authors: Guillaume Craveur

Abstract:

The evaluation of the passenger exchange time is necessary for railway operators in order to optimize and dimension rail traffic. Several influential parameters are identified and studied. Each parameter leads to a modeling completed with the buildingEXODUS software. The objective is the modelling of passenger exchanges measured by passenger counting. Population size is dimensioned using passenger counting files which are a report of the train service and contain following useful informations: number of passengers who get on and leave the train, exchange time. These information are collected by sensors placed at the top of each train door. With passenger counting files it is possible to know how many people are engaged in the exchange and how long is the exchange, but it is not possible to know passenger flow of the door. All the information about observed exchanges are thus not available. For this reason and in order to minimize inaccuracies, only short exchanges (less than 30 seconds) with a maximum of people are performed.

Keywords: passengers exchange, numerical tools, rolling stock, platforms

Procedia PDF Downloads 203
219 Development and Verification of the Idom Shielding Optimization Tool

Authors: Omar Bouhassoun, Cristian Garrido, César Hueso

Abstract:

The radiation shielding design is an optimization problem with multiple -constrained- objective functions (radiation dose, weight, price, etc.) that depend on several parameters (material, thickness, position, etc.). The classical approach for shielding design consists of a brute force trial-and-error process subject to previous designer experience. Therefore, the result is an empirical solution but not optimal, which can degrade the overall performance of the shielding. In order to automate the shielding design procedure, the IDOM Shielding Optimization Tool (ISOT) has been developed. This software combines optimization algorithms with the capabilities to read/write input files, run calculations, as well as parse output files for different radiation transport codes. In the first stage, the software was established to adjust the input files for two well-known Monte Carlo codes (MCNP and Serpent) and optimize the result (weight, volume, price, dose rate) using multi-objective genetic algorithms. Nevertheless, its modular implementation easily allows the inclusion of more radiation transport codes and optimization algorithms. The work related to the development of ISOT and its verification on a simple 3D multi-layer shielding problem using both MCNP and Serpent will be presented. ISOT looks very promising for achieving an optimal solution to complex shielding problems.

Keywords: optimization, shielding, nuclear, genetic algorithm

Procedia PDF Downloads 76
218 Comparison of the Cyclic Fatigue Resistance of Endoart Gold, Endoart Blue, Protaper Universal, and Protaper Gold Files at Body Temperature

Authors: Ayhan Eymirli, Sila N. Usta

Abstract:

The aim of this study is the comparison of the cyclic fatigue resistance of EndoArt Gold (EAG, Inci Dental, Istanbul, Turkey), EndoArt Blue (EAB, Inci Dental, Istanbul, Turkey), ProTaper Universal (PTU, Dentsply Tulsa Dental Specialties), and ProTaper Gold (PTG, Dentsply Tulsa Dental Specialties) files at body temperature. Twelve instruments of each EAG, EAB, PTU, PTG file system were included in this study. All selected files were rotated in the artificial canals, which have a 60° angle and a 5-mm radius of curvature until fracture occurred. The time to fracture (Ttf) was measured in seconds by a chronometer in the control panel that presents in the cyclic fatigue testing device when a fracture was detected visually and/or audibly. The lengths of the fractured fragments (FL) were also measured with a digital microcaliper. The data of Ttf and FL were analyzed using Kruskal-Wallis, one-way ANOVA and post hoc Bonferroni tests at the 5% significance level. There was a statistically significant difference among the file systems (p < 0.05). EAB had the statistically highest fatigue resistance, and PTU had the statistically lowest fatigue resistance (p < 0.05). PTG system had a statistically higher FL means than EAB and PTU file systems (p < 0.05). EAB had the greatest cyclic fatigue resistance amongst the other file systems. It can be stated that heat treatments may be a factor that increases fatigue resistance.

Keywords: cyclic fatigue resistance, Endo art blue, Endo art gold, pro taper gold, pro taper universal

Procedia PDF Downloads 104
217 Surge Analysis of Water Transmission Mains in Una, Himachal Pradesh, India

Authors: Baldev Setia, Raj Rajeshwari, Maneesh Kumar

Abstract:

Present paper is an analysis of water transmission mains failed due to surge analysis by using basic software known as Surge Analysis Program (SAP). It is a real time failure case study of a pipe laid in Una, Himachal Pradesh. The transmission main is a 13 kilometer long pipe with 7.9 kilometers as pumping main and 5.1 kilometers as gravitational main. The analysis deals with mainly pumping mains. The results are available in two text files. Besides, several files are prepared with specific view to obtain results in a graphical form. These results help to observe the pressure difference and surge occurrence at different locations along the pipe profile, which help to redesign the transmission main with different but suitable safety measures against possible surge. A technically viable and economically feasible design has been provided as per the relevant manual and standard code of practice.

Keywords: surge, water hammer, transmission mains, SAP 2000

Procedia PDF Downloads 323
216 Surge Analysis of Water Transmission Mains in Una, Himachal Pradesh (India)

Authors: Baldev Setia, Raj Rajeshwari, Maneesh Kumar

Abstract:

Present paper is an analysis of water transmission mains failed due to surge analysis by using basic software known as Surge Analysis Program (SAP). It is a real time failure case study of a pipe laid in Una, Himachal Pradesh. The transmission main is a 13 kilometres long pipe with 7.9 kilometres as pumping main and 5.1 kilometres as gravitational main. The analysis deals with mainly pumping mains. The results are available in two text files. Besides, several files are prepared with specific view to obtain results in a graphical form. These results help to observe the pressure difference and surge occurrence at different locations along the pipe profile, which help to redesign the transmission main with different but suitable safety measures against possible surge. A technically viable and economically feasible design has been provided as per the relevant manual and standard code of practice.

Keywords: surge, water hammer, transmission mains, SAP 2000

Procedia PDF Downloads 376
215 Hash Based Block Matching for Digital Evidence Image Files from Forensic Software Tools

Authors: M. Kaya, M. Eris

Abstract:

Internet use, intelligent communication tools, and social media have all become an integral part of our daily life as a result of rapid developments in information technology. However, this widespread use increases crimes committed in the digital environment. Therefore, digital forensics, dealing with various crimes committed in digital environment, has become an important research topic. It is in the research scope of digital forensics to investigate digital evidences such as computer, cell phone, hard disk, DVD, etc. and to report whether it contains any crime related elements. There are many software and hardware tools developed for use in the digital evidence acquisition process. Today, the most widely used digital evidence investigation tools are based on the principle of finding all the data taken place in digital evidence that is matched with specified criteria and presenting it to the investigator (e.g. text files, files starting with letter A, etc.). Then, digital forensics experts carry out data analysis to figure out whether these data are related to a potential crime. Examination of a 1 TB hard disk may take hours or even days, depending on the expertise and experience of the examiner. In addition, it depends on examiner’s experience, and may change overall result involving in different cases overlooked. In this study, a hash-based matching and digital evidence evaluation method is proposed, and it is aimed to automatically classify the evidence containing criminal elements, thereby shortening the time of the digital evidence examination process and preventing human errors.

Keywords: block matching, digital evidence, hash list, evaluation of digital evidence

Procedia PDF Downloads 227
214 Musical Composition by Computer with Inspiration from Files of Different Media Types

Authors: Cassandra Pratt Romero, Andres Gomez de Silva Garza

Abstract:

This paper describes a computational system designed to imitate human inspiration during musical composition. The system is called MIS (Musical Inspiration Simulator). The MIS system is inspired by media to which human beings are exposed daily (visual, textual, or auditory) to create new musical compositions based on the emotions detected in said media. After building the system we carried out a series of evaluations with volunteer users who used MIS to compose music based on images, texts, and audio files. The volunteers were asked to judge the harmoniousness and innovation in the system's compositions. An analysis of the results points to the difficulty of computational analysis of the characteristics of the media to which we are exposed daily, as human emotions have a subjective character. This observation will direct future improvements in the system.

Keywords: human inspiration, musical composition, musical composition by computer, theory of sensation and human perception

Procedia PDF Downloads 145
213 DURAFILE: A Collaborative Tool for Preserving Digital Media Files

Authors: Santiago Macho, Miquel Montaner, Raivo Ruusalepp, Ferran Candela, Xavier Tarres, Rando Rostok

Abstract:

During our lives, we generate a lot of personal information such as photos, music, text documents and videos that link us with our past. This data that used to be tangible is now digital information stored in our computers, which implies a software dependence to make them accessible in the future. Technology, however, constantly evolves and goes through regular shifts, quickly rendering various file formats obsolete. The need for accessing data in the future affects not only personal users but also organizations. In a digital environment, a reliable preservation plan and the ability to adapt to fast changing technology are essential for maintaining data collections in the long term. We present in this paper the European FP7 project called DURAFILE that provides the technology to preserve media files for personal users and organizations while maintaining their quality.

Keywords: artificial intelligence, digital preservation, social search, digital preservation plans

Procedia PDF Downloads 419
212 A Hybrid Watermarking Model Based on Frequency of Occurrence

Authors: Hamza A. A. Al-Sewadi, Adnan H. M. Al-Helali, Samaa A. K. Khamis

Abstract:

Ownership proofs of multimedia such as text, image, audio or video files can be achieved by the burial of watermark is them. It is achieved by introducing modifications into these files that are imperceptible to the human senses but easily recoverable by a computer program. These modifications would be in the time domain or frequency domain or both. This paper presents a procedure for watermarking by mixing amplitude modulation with frequency transformation histogram; namely a specific value is used to modulate the intensity component Y of the YIQ components of the carrier image. This scheme is referred to as histogram embedding technique (HET). Results comparison with those of other techniques such as discrete wavelet transform (DWT), discrete cosine transform (DCT) and singular value decomposition (SVD) have shown an enhance efficiency in terms of ease and performance. It has manifested a good degree of robustness against various environment effects such as resizing, rotation and different kinds of noise. This method would prove very useful technique for copyright protection and ownership judgment.

Keywords: authentication, copyright protection, information hiding, ownership, watermarking

Procedia PDF Downloads 536
211 A Graph Library Development Based on the Service-‎Oriented Architecture: Used for Representation of the ‎Biological ‎Systems in the Computer Algorithms

Authors: Mehrshad Khosraviani, Sepehr Najjarpour

Abstract:

Considering the usage of graph-based approaches in systems and synthetic biology, and the various types of ‎the graphs employed by them, a comprehensive graph library based ‎on the three-tier architecture (3TA) was previously introduced for full representation of the biological systems. Although proposing a 3TA-based graph library, three following reasons motivated us to redesign the graph ‎library based on the service-oriented architecture (SOA): (1) Maintaining the accuracy of the data related to an input graph (including its edges, its ‎vertices, its topology, etc.) without involving the end user:‎ Since, in the case of using 3TA, the library files are available to the end users, they may ‎be utilized incorrectly, and consequently, the invalid graph data will be provided to the ‎computer algorithms. However, considering the usage of the SOA, the operation of the ‎graph registration is specified as a service by encapsulation of the library files. In other words, overall control operations needed for registration of the valid data will be the ‎responsibility of the services. (2) Partitioning of the library product into some different parts: Considering 3TA, a whole library product was provided in general. While here, the product ‎can be divided into smaller ones, such as an AND/OR graph drawing service, and each ‎one can be provided individually. As a result, the end user will be able to select any ‎parts of the library product, instead of all features, to add it to a project. (3) Reduction of the complexities: While using 3TA, several other libraries must be needed to add for connecting to the ‎database, responsibility of the provision of the needed library resources in the SOA-‎based graph library is entrusted with the services by themselves. Therefore, the end user ‎who wants to use the graph library is not involved with its complexity. In the end, in order to ‎make ‎the library easier to control in the system, and to restrict the end user from accessing the files, ‎it was preferred to use the service-oriented ‎architecture ‎‎(SOA) over the three-tier architecture (3TA) and to redevelop the previously proposed graph library based on it‎.

Keywords: Bio-Design Automation, Biological System, Graph Library, Service-Oriented Architecture, Systems and Synthetic Biology

Procedia PDF Downloads 284
210 Frequency of Occurrence Hybrid Watermarking Scheme

Authors: Hamza A. Ali, Adnan H. M. Al-Helali

Abstract:

Generally, a watermark is information that identifies the ownership of multimedia (text, image, audio or video files). It is achieved by introducing modifications into these files that are imperceptible to the human senses but easily recoverable by a computer program. These modifications are done according to a secret key in a descriptive model that would be either in the time domain or frequency domain or both. This paper presents a procedure for watermarking by mixing amplitude modulation with frequency transformation histogram; namely a specific value is used to modulate the intensity component Y of the YIQ components of the carrier image. This scheme is referred to as histogram embedding technique (HET). Results comparison with those of other techniques such as discrete wavelet transform (DWT), discrete cosine transform (DCT) and singular value decomposition (SVD) have shown an enhance efficiency in terms of ease and performance. It has manifested a good degree of robustness against various environment effects such as resizing, rotation and different kinds of noise. This method would prove very useful technique for copyright protection and ownership judgment.

Keywords: watermarking, ownership, copyright protection, steganography, information hiding, authentication

Procedia PDF Downloads 345
209 The Effectiveness of Using MS SharePoint for the Curriculum Repository System

Authors: Misook Ahn

Abstract:

This study examines the Institutional Curriculum Repository (ICR) developed with MS SharePoint. The purpose of using MS SharePoint is to organize, share, and manage the curriculum data. The ICR aims to build a centralized curriculum infrastructure, preserve all curriculum materials, and provide academic service to users (faculty, students, or other agencies). The ICR collection includes core language curriculum materials developed by each language school—foreign language textbooks, language survival kits, and audio files currently in or not in use at the schools. All core curriculum materials with audio and video files have been coded, collected, and preserved at the ICR. All metadata for the collected curriculum materials have been input by language, code, year, book type, level, user, version, and current status (in use/not in use). The qualitative content analysis, including the survey data, is used to evaluate the effectiveness of using MS SharePoint for the repository system. This study explains how to manage and preserve curriculum materials with MS SharePoint, along with challenges and suggestions for further research. This study will be beneficial to other universities or organizations considering archiving or preserving educational materials.

Keywords: digital preservation, ms sharepoint, repository, curriculum materials

Procedia PDF Downloads 68