Search results for: real property
5770 Benchmarking of Pentesting Tools
Authors: Esteban Alejandro Armas Vega, Ana Lucila Sandoval Orozco, Luis Javier García Villalba
Abstract:
The benchmarking of tools for dynamic analysis of vulnerabilities in web applications is something that is done periodically, because these tools from time to time update their knowledge base and search algorithms, in order to improve their accuracy. Unfortunately, the vast majority of these evaluations are made by software enthusiasts who publish their results on blogs or on non-academic websites and always with the same evaluation methodology. Similarly, academics who have carried out this type of analysis from a scientific approach, the majority, make their analysis within the same methodology as well the empirical authors. This paper is based on the interest of finding answers to questions that many users of this type of tools have been asking over the years, such as, to know if the tool truly test and evaluate every vulnerability that it ensures do, or if the tool, really, deliver a real report of all the vulnerabilities tested and exploited. This kind of questions have also motivated previous work but without real answers. The aim of this paper is to show results that truly answer, at least on the tested tools, all those unanswered questions. All the results have been obtained by changing the common model of benchmarking used for all those previous works.Keywords: cybersecurity, IDS, security, web scanners, web vulnerabilities
Procedia PDF Downloads 3195769 Empirical Study of Running Correlations in Exam Marks: Same Statistical Pattern as Chance
Authors: Weisi Guo
Abstract:
It is well established that there may be running correlations in sequential exam marks due to students sitting in the order of course registration patterns. As such, a random and non-sequential sampling of exam marks is a standard recommended practice. Here, the paper examines a large number of exam data stretching several years across different modules to see the degree to which it is true. Using the real mark distribution as a generative process, it was found that random simulated data had no more sequential randomness than the real data. That is to say, the running correlations that one often observes are statistically identical to chance. Digging deeper, it was found that some high running correlations have students that indeed share a common course history and make similar mistakes. However, at the statistical scale of a module question, the combined effect is statistically similar to the random shuffling of papers. As such, there may not be the need to take random samples for marks, but it still remains good practice to mark papers in a random sequence to reduce the repetitive marking bias and errors.Keywords: data analysis, empirical study, exams, marking
Procedia PDF Downloads 1815768 Corrosion of Concrete Reinforcing Steel Bars Tested and Compared Between Various Protection Methods
Authors: P. van Tonder, U. Bagdadi, B. M. D. Lario, Z. Masina, T. R. Motshwari
Abstract:
This paper analyses how concrete reinforcing steel bars corrode and how it can be minimised through the use of various protection methods against corrosion, such as metal-based paint, alloying, cathodic protection and electroplating. Samples of carbon steel bars were protected, using these four methods. Tests performed on the samples included durability, electrical resistivity and bond strength. Durability results indicated relatively low corrosion rates for alloying, cathodic protection, electroplating and metal-based paint. The resistivity results indicate all samples experienced a downward trend, despite erratic fluctuations in the data, indicating an inverse relationship between electrical resistivity and corrosion rate. The results indicated lowered bond strengths when the reinforced concrete was cured in seawater compared to being cured in normal water. It also showed that higher design compressive strengths lead to higher bond strengths which can be used to compensate for the loss of bond strength due to corrosion in a real-world application. In terms of implications, all protection methods have the potential to be effective at resisting corrosion in real-world applications, especially the alloying, cathodic protection and electroplating methods. The metal-based paint underperformed by comparison, most likely due to the nature of paint in general which can fade and chip away, revealing the steel samples and exposing them to corrosion. For alloying, stainless steel is the suggested material of choice, where Y-bars are highly recommended as smooth bars have a much-lowered bond strength. Cathodic protection performed the best of all in protecting the sample from corrosion, however, its real-world application would require significant evaluation into the feasibility of such a method.Keywords: protection methods, corrosion, concrete, reinforcing steel bars
Procedia PDF Downloads 1735767 Hands-off Parking: Deep Learning Gesture-based System for Individuals with Mobility Needs
Authors: Javier Romera, Alberto Justo, Ignacio Fidalgo, Joshue Perez, Javier Araluce
Abstract:
Nowadays, individuals with mobility needs face a significant challenge when docking vehicles. In many cases, after parking, they encounter insufficient space to exit, leading to two undesired outcomes: either avoiding parking in that spot or settling for improperly placed vehicles. To address this issue, the following paper presents a parking control system employing gestural teleoperation. The system comprises three main phases: capturing body markers, interpreting gestures, and transmitting orders to the vehicle. The initial phase is centered around the MediaPipe framework, a versatile tool optimized for real-time gesture recognition. MediaPipe excels at detecting and tracing body markers, with a special emphasis on hand gestures. Hands detection is done by generating 21 reference points for each hand. Subsequently, after data capture, the project employs the MultiPerceptron Layer (MPL) for indepth gesture classification. This tandem of MediaPipe's extraction prowess and MPL's analytical capability ensures that human gestures are translated into actionable commands with high precision. Furthermore, the system has been trained and validated within a built-in dataset. To prove the domain adaptation, a framework based on the Robot Operating System (ROS), as a communication backbone, alongside CARLA Simulator, is used. Following successful simulations, the system is transitioned to a real-world platform, marking a significant milestone in the project. This real vehicle implementation verifies the practicality and efficiency of the system beyond theoretical constructs.Keywords: gesture detection, mediapipe, multiperceptron layer, robot operating system
Procedia PDF Downloads 1005766 A Laser Instrument Rapid-E+ for Real-Time Measurements of Airborne Bioaerosols Such as Bacteria, Fungi, and Pollen
Authors: Minghui Zhang, Sirine Fkaier, Sabri Fernana, Svetlana Kiseleva, Denis Kiselev
Abstract:
The real-time identification of bacteria and fungi is difficult because they emit much weaker signals than pollen. In 2020, Plair developed Rapid-E+, which extends abilities of Rapid-E to detect smaller bioaerosols such as bacteria and fungal spores with diameters down to 0.3 µm, while keeping the similar or even better capability for measurements of large bioaerosols like pollen. Rapid-E+ enables simultaneous measurements of (1) time-resolved, polarization and angle dependent Mie scattering patterns, (2) fluorescence spectra resolved in 16 channels, and (3) fluorescence lifetime of individual particles. Moreover, (4) it provides 2D Mie scattering images which give the full information on particle morphology. The parameters of every single bioaerosol aspired into the instrument are subsequently analysed by machine learning. Firstly, pure species of microbes, e.g., Bacillus subtilis (a species of bacteria), and Penicillium chrysogenum (a species of fungal spores), were aerosolized in a bioaerosol chamber for Rapid-E+ training. Afterwards, we tested microbes under different concentrations. We used several steps of data analysis to classify and identify microbes. All single particles were analysed by the parameters of light scattering and fluorescence in the following steps. (1) They were treated with a smart filter block to get rid of non-microbes. (2) By classification algorithm, we verified the filtered particles were microbes based on the calibration data. (3) The probability threshold (defined by the user) step provides the probability of being microbes ranging from 0 to 100%. We demonstrate how Rapid-E+ identified simultaneously microbes based on the results of Bacillus subtilis (bacteria) and Penicillium chrysogenum (fungal spores). By using machine learning, Rapid-E+ achieved identification precision of 99% against the background. The further classification suggests the precision of 87% and 89% for Bacillus subtilis and Penicillium chrysogenum, respectively. The developed algorithm was subsequently used to evaluate the performance of microbe classification and quantification in real-time. The bacteria and fungi were aerosolized again in the chamber with different concentrations. Rapid-E+ can classify different types of microbes and then quantify them in real-time. Rapid-E+ enables classifying different types of microbes and quantifying them in real-time. Rapid-E+ can identify pollen down to species with similar or even better performance than the previous version (Rapid-E). Therefore, Rapid-E+ is an all-in-one instrument which classifies and quantifies not only pollen, but also bacteria and fungi. Based on the machine learning platform, the user can further develop proprietary algorithms for specific microbes (e.g., virus aerosols) and other aerosols (e.g., combustion-related particles that contain polycyclic aromatic hydrocarbons).Keywords: bioaerosols, laser-induced fluorescence, Mie-scattering, microorganisms
Procedia PDF Downloads 905765 Analysis of Real Time Seismic Signal Dataset Using Machine Learning
Authors: Sujata Kulkarni, Udhav Bhosle, Vijaykumar T.
Abstract:
Due to the closeness between seismic signals and non-seismic signals, it is vital to detect earthquakes using conventional methods. In order to distinguish between seismic events and non-seismic events depending on their amplitude, our study processes the data that come from seismic sensors. The authors suggest a robust noise suppression technique that makes use of a bandpass filter, an IIR Wiener filter, recursive short-term average/long-term average (STA/LTA), and Carl short-term average (STA)/long-term average for event identification (LTA). The trigger ratio used in the proposed study to differentiate between seismic and non-seismic activity is determined. The proposed work focuses on significant feature extraction for machine learning-based seismic event detection. This serves as motivation for compiling a dataset of all features for the identification and forecasting of seismic signals. We place a focus on feature vector dimension reduction techniques due to the temporal complexity. The proposed notable features were experimentally tested using a machine learning model, and the results on unseen data are optimal. Finally, a presentation using a hybrid dataset (captured by different sensors) demonstrates how this model may also be employed in a real-time setting while lowering false alarm rates. The planned study is based on the examination of seismic signals obtained from both individual sensors and sensor networks (SN). A wideband seismic signal from BSVK and CUKG station sensors, respectively located near Basavakalyan, Karnataka, and the Central University of Karnataka, makes up the experimental dataset.Keywords: Carl STA/LTA, features extraction, real time, dataset, machine learning, seismic detection
Procedia PDF Downloads 1245764 The Effect of Object Presentation on Action Memory in School-Aged Children
Authors: Farzaneh Badinlou, Reza Kormi-Nouri, Monika Knopf
Abstract:
Enacted tasks are typically remembered better than when the same task materials are only verbally encoded, a robust finding referred to as the enactment effect. It has been assumed that enactment effect is independent of object presence but the size of enactment effect can be increased by providing objects at study phase in adults. To clarify the issues in children, free recall and cued recall performance of action phrases with or without using real objects were compared in 410 school-aged children from four age groups (8, 10, 12 and 14 years old). In this study, subjects were instructed to learn a series of action phrases under three encoding conditions, participants listened to verbal action phrases (VTs), performed the phrases (SPTs: subject-performed tasks), and observed the experimenter perform the phrases (EPTs: experimenter-performed tasks). Then, free recall and cued recall memory tests were administrated. The results revealed that the real object compared with imaginary objects improved recall performance in SPTs and EPTs, but more so in VTs. It was also found that the object presence was not necessary for the occurrence of the enactment effect but it was changed the size of enactment effect in all age groups. The size of enactment effect was more pronounced for imaginary objects than the real object in both free recall and cued recall memory tests in children. It was discussed that SPTs and EPTs deferentially facilitate item-specific and relation information processing and providing the objects can moderate the processing underlying the encoding conditions.Keywords: action memory, enactment effect, item-specific processing, object, relational processing, school-aged children
Procedia PDF Downloads 2385763 Outwrestling Cataclysmic Tsunamis at Hilo, Hawaii: Using Technical Developments of the past 50 Years to Improve Performance
Authors: Mark White
Abstract:
The best practices for owners and urban planners to manage tsunami risk have evolved during the last fifty years, and related technical advances have created opportunities for them to obtain better performance than in earlier cataclysmic tsunami inundations. This basic pattern is illustrated at Hilo Bay, the waterfront area of Hilo, Hawaii, an urban seaport which faces the most severe tsunami hazard of the Hawaiian archipelago. Since April 1, 1946, Hilo Bay has endured tsunami waves with a maximum water height exceeding 2.5 meters following four severe earthquakes: Unimak Island (Mw 8.6, 6.1 m) in 1946; Valdiva (Mw 9.5, the largest earthquake of the 20th century, 10.6 m) in 1960; William Prince Sound (Mw 9.2, 3.8 m) in 1964; and Kalapana (Mw 7.7, the largest earthquake in Hawaii since 1868, 2.6 m) in 1975. Ignoring numerous smaller tsunamis during the same time frame, these four cataclysmic tsunamis have caused property losses in Hilo to exceed $1.25 billion and more than 150 deaths. It is reasonable to foresee another cataclysmic tsunami inundating the urban core of Hilo in the next 50 years, which, if unchecked, could cause additional deaths and losses in the hundreds of millions of dollars. Urban planners and individual owners are now in a position to reduce these losses in the next foreseeable tsunami that generates maximum water heights between 2.5 and 10 meters in Hilo Bay. Since 1946, Hilo planners and individual owners have already created buffer zones between the shoreline and its historic downtown area. As these stakeholders make inevitable improvements to the built environment along and adjacent to the shoreline, they should incorporate new methods for better managing the obvious tsunami risk at Hilo. At the planning level, new manmade land forms, such as tsunami parks and inundation reservoirs, should be developed. Individual owners should require their design professionals to include sacrificial seismic and tsunami fuses that will perform well in foreseeable severe events and that can be easily repaired in the immediate aftermath. These investments before the next cataclysmic tsunami at Hilo will yield substantial reductions in property losses and fatalities.Keywords: hilo, tsunami parks, reservoirs, fuse systems, risk managment
Procedia PDF Downloads 1655762 The Establishment and Application of TRACE/FRAPTRAN Model for Kuosheng Nuclear Power Plant
Authors: S. W. Chen, W. K. Lin, J. R. Wang, C. Shih, H. T. Lin, H. C. Chang, W. Y. Li
Abstract:
Kuosheng nuclear power plant (NPP) is a BWR/6 type NPP and located on the northern coast of Taiwan. First, Kuosheng NPP TRACE model were developed in this research. In order to assess the system response of Kuosheng NPP TRACE model, startup tests data were used to evaluate Kuosheng NPP TRACE model. Second, the over pressurization transient analysis of Kuosheng NPP TRACE model was performed. Besides, in order to confirm the mechanical property and integrity of fuel rods, FRAPTRAN analysis was also performed in this study.Keywords: TRACE, safety analysis, BWR/6, FRAPTRA
Procedia PDF Downloads 5635761 How Hormesis Impacts Practice of Ecological Risk Assessment and Food Safety Assessment
Authors: Xiaoxian Zhang
Abstract:
Guidelines of ecological risk assessment (ERA) and food safety assessment (FSA) used nowadays, based on an S-shaped threshold dose-response curve (SDR), fail to consider hormesis, a reproducible biphasic dose-response model represented as a J-shaped or an inverted U-shaped curve, that occurs in the real-life environment across multitudinous compounds on cells, organisms, populations, and even the ecosystem. Specifically, in SDR-based ERA and FSA practice, predicted no effect concentration (PNEC) is calculated separately for individual substances from no observed effect concentration (NOEC, usually equivalent to 10% effect concentration (EC10) of a contaminant or food condiment) over an assessment coefficient that is bigger than 1. Experienced researchers doubted that hormesis in the real-life environment might lead to a waste of limited human and material resources in ERA and FSA practice, but related data are scarce. In this study, hormetic effects on bioluminescence of Aliivibrio fischeri (A. f) induced by sulfachloropyridazine (SCP) under 40 conditions to simulate the real-life scenario were investigated, and hormetic effects on growth of human MCF-7 cells caused by brown sugar and mascavado sugar were found likewise. After comparison of related parameters, it has for the first time been proved that there is a 50% probability for safe concentration (SC) of contaminants and food condiments to fall within the hormetic-stimulatory range (HSR) or left to HSR, revealing the unreliability of traditional parameters in standardized (eco)toxicological studies, and supporting qualitatively and quantitatively the over-strictness of ERA and FSA resulted from misuse of SDR. This study provides a novel perspective for ERA and FSA practitioners that hormesis should dominate and conditions where SDR works should only be singled out on a specific basis.Keywords: dose-response relationship, food safety, ecological risk assessment, hormesis
Procedia PDF Downloads 1455760 The Factors Affecting the Development of the Media and Animations for Vocational School in Thailand
Authors: Tanit Pruktara
Abstract:
The research aimed to study the students’ learning achievement and awareness level on electrical energy consumption and conservation and also to investigate the students’ attitude on the developed multimedia supplemented instructional unit for learning household electrical energy consumption and conservation in grade 10 Thailand student. This study used a quantitative method using MCQ for pre and post-achievement tests and Likert scales for awareness and attitude survey questionnaires. The results from this were employed to improve the multimedia to be appropriate for the classroom and with real life situations in the second phase, the main study. The experimental results showed that the developed learning unit significantly improved the students’ learning achievement as well as their awareness of electric energy conservation. Additional we found the student will enjoy participating in class activities when the lessons are taught using multimedia and helps them to develop the relevance between the course and real world situations.Keywords: lesson plan, media and animations, training course, vocational school in Thailand
Procedia PDF Downloads 1775759 Advanced Biosensor Characterization of Phage-Mediated Lysis in Real-Time and under Native Conditions
Authors: Radka Obořilová, Hana Šimečková, Matěj Pastucha, Jan Přibyl, Petr Skládal, Ivana Mašlaňová, Zdeněk Farka
Abstract:
Due to the spreading of antimicrobial resistance, alternative approaches to combat superinfections are being sought, both in the field of lysing agents and methods for studying bacterial lysis. A suitable alternative to antibiotics is phage therapy and enzybiotics, for which it is also necessary to study the mechanism of their action. Biosensor-based techniques allow rapid detection of pathogens in real time, verification of sensitivity to commonly used antimicrobial agents, and selection of suitable lysis agents. The detection of lysis takes place on the surface of the biosensor with immobilized bacteria, which has the potential to be used to study biofilms. An example of such a biosensor is surface plasmon resonance (SPR), which records the kinetics of bacterial lysis based on a change in the resonance angle. The bacteria are immobilized on the surface of the SPR chip, and the action of phage as the mass loss is monitored after a typical lytic cycle delay. Atomic force microscopy (AFM) is a technique for imaging of samples on the surface. In contrast to electron microscopy, it has the advantage of real-time imaging in the native conditions of the nutrient medium. In our case, Staphylococcus aureus was lysed using the enzyme lysostaphin and phage P68 from the familyPodoviridae at 37 ° C. In addition to visualization, AFM was used to study changes in mechanical properties during lysis, which resulted in a reduction of Young’s modulus (E) after disruption of the bacterial wall. Changes in E reflect the stiffness of the bacterium. These advanced methods provide deeper insight into bacterial lysis and can help to fight against bacterial diseases.Keywords: biosensors, atomic force microscopy, surface plasmon resonance, bacterial lysis, staphylococcus aureus, phage P68
Procedia PDF Downloads 1345758 Spatial Analysis of Park and Ride Users’ Dynamic Accessibility to Train Station: A Case Study in Perth
Authors: Ting (Grace) Lin, Jianhong (Cecilia) Xia, Todd Robinson
Abstract:
Accessibility analysis, examining people’s ability to access facilities and destinations, is a fundamental assessment for transport planning, policy making, and social exclusion research. Dynamic accessibility which measures accessibility in real-time traffic environment has been an advanced accessibility indicator in transport research. It is also a useful indicator to help travelers to understand travel time daily variability, assists traffic engineers to monitor traffic congestions, and finally develop effective strategies in order to mitigate traffic congestions. This research involved real-time traffic information by collecting travel time data with 15-minute interval via the TomTom® API. A framework for measuring dynamic accessibility was then developed based on the gravity theory and accessibility dichotomy theory through space and time interpolation. Finally, the dynamic accessibility can be derived at any given time and location under dynamic accessibility spatial analysis framework.Keywords: dynamic accessibility, hot spot, transport research, TomTom® API
Procedia PDF Downloads 3885757 Weight Estimation Using the K-Means Method in Steelmaking’s Overhead Cranes in Order to Reduce Swing Error
Authors: Seyedamir Makinejadsanij
Abstract:
One of the most important factors in the production of quality steel is to know the exact weight of steel in the steelmaking area. In this study, a calculation method is presented to estimate the exact weight of the melt as well as the objects transported by the overhead crane. Iran Alloy Steel Company's steelmaking area has three 90-ton cranes, which are responsible for transferring the ladles and ladle caps between 34 areas in the melt shop. Each crane is equipped with a Disomat Tersus weighing system that calculates and displays real-time weight. The moving object has a variable weight due to swinging, and the weighing system has an error of about +-5%. This means that when the object is moving by a crane, which weighs about 80 tons, the device (Disomat Tersus system) calculates about 4 tons more or 4 tons less, and this is the biggest problem in calculating a real weight. The k-means algorithm is an unsupervised clustering method that was used here. The best result was obtained by considering 3 centers. Compared to the normal average(one) or two, four, five, and six centers, the best answer is with 3 centers, which is logically due to the elimination of noise above and below the real weight. Every day, the standard weight is moved with working cranes to test and calibrate cranes. The results are shown that the accuracy is about 40 kilos per 60 tons (standard weight). As a result, with this method, the accuracy of moving weight is calculated as 99.95%. K-means is used to calculate the exact mean of objects. The stopping criterion of the algorithm is also the number of 1000 repetitions or not moving the points between the clusters. As a result of the implementation of this system, the crane operator does not stop while moving objects and continues his activity regardless of weight calculations. Also, production speed increased, and human error decreased.Keywords: k-means, overhead crane, melt weight, weight estimation, swing problem
Procedia PDF Downloads 905756 Intelligent Process Data Mining for Monitoring for Fault-Free Operation of Industrial Processes
Authors: Hyun-Woo Cho
Abstract:
The real-time fault monitoring and diagnosis of large scale production processes is helpful and necessary in order to operate industrial process safely and efficiently producing good final product quality. Unusual and abnormal events of the process may have a serious impact on the process such as malfunctions or breakdowns. This work try to utilize process measurement data obtained in an on-line basis for the safe and some fault-free operation of industrial processes. To this end, this work evaluated the proposed intelligent process data monitoring framework based on a simulation process. The monitoring scheme extracts the fault pattern in the reduced space for the reliable data representation. Moreover, this work shows the results of using linear and nonlinear techniques for the monitoring purpose. It has shown that the nonlinear technique produced more reliable monitoring results and outperforms linear methods. The adoption of the qualitative monitoring model helps to reduce the sensitivity of the fault pattern to noise.Keywords: process data, data mining, process operation, real-time monitoring
Procedia PDF Downloads 6405755 Dynamic Background Updating for Lightweight Moving Object Detection
Authors: Kelemewerk Destalem, Joongjae Cho, Jaeseong Lee, Ju H. Park, Joonhyuk Yoo
Abstract:
Background subtraction and temporal difference are often used for moving object detection in video. Both approaches are computationally simple and easy to be deployed in real-time image processing. However, while the background subtraction is highly sensitive to dynamic background and illumination changes, the temporal difference approach is poor at extracting relevant pixels of the moving object and at detecting the stopped or slowly moving objects in the scene. In this paper, we propose a moving object detection scheme based on adaptive background subtraction and temporal difference exploiting dynamic background updates. The proposed technique consists of a histogram equalization, a linear combination of background and temporal difference, followed by the novel frame-based and pixel-based background updating techniques. Finally, morphological operations are applied to the output images. Experimental results show that the proposed algorithm can solve the drawbacks of both background subtraction and temporal difference methods and can provide better performance than that of each method.Keywords: background subtraction, background updating, real time, light weight algorithm, temporal difference
Procedia PDF Downloads 3425754 Ship Detection Requirements Analysis for Different Sea States: Validation on Real SAR Data
Authors: Jaime Martín-de-Nicolás, David Mata-Moya, Nerea del-Rey-Maestre, Pedro Gómez-del-Hoyo, María-Pilar Jarabo-Amores
Abstract:
Ship detection is nowadays quite an important issue in tasks related to sea traffic control, fishery management and ship search and rescue. Although it has traditionally been carried out by patrol ships or aircrafts, coverage and weather conditions and sea state can become a problem. Synthetic aperture radars can surpass these coverage limitations and work under any climatological condition. A fast CFAR ship detector based on a robust statistical modeling of sea clutter with respect to sea states in SAR images is used. In this paper, the minimum SNR required to obtain a given detection probability with a given false alarm rate for any sea state is determined. A Gaussian target model using real SAR data is considered. Results show that SNR does not depend heavily on the class considered. Provided there is some variation in the backscattering of targets in SAR imagery, the detection probability is limited and a post-processing stage based on morphology would be suitable.Keywords: SAR, generalized gamma distribution, detection curves, radar detection
Procedia PDF Downloads 4525753 Advantages of Fuzzy Control Application in Fast and Sensitive Technological Processes
Authors: Radim Farana, Bogdan Walek, Michal Janosek, Jaroslav Zacek
Abstract:
This paper presents the advantages of fuzzy control use in technological processes control. The paper presents a real application of the Linguistic Fuzzy-Logic Control, developed at the University of Ostrava for the control of physical models in the Intelligent Systems Laboratory. The paper presents an example of a sensitive non-linear model, such as a magnetic levitation model and obtained results which show how modern information technologies can help to solve actual technical problems. A special method based on the LFLC controller with partial components is presented in this paper followed by the method of automatic context change, which is very helpful to achieve more accurate control results. The main advantage of the used system is its robustness in changing conditions demonstrated by comparing with conventional PID controller. This technology and real models are also used as a background for problem-oriented teaching, realized at the department for master students and their collaborative as well as individual final projects.Keywords: control, fuzzy logic, sensitive system, technological proves
Procedia PDF Downloads 4695752 Relevance of Copyright and Trademark in the Gaming Industry
Authors: Deeksha Karunakar
Abstract:
The gaming industry is one of the biggest industries in the world. Video games are interactive works of authorship that require the execution of a computer programme on specialized hardware but which also incorporate a wide variety of other artistic mediums, such as music, scripts, stories, video, paintings, and characters, into which the player takes an active role. Therefore, video games are not made as singular, simple works but rather as a collection of elements that, if they reach a certain level of originality and creativity, can each be copyrighted on their own. A video game is made up of a wide variety of parts, all of which combine to form the overall sensation that we, the players, have while playing. The entirety of the components is implemented in the form of software code, which is then translated into the game's user interface. Even while copyright protection is already in place for the coding of software, the work that is produced because of that coding can also be protected by copyright. This includes the game's storyline or narrative, its characters, and even elements of the code on their own. In each sector, there is a potential legal framework required, and the gaming industry also requires legal frameworks. This represents the importance of intellectual property laws in each sector. This paper will explore the beginnings of video games, the various aspects of game copyrights, and the approach of the courts, including examples of a few different instances. Although the creative arts have always been known to draw inspiration from and build upon the works of others, it has not always been simple to evaluate whether a game has been cloned. The video game business is experiencing growth as it has never seen before today. The majority of today's video games are both pieces of software and works of audio-visual art. Even though the existing legal framework does not have a clause specifically addressing video games, it is clear that there is a great many alternative means by which this protection can be granted. This paper will represent the importance of copyright and trademark laws in the gaming industry and its regulations with the help of relevant case laws via utilizing doctrinal methodology to support its findings. The aim of the paper is to make aware of the applicability of intellectual property laws in the gaming industry and how the justice system is evolving to adapt to such new industries. Furthermore, it will provide in-depth knowledge of their relationship with each other.Keywords: copyright, DMCA, gaming industry, trademark, WIPO
Procedia PDF Downloads 675751 A Study on Real-Time Fluorescence-Photoacoustic Imaging System for Mouse Thrombosis Monitoring
Authors: Sang Hun Park, Moung Young Lee, Su Min Yu, Hyun Sang Jo, Ji Hyeon Kim, Chul Gyu Song
Abstract:
A near-infrared light source used as a light source in the fluorescence imaging system is suitable for use in real-time during the operation since it has no interference in surgical vision. However, fluorescence images do not have depth information. In this paper, we configured the device with the research on molecular imaging systems for monitoring thrombus imaging using fluorescence and photoacoustic. Fluorescence imaging was performed using a phantom experiment in order to search the exact location, and the Photoacoustic image was in order to detect the depth. Fluorescence image obtained when evaluated through current phantom experiments when the concentration of the contrast agent is 25μg / ml, it was confirmed that it looked sharper. The phantom experiment is has shown the possibility with the fluorescence image and photoacoustic image using an indocyanine green contrast agent. For early diagnosis of cardiovascular diseases, more active research with the fusion of different molecular imaging devices is required.Keywords: fluorescence, photoacoustic, indocyanine green, carotid artery
Procedia PDF Downloads 6015750 Transition Economies, Typology, and Models: The Case of Libya
Authors: Abderahman Efhialelbum
Abstract:
The period since the fall of the Berlin Wall on November 9, 1989, and the collapse of the former Soviet Union in December 1985 has seen a major change in the economies and labour markets of Eastern Europe. The events also had reverberating effects across Asia and South America and parts of Africa, including Libya. This article examines the typologies and the models of transition economies. Also, it sheds light on the Libyan transition in particular and the impact of Qadhafi’s regime on the transition process. Finally, it illustrates how the Libyan transition process followed the trajectory of other countries using economic indicators such as free trade, property rights, and inflation.Keywords: transition, economy, typology, model, Libya
Procedia PDF Downloads 1565749 Use of Real Time Ultrasound for the Prediction of Carcass Composition in Serrana Goats
Authors: Antonio Monteiro, Jorge Azevedo, Severiano Silva, Alfredo Teixeira
Abstract:
The objective of this study was to compare the carcass and in vivo real-time ultrasound measurements (RTU) and their capacity to predict the composition of Serrana goats up to 40% of maturity. Twenty one females (11.1 ± 3.97 kg) and Twenty one males (15.6 ± 5.38 kg) were utilized to made in vivo measurements with a 5 MHz probe (ALOKA 500V scanner) at the 9th-10th, 10th-11th thoracic vertebrae (uT910 and uT1011, respectively), at the 1st- 2nd, 3rd-4th, and 4th-5th lumbar vertebrae (uL12, ul34 and uL45, respectively) and also at the 3rd-4th sternebrae (EEST). It was recorded the images of RTU measurements of Longissimus thoracis et lumborum muscle (LTL) depth (EM), width (LM), perimeter (PM), area (AM) and subcutaneous fat thickness (SFD) above the LTL, as well as the depth of tissues of the sternum (EEST) between the 3rd-4th sternebrae. All RTU images were analyzed using the ImageJ software. After slaughter, the carcasses were stored at 4 ºC for 24 h. After this period the carcasses were divided and the left half was entirely dissected into muscle, dissected fat (subcutaneous fat plus intermuscular fat) and bone. Prior to the dissection measurements equivalent to those obtained in vivo with RTU were recorded. Using the Statistica 5, correlation and regression analyses were performed. The prediction of carcass composition was achieved by stepwise regression procedure, with live weight and RTU measurements with and without transformation of variables to the same dimension. The RTU and carcass measurements, except for SFD measurements, showed high correlation (r > 0.60, P < 0.001). The RTU measurements and the live weight, showed ability to predict carcass composition on muscle (R2 = 0.99, P < 0.001), subcutaneous fat (R2 = 0.41, P < 0.001), intermuscular fat (R2 = 0.84, P < 0.001), dissected fat (R2 = 0.71, P < 0.001) and bone (R2 = 0.94, P < 0.001). The transformation of variables allowed a slight increase of precision, but with the increase in the number of variables, with the exception of subcutaneous fat prediction. In vivo measurements by RTU can be applied to predict kid goat carcass composition, from 5 measurements of RTU and the live weight.Keywords: carcass, goats, real time, ultrasound
Procedia PDF Downloads 2615748 Air Handling Units Power Consumption Using Generalized Additive Model for Anomaly Detection: A Case Study in a Singapore Campus
Authors: Ju Peng Poh, Jun Yu Charles Lee, Jonathan Chew Hoe Khoo
Abstract:
The emergence of digital twin technology, a digital replica of physical world, has improved the real-time access to data from sensors about the performance of buildings. This digital transformation has opened up many opportunities to improve the management of the building by using the data collected to help monitor consumption patterns and energy leakages. One example is the integration of predictive models for anomaly detection. In this paper, we use the GAM (Generalised Additive Model) for the anomaly detection of Air Handling Units (AHU) power consumption pattern. There is ample research work on the use of GAM for the prediction of power consumption at the office building and nation-wide level. However, there is limited illustration of its anomaly detection capabilities, prescriptive analytics case study, and its integration with the latest development of digital twin technology. In this paper, we applied the general GAM modelling framework on the historical data of the AHU power consumption and cooling load of the building between Jan 2018 to Aug 2019 from an education campus in Singapore to train prediction models that, in turn, yield predicted values and ranges. The historical data are seamlessly extracted from the digital twin for modelling purposes. We enhanced the utility of the GAM model by using it to power a real-time anomaly detection system based on the forward predicted ranges. The magnitude of deviation from the upper and lower bounds of the uncertainty intervals is used to inform and identify anomalous data points, all based on historical data, without explicit intervention from domain experts. Notwithstanding, the domain expert fits in through an optional feedback loop through which iterative data cleansing is performed. After an anomalously high or low level of power consumption detected, a set of rule-based conditions are evaluated in real-time to help determine the next course of action for the facilities manager. The performance of GAM is then compared with other approaches to evaluate its effectiveness. Lastly, we discuss the successfully deployment of this approach for the detection of anomalous power consumption pattern and illustrated with real-world use cases.Keywords: anomaly detection, digital twin, generalised additive model, GAM, power consumption, supervised learning
Procedia PDF Downloads 1545747 2D-Modeling with Lego Mindstorms
Authors: Miroslav Popelka, Jakub Nozicka
Abstract:
The whole work is based on possibility to use Lego Mindstorms robotics systems to reduce costs. Lego Mindstorms consists of a wide variety of hardware components necessary to simulate, programme and test of robotics systems in practice. To programme algorithm, which simulates space using the ultrasonic sensor, was used development environment supplied with kit. Software Matlab was used to render values afterwards they were measured by ultrasonic sensor. The algorithm created for this paper uses theoretical knowledge from area of signal processing. Data being processed by algorithm are collected by ultrasonic sensor that scans 2D space in front of it. Ultrasonic sensor is placed on moving arm of robot which provides horizontal moving of sensor. Vertical movement of sensor is provided by wheel drive. The robot follows map in order to get correct positioning of measured data. Based on discovered facts it is possible to consider Lego Mindstorm for low-cost and capable kit for real-time modelling.Keywords: LEGO Mindstorms, ultrasonic sensor, real-time modeling, 2D object, low-cost robotics systems, sensors, Matlab, EV3 Home Edition Software
Procedia PDF Downloads 4735746 Design and Development of Fleet Management System for Multi-Agent Autonomous Surface Vessel
Authors: Zulkifli Zainal Abidin, Ahmad Shahril Mohd Ghani
Abstract:
Agent-based systems technology has been addressed as a new paradigm for conceptualizing, designing, and implementing software systems. Agents are sophisticated systems that act autonomously across open and distributed environments in solving problems. Nevertheless, it is impractical to rely on a single agent to do all computing processes in solving complex problems. An increasing number of applications lately require multiple agents to work together. A multi-agent system (MAS) is a loosely coupled network of agents that interact to solve problems that are beyond the individual capacities or knowledge of each problem solver. However, the network of MAS still requires a main system to govern or oversees the operation of the agents in order to achieve a unified goal. We had developed a fleet management system (FMS) in order to manage the fleet of agents, plan route for the agents, perform real-time data processing and analysis, and issue sets of general and specific instructions to the agents. This FMS should be able to perform real-time data processing, communicate with the autonomous surface vehicle (ASV) agents and generate bathymetric map according to the data received from each ASV unit. The first algorithm is developed to communicate with the ASV via radio communication using standard National Marine Electronics Association (NMEA) protocol sentences. Next, the second algorithm will take care of the path planning, formation and pattern generation is tested using various sample data. Lastly, the bathymetry map generation algorithm will make use of data collected by the agents to create bathymetry map in real-time. The outcome of this research is expected can be applied on various other multi-agent systems.Keywords: autonomous surface vehicle, fleet management system, multi agent system, bathymetry
Procedia PDF Downloads 2715745 Field Evaluation of Pile Behavior in Sandy Soil Underlain by Clay
Authors: R. Bakr, M. Elmeligy, A. Ibrahim
Abstract:
When the building loads are relatively small, challenges are often facing the foundation design especially when inappropriate soil conditions exist. These may be represented in the existence of soft soil in the upper layers of soil while sandy soil or firm cohesive soil exist in the deeper layers. In such cases, the design becomes infeasible if the piles are extended to the deeper layers, especially when there are sandy layers existing at shallower depths underlain by stiff clayey soil. In this research, models of piles terminated in sand underlain by clay soils are numerically simulated by different modelling theories. Finite element software, Plaxis 3-D Foundation was used to evaluate the pile behavior under different loading scenarios. The standard static load test according to ASTM D-1143 was simulated and compared with the real-life loading scenario. The results showed that the pile behavior obtained from the current static load test do not realistically represent that obtained from real-life loading. Attempts were carried out to capture the proper numerical loading scenario that simulates the pile behavior in real-life loading including the long-term effect. A modified method based on this research findings is proposed for the static pile loading tests. Field loading tests were carried out to validate the new method. Results obtained from both numerical and field tests by using the modified method prove that this method is more accurate in predicting the pile behavior in sand soil underlain by clay more than the current standard static load.Keywords: numerical simulation, static load test, pile behavior, sand underlain with clay, creep
Procedia PDF Downloads 3225744 Digital Twin for Retail Store Security
Authors: Rishi Agarwal
Abstract:
Digital twins are emerging as a strong technology used to imitate and monitor physical objects digitally in real time across sectors. It is not only dealing with the digital space, but it is also actuating responses in the physical space in response to the digital space processing like storage, modeling, learning, simulation, and prediction. This paper explores the application of digital twins for enhancing physical security in retail stores. The retail sector still relies on outdated physical security practices like manual monitoring and metal detectors, which are insufficient for modern needs. There is a lack of real-time data and system integration, leading to ineffective emergency response and preventative measures. As retail automation increases, new digital frameworks must control safety without human intervention. To address this, the paper proposes implementing an intelligent digital twin framework. This collects diverse data streams from in-store sensors, surveillance, external sources, and customer devices and then Advanced analytics and simulations enable real-time monitoring, incident prediction, automated emergency procedures, and stakeholder coordination. Overall, the digital twin improves physical security through automation, adaptability, and comprehensive data sharing. The paper also analyzes the pros and cons of implementation of this technology through an Emerging Technology Analysis Canvas that analyzes different aspects of this technology through both narrow and wide lenses to help decision makers in their decision of implementing this technology. On a broader scale, this showcases the value of digital twins in transforming legacy systems across sectors and how data sharing can create a safer world for both retail store customers and owners.Keywords: digital twin, retail store safety, digital twin in retail, digital twin for physical safety
Procedia PDF Downloads 725743 Solving Dimensionality Problem and Finding Statistical Constructs on Latent Regression Models: A Novel Methodology with Real Data Application
Authors: Sergio Paez Moncaleano, Alvaro Mauricio Montenegro
Abstract:
This paper presents a novel statistical methodology for measuring and founding constructs in Latent Regression Analysis. This approach uses the qualities of Factor Analysis in binary data with interpretations on Item Response Theory (IRT). In addition, based on the fundamentals of submodel theory and with a convergence of many ideas of IRT, we propose an algorithm not just to solve the dimensionality problem (nowadays an open discussion) but a new research field that promises more fear and realistic qualifications for examiners and a revolution on IRT and educational research. In the end, the methodology is applied to a set of real data set presenting impressive results for the coherence, speed and precision. Acknowledgments: This research was financed by Colciencias through the project: 'Multidimensional Item Response Theory Models for Practical Application in Large Test Designed to Measure Multiple Constructs' and both authors belong to SICS Research Group from Universidad Nacional de Colombia.Keywords: item response theory, dimensionality, submodel theory, factorial analysis
Procedia PDF Downloads 3725742 Quality of Today's Teachers: Post-Certified Teachers' Competence in Alleviating Poverties towards a Sustainable Development
Authors: Sudirman
Abstract:
Competence is a term describing capability that correlates with a person’s occupation. The competence of a teacher consists of four, i.e., pedagogical, professional, personality and social competence. These four components are implemented during interacting with students to motivate the students and improve their achievement. The objective of this qualitative study is to explore the roles and contributions of certified teachers in alleviating the issue of poverty to promote a sustainable development. The data comprise primary and secondary data which were generated from observation, interview, documentation and library research. Furthermore, this study offers in-depth information regarding the performance of the teachers in coping with poverty and sustaining development. The result shows that the teacher’s competence positively contributes to the improvement of students’ achievement. This helps the students to prepare for the real work experience by which it results in a better income and, therefore, alleviate poverty. All in all, the quality of today’s teachers can be measured by their contribution in enhancing the students’ competence prior to entering real work, resulting in a wealthy society. This is to deal with poverty and conceptualizing a sustainable development.Keywords: competence, development, poverty, teachers
Procedia PDF Downloads 1505741 Exposure Assessment to Airborne Particulate Matter in Agriculture
Authors: K. Rumchev, S. Gilbey
Abstract:
Airborne particulate matter is a known hazard to human health, with a considerable body of evidence linking agricultural dust exposures to adverse human health effects in exposed populations. It is also known that agricultural workers are exposed to high levels of soil dust and other types of airborne particulate matter within the farming environment. The aim of this study was to examine exposure to agricultural dust among farm workers during the seeding season. Twenty-one wheat-belt farms consented to participate in the study with 30 workers being monitored for dust exposure whilst seeding or undertaking seeding associated tasks. Each farm was visited once and farmers’ were asked to wear a personal air sampler for a 4-hour sampling period. Simultaneous, real-time, tractor cabin air quality monitoring was also undertaken. Data for this study was collected using real-time aerosol dust monitors to determine in-tractor cabin PM exposure to five size fractions (total, PM10, respirable, PM2.5 and PM1), and personal sampling was undertaken to establish individual exposure to inhalable and respirable dust concentrations. The study established a significant difference between personal exposures and simultaneous real-time in-cabin exposures for both inhalable and respirable fractions. No significant difference was shown between in-cabin and personal inhalable dust concentrations during seeding and spraying tasks, although both in-cabin and personal concentrations were two times greater for seeding than spraying. Future research should focus on educating and providing farm owners and workers with more information on adopting safe work practices to minimise harmful exposures to agricultural dust.Keywords: agriculture, air quality, Australia, particulate matter
Procedia PDF Downloads 217