Search results for: location identification
4089 Detecting Anomalous Matches: An Empirical Study from National Basketball Association
Authors: Jacky Liu, Dulani Jayasuriya, Ryan Elmore
Abstract:
Match fixing and anomalous sports events have increasingly threatened the integrity of professional sports, prompting concerns about existing detection methods. This study addresses prior research limitations in match fixing detection, improving the identification of potential fraudulent matches by incorporating advanced anomaly detection techniques. We develop a novel method to identify anomalous matches and player performances by examining series of matches, such as playoffs. Additionally, we investigate bettors' potential profits when avoiding anomaly matches and explore factors behind unusual player performances. Our literature review covers match fixing detection, match outcome forecasting models, and anomaly detection methods, underscoring current limitations and proposing a new sports anomaly detection method. Our findings reveal anomalous series in the 2022 NBA playoffs, with the Phoenix Suns vs Dallas Mavericks series having the lowest natural occurrence probability. We identify abnormal player performances and bettors' profits significantly decrease when post-season matches are included. This study contributes by developing a new approach to detect anomalous matches and player performances, and assisting investigators in identifying responsible parties. While we cannot conclusively establish reasons behind unusual player performances, our findings suggest factors such as team financial difficulties, executive mismanagement, and individual player contract issues.Keywords: anomaly match detection, match fixing, match outcome forecasting, problematic players identification
Procedia PDF Downloads 784088 Groundwater Potential Zone Identification in Unconsolidated Aquifer Using Geophysical Techniques around Tarbela Ghazi, District Haripur, Pakistan
Authors: Syed Muzyan Shahzad, Liu Jianxin, Asim Shahzad, Muhammad Sharjeel Raza, Sun Ya, Fanidi Meryem
Abstract:
Electrical resistivity investigation was conducted in vicinity of Tarbela Ghazi, in order to study the subsurface layer with a view of determining the depth to the aquifer and thickness of groundwater potential zones. Vertical Electrical Sounding (VES) using Schlumberger array was carried out at 16 VES stations. Well logging data at four tube wells have been used to mark the super saturated zones with great discharge rate. The present paper shows a geoelectrical identification of the lithology and an estimate of the relationship between the resistivity and Dar Zarrouk parameters (transverse unit resistance and longitudinal unit conductance). The VES results revealed both homogeneous and heterogeneous nature of the subsurface strata. Aquifer is unconfined to confine in nature, and at few locations though perched aquifer has been identified, groundwater potential zones are developed in unconsolidated deposits layers and more than seven geo-electric layers are observed at some VES locations. Saturated zones thickness ranges from 5 m to 150 m, whereas at few area aquifer is beyond 150 m thick. The average anisotropy, transvers resistance and longitudinal conductance values are 0.86 %, 35750.9821 Ω.m2, 0.729 Siemens, respectively. The transverse unit resistance values fluctuate all over the aquifer system, whereas below at particular depth high values are observed, that significantly associated with the high transmissivity zones. The groundwater quality in all analyzed samples is below permissible limit according to World Health Standard (WHO).Keywords: aquifer, Dar Zarrouk parameters, geoelectric layers, Tarbela Ghazi
Procedia PDF Downloads 1954087 Pavement Failures and Its Maintenance
Authors: Maulik L. Sisodia, Tirth K. Raval, Aarsh S. Mistry
Abstract:
This paper summarizes the ongoing researches about the defects in both flexible and rigid pavement and the maintenance in both flexible and rigid pavements. Various defects in pavements have been identified since the existence of both flexible and rigid pavement. Flexible Pavement failure is defined in terms of decreasing serviceability caused by the development of cracks, ruts, potholes etc. Flexible Pavement structure can be destroyed in a single season due to water penetration. Defects in flexible pavements is a problem of multiple dimensions, phenomenal growth of vehicular traffic (in terms of no. of axle loading of commercial vehicles), the rapid expansion in the road network, non-availability of suitable technology, material, equipment, skilled labor and poor funds allocation have all added complexities to the problem of flexible pavements. In rigid pavements due to different type of destress the failure like joint spalling, faulting, shrinkage cracking, punch out, corner break etc. Application of correction in the existing surface will enhance the life of maintenance works as well as that of strengthening layer. Maintenance of a road network involves a variety of operations, i.e., identification of deficiencies and planning, programming and scheduling for actual implementation in the field and monitoring. The essential objective should be to keep the road surface and appurtenances in good condition and to extend the life of the road assets to its design life. The paper describes lessons learnt from pavement failures and problems experienced during the last few years on a number of projects in India. Broadly, the activities include identification of defects and the possible cause there off, determination of appropriate remedial measures; implement these in the field and monitoring of the results.Keywords: Flexible Pavements, Rigid Pavements, Defects, Maintenance
Procedia PDF Downloads 1704086 Improving Fingerprinting-Based Localization System Using Generative Artificial Intelligence
Authors: Getaneh Berie Tarekegn
Abstract:
A precise localization system is crucial for many artificial intelligence Internet of Things (AI-IoT) applications in the era of smart cities. Their applications include traffic monitoring, emergency alarming, environmental monitoring, location-based advertising, intelligent transportation, and smart health care. The most common method for providing continuous positioning services in outdoor environments is by using a global navigation satellite system (GNSS). Due to nonline-of-sight, multipath, and weather conditions, GNSS systems do not perform well in dense urban, urban, and suburban areas.This paper proposes a generative AI-based positioning scheme for large-scale wireless settings using fingerprinting techniques. In this article, we presented a novel semi-supervised deep convolutional generative adversarial network (S-DCGAN)-based radio map construction method for real-time device localization. We also employed a reliable signal fingerprint feature extraction method with t-distributed stochastic neighbor embedding (t-SNE), which extracts dominant features while eliminating noise from hybrid WLAN and long-term evolution (LTE) fingerprints. The proposed scheme reduced the workload of site surveying required to build the fingerprint database by up to 78.5% and significantly improved positioning accuracy. The results show that the average positioning error of GAILoc is less than 39 cm, and more than 90% of the errors are less than 82 cm. That is, numerical results proved that, in comparison to traditional methods, the proposed SRCLoc method can significantly improve positioning performance and reduce radio map construction costs.Keywords: location-aware services, feature extraction technique, generative adversarial network, long short-term memory, support vector machine
Procedia PDF Downloads 704085 Permeability Prediction Based on Hydraulic Flow Unit Identification and Artificial Neural Networks
Authors: Emad A. Mohammed
Abstract:
The concept of hydraulic flow units (HFU) has been used for decades in the petroleum industry to improve the prediction of permeability. This concept is strongly related to the flow zone indicator (FZI) which is a function of the reservoir rock quality index (RQI). Both indices are based on reservoir porosity and permeability of core samples. It is assumed that core samples with similar FZI values belong to the same HFU. Thus, after dividing the porosity-permeability data based on the HFU, transformations can be done in order to estimate the permeability from the porosity. The conventional practice is to use the power law transformation using conventional HFU where percentage of error is considerably high. In this paper, neural network technique is employed as a soft computing transformation method to predict permeability instead of power law method to avoid higher percentage of error. This technique is based on HFU identification where Amaefule et al. (1993) method is utilized. In this regard, Kozeny and Carman (K–C) model, and modified K–C model by Hasan and Hossain (2011) are employed. A comparison is made between the two transformation techniques for the two porosity-permeability models. Results show that the modified K-C model helps in getting better results with lower percentage of error in predicting permeability. The results also show that the use of artificial intelligence techniques give more accurate prediction than power law method. This study was conducted on a heterogeneous complex carbonate reservoir in Oman. Data were collected from seven wells to obtain the permeability correlations for the whole field. The findings of this study will help in getting better estimation of permeability of a complex reservoir.Keywords: permeability, hydraulic flow units, artificial intelligence, correlation
Procedia PDF Downloads 1354084 Human-Computer Interaction: Strategies for Ensuring the Design of User-Centered Web Interfaces for Smartphones
Authors: Byron Joseph A. Hallar, Annjeannette Alain D. Galang, Maria Visitacion N. Gumabay
Abstract:
The widespread adoption and increasing proliferation of smartphones that started during the first decade of the twenty-first century have enabled their users to communicate and access information in ways that were merely thought of as possibilities in the few years before the smartphone revolution. A product of the convergence of the cellular phone and portable computer, the smartphone provides an additional important function that used to be the exclusive domain of desktop-bound computers and portable computers: Web Browsing. For increasing numbers of users, the smartphone and allied devices such as tablet computers have become their first and often their only means of accessing the World Wide Web. This has led to the development of websites that cater to the needs of the new breed of smartphone-carrying web users. The smaller size of smartphones as compared with conventional computers has provided unique challenges to web interface designers. The smaller screen size and touchscreen interface have made it much more difficult to read and navigate through web pages that were in most part designed for traditional desktop and portable computers. Although increasing numbers of websites now provide an alternate website formatted for smartphones, problems with ease of use, reliability and usability still remain. This study focuses on the identification of the problems associated with smartphone web interfaces, the compliance with accepted standards of user-oriented web interface design, the strategies that could be utilized to ensure the design of user-centric web interfaces for smartphones, and the identification of the current trends and developments related to user-centric web interface design intended for the consumption of smartphone users.Keywords: human-computer interaction, user-centered design, web interface, mobile, smartphone
Procedia PDF Downloads 3564083 Determination of Identification and Antibiotic Resistance Rates of Pseudomonas aeruginosa Strains from Various Clinical Specimens in a University Hospital for Two Years, 2013-2015
Authors: Recep Kesli, Gulsah Asik, Cengiz Demir, Onur Turkyilmaz
Abstract:
Objective: Pseudomonas aeruginosa (P. aeruginosa) is an important nosocomial pathogen which causes serious hospital infections and is resistant to many commonly used antibiotics. P. aeruginosa can develop resistance during therapy and also it is very resistant to disinfectant chemicals. It may be found in respiratory support devices in hospitals. In this study, the antibiotic resistance of P. aeruginosa strains isolated from bronchial aspiration samples was evaluated retrospectively. Methods: Between October 2013 and September 2015, a total of 318 P. aeruginosa were isolated from clinical samples obtained from various intensive care units and inpatient patients hospitalized at Afyon Kocatepe University, ANS Practice and Research Hospital. Isolated bacteria identified by using both the conventional methods and automated identification system-VITEK 2 (bioMerieux, Marcy l’etoile France). Antibacterial resistance tests were performed by using Kirby-Bauer disc (Oxoid, Hampshire, England) diffusion method following the recommendations of CLSI. Results: Antibiotic resistance rates of identified 318 P. aeruginosa strains were found as follows for tested antibiotics; 32 % amikacin, 42% gentamicin, 43% imipenem, 43% meropenem, 50% ciprofloxacin, 57% levofloxacin, 38% cefepime, 63% ceftazidime, and 85% piperacillin/tazobactam. Conclusion: Resistance profiles change according to years and provinces for P. aeruginosa, so these findings should be considered empirical treatment choices. In this study, the highest and lowest resistance rates found against piperacillin/tazobactam % 85, and amikacin %32.Keywords: Pseudomonas aeruginosa, antibiotic resistance rates, intensive care unit, Pseudomonas spp.
Procedia PDF Downloads 2874082 Autonomous Exploration, Navigation and Mapping Payload Integrated on a Quadruped Robot
Authors: Julian Y. Raheema, Michael R. Hess, Raymond C. Provost, Mark Bilinski
Abstract:
The world is rapidly moving towards advancing and utilizing artificial intelligence and autonomous robotics. The ground-breaking Boston Dynamics quadruped robot, SPOT, was designed for industrial and commercial tasks requiring limited autonomous navigation. Out of the box, SPOT has route memorization and playback – it can repeat a path that it has been manually piloted through, but it cannot autonomously navigate an area that has not been previously explored. The presented SPOT payload package is built on ROS framework to support autonomous navigation and mapping of an unexplored environment. The package is fully integrated with SPOT to take advantage of motor controls and collision avoidance that comes natively with the robot. The payload runs all computations onboard, takes advantage of visual odometry SLAM and uses an Intel RealSense depth camera and Velodyne LiDAR sensor to generate 2D and 3D maps while in autonomous navigation mode. These maps are fused into the navigation stack to generate a costmap to enable the robot to safely navigate the environment without causing damage to the surroundings or the robot. The operator defines the operational zone and start location and then sends the explore command to have SPOT explore, generate 2D and 3D maps of the environment and return to the start location to await the operator's next command. The benefit of the presented package is that it is much lighter weight and less expensive than previous approaches and, importantly, operates in GPS-denied scenarios, which is ideal for indoor mapping. There are numerous applications that are hazardous to humans for SPOT enhanced with the autonomy payload, including disaster response, nuclear inspection, mine inspection, and so on. Other less extreme uses cases include autonomous 3D and 2D scanning of facilities for inspection, engineering and construction purposes.Keywords: autonomous, SLAM, quadruped, mapping, exploring, ROS, robotics, navigation
Procedia PDF Downloads 884081 Financial Performance Model of Local Economic Enterprises in Matalam, Cotabato
Authors: Kristel Faye Tandog
Abstract:
The State Owned Enterprise (SOE) or also called Public Enterprise (PE) has been playing a vital role in a country’s social and economic development. Following this idea, this study focused on the Factor Structures of Financial Performance of the Local Economic Enterprises (LEEs) namely: Food Court, Market, Slaughterhouse, and Terminal in Matalam, Cotabato. It aimed to determine the profile of the LEEs in terms of organizational structure, manner of creation, years in operation, source of initial operating requirements, annual operating budget, geographical location, and size or description of the facility. This study also included the different financial ratios of LEE that covered a five year period from Calendar Year 2009 to 2013. Primary data using survey questionnaire was administered to 468 respondents and secondary data were sourced out from the government archives and financial documents of the said LGU. There were 12 dominant factors identified namely: “management”, “enforcement of laws”, “strategic location”, “existence of non-formal competitors”, “proper maintenance”, “pricing”, “customer service”, “collection process”, “rentals and services”, “efficient use of resources”, “staffing”, and “timeliness and accuracy”. On the other hand, the financial performance of the LEE of Matalam, Cotabato using financial ratios needs reformatting. This denotes that refinement as to the following ratios: Cash Flow Indicator, Activity, Profitability and Growth is necessary. The cash flow indicator ratio showed difficulty in covering its debts in successive years. Likewise, the activity ratios showed that the LEE had not been effective in putting its investment at work. Moreover, profitability ratios revealed that it had operated in minimum capacity and had incurred net losses and thus, it had a weak profit performance. Furthermore, growth ratios showed that LEE had a declining growth trend particularly in net income.Keywords: factor structures, financial performance, financial ratios, state owned enterprises
Procedia PDF Downloads 2554080 Systematic Identification of Noncoding Cancer Driver Somatic Mutations
Authors: Zohar Manber, Ran Elkon
Abstract:
Accumulation of somatic mutations (SMs) in the genome is a major driving force of cancer development. Most SMs in the tumor's genome are functionally neutral; however, some cause damage to critical processes and provide the tumor with a selective growth advantage (termed cancer driver mutations). Current research on functional significance of SMs is mainly focused on finding alterations in protein coding sequences. However, the exome comprises only 3% of the human genome, and thus, SMs in the noncoding genome significantly outnumber those that map to protein-coding regions. Although our understanding of noncoding driver SMs is very rudimentary, it is likely that disruption of regulatory elements in the genome is an important, yet largely underexplored mechanism by which somatic mutations contribute to cancer development. The expression of most human genes is controlled by multiple enhancers, and therefore, it is conceivable that regulatory SMs are distributed across different enhancers of the same target gene. Yet, to date, most statistical searches for regulatory SMs have considered each regulatory element individually, which may reduce statistical power. The first challenge in considering the cumulative activity of all the enhancers of a gene as a single unit is to map enhancers to their target promoters. Such mapping defines for each gene its set of regulating enhancers (termed "set of regulatory elements" (SRE)). Considering multiple enhancers of each gene as one unit holds great promise for enhancing the identification of driver regulatory SMs. However, the success of this approach is greatly dependent on the availability of comprehensive and accurate enhancer-promoter (E-P) maps. To date, the discovery of driver regulatory SMs has been hindered by insufficient sample sizes and statistical analyses that often considered each regulatory element separately. In this study, we analyzed more than 2,500 whole-genome sequence (WGS) samples provided by The Cancer Genome Atlas (TCGA) and The International Cancer Genome Consortium (ICGC) in order to identify such driver regulatory SMs. Our analyses took into account the combinatorial aspect of gene regulation by considering all the enhancers that control the same target gene as one unit, based on E-P maps from three genomics resources. The identification of candidate driver noncoding SMs is based on their recurrence. We searched for SREs of genes that are "hotspots" for SMs (that is, they accumulate SMs at a significantly elevated rate). To test the statistical significance of recurrence of SMs within a gene's SRE, we used both global and local background mutation rates. Using this approach, we detected - in seven different cancer types - numerous "hotspots" for SMs. To support the functional significance of these recurrent noncoding SMs, we further examined their association with the expression level of their target gene (using gene expression data provided by the ICGC and TCGA for samples that were also analyzed by WGS).Keywords: cancer genomics, enhancers, noncoding genome, regulatory elements
Procedia PDF Downloads 1024079 GAILoc: Improving Fingerprinting-Based Localization System Using Generative Artificial Intelligence
Authors: Getaneh Berie Tarekegn
Abstract:
A precise localization system is crucial for many artificial intelligence Internet of Things (AI-IoT) applications in the era of smart cities. Their applications include traffic monitoring, emergency alarming, environmental monitoring, location-based advertising, intelligent transportation, and smart health care. The most common method for providing continuous positioning services in outdoor environments is by using a global navigation satellite system (GNSS). Due to nonline-of-sight, multipath, and weather conditions, GNSS systems do not perform well in dense urban, urban, and suburban areas.This paper proposes a generative AI-based positioning scheme for large-scale wireless settings using fingerprinting techniques. In this article, we presented a novel semi-supervised deep convolutional generative adversarial network (S-DCGAN)-based radio map construction method for real-time device localization. We also employed a reliable signal fingerprint feature extraction method with t-distributed stochastic neighbor embedding (t-SNE), which extracts dominant features while eliminating noise from hybrid WLAN and long-term evolution (LTE) fingerprints. The proposed scheme reduced the workload of site surveying required to build the fingerprint database by up to 78.5% and significantly improved positioning accuracy. The results show that the average positioning error of GAILoc is less than 39 cm, and more than 90% of the errors are less than 82 cm. That is, numerical results proved that, in comparison to traditional methods, the proposed SRCLoc method can significantly improve positioning performance and reduce radio map construction costs.Keywords: location-aware services, feature extraction technique, generative adversarial network, long short-term memory, support vector machine
Procedia PDF Downloads 724078 Improving Fingerprinting-Based Localization System Using Generative AI
Authors: Getaneh Berie Tarekegn, Li-Chia Tai
Abstract:
With the rapid advancement of artificial intelligence, low-power built-in sensors on Internet of Things devices, and communication technologies, location-aware services have become increasingly popular and have permeated every aspect of people’s lives. Global navigation satellite systems (GNSSs) are the default method of providing continuous positioning services for ground and aerial vehicles, as well as consumer devices (smartphones, watches, notepads, etc.). However, the environment affects satellite positioning systems, particularly indoors, in dense urban and suburban cities enclosed by skyscrapers, or when deep shadows obscure satellite signals. This is because (1) indoor environments are more complicated due to the presence of many objects surrounding them; (2) reflection within the building is highly dependent on the surrounding environment, including the positions of objects and human activity; and (3) satellite signals cannot be reached in an indoor environment, and GNSS doesn't have enough power to penetrate building walls. GPS is also highly power-hungry, which poses a severe challenge for battery-powered IoT devices. Due to these challenges, IoT applications are limited. Consequently, precise, seamless, and ubiquitous Positioning, Navigation and Timing (PNT) systems are crucial for many artificial intelligence Internet of Things (AI-IoT) applications in the era of smart cities. Their applications include traffic monitoring, emergency alarms, environmental monitoring, location-based advertising, intelligent transportation, and smart health care. This paper proposes a generative AI-based positioning scheme for large-scale wireless settings using fingerprinting techniques. In this article, we presented a semi-supervised deep convolutional generative adversarial network (S-DCGAN)-based radio map construction method for real-time device localization. We also employed a reliable signal fingerprint feature extraction method with t-distributed stochastic neighbor embedding (t-SNE), which extracts dominant features while eliminating noise from hybrid WLAN and long-term evolution (LTE) fingerprints. The proposed scheme reduced the workload of site surveying required to build the fingerprint database by up to 78.5% and significantly improved positioning accuracy. The results show that the average positioning error of GAILoc is less than 0.39 m, and more than 90% of the errors are less than 0.82 m. According to numerical results, SRCLoc improves positioning performance and reduces radio map construction costs significantly compared to traditional methods.Keywords: location-aware services, feature extraction technique, generative adversarial network, long short-term memory, support vector machine
Procedia PDF Downloads 414077 Quantification and Identification of the Main Components of the Biomass of the Microalgae Scenedesmus SP. – Prospection of Molecules of Commercial Interest
Authors: Carolina V. Viegas, Monique Gonçalves, Gisel Chenard Diaz, Yordanka Reyes Cruz, Donato Alexandre Gomes Aranda
Abstract:
To develop the massive cultivation of microalgae, it is necessary to isolate and characterize the species, improving genetic tools in search of specific characteristics. Therefore, the detection, identification and quantification of the compounds that compose the Scenedesmus sp. were prerequisites to verify the potential of these microalgae. The main objective of this work was to carry out the characterization of Scenedesmus sp. as to the content of ash, carbohydrates, proteins and lipids as well as the determination of the composition of their lipid classes and main fatty acids. The biomass of Scenedesmus sp, showed 15,29 ± 0,23 % of ash and CaO (36,17 %) was the main component of this fraction, The total protein and carbohydrate content of the biomass was 40,74 ± 1,01 % and 23,37 ± 0,95 %, respectively, proving to be a potential source of proteins as well as carbohydrates for the production of ethanol via fermentation, The lipid contents extracted via Bligh & Dyer and in situ saponification were 8,18 ± 0,13 % and 4,11 ± 0,11 %, respectively. In the lipid extracts obtained via Bligh & Dyer, approximately 50 % of the composition of this fraction consists of fatty compounds, while the other half is composed of an unsaponifiable fraction composed mainly of chlorophylls, phytosterols and carotenes. From the lowest yield, it was possible to obtain a selectivity of 92,14 % for fatty components (fatty acids and fatty esters) confirmed through the infrared spectroscopy technique. The presence of polyunsaturated acids (~45 %) in the lipid extracts indicated the potential of this fraction as a source of nutraceuticals. The results indicate that the biomass of Scenedesmus sp, can become a promising potential source for obtaining polyunsaturated fatty acids, carotenoids and proteins as well as the simultaneous obtainment of different compounds of high commercial value.Keywords: microalgae, Desmodesmus, lipid classes, fatty acid profile, proteins, carbohydrates
Procedia PDF Downloads 954076 Hemispheric Locus and Gender Predict the Delay between the Moment of Stroke and Hospitalization
Authors: D. Anderlini, G. Wallis
Abstract:
Background: The number of people experiencing stroke is steadily increasing due to changes in diet and lifestyle, to longer life expectancy resulting in older population, to higher survival rates as a consequence of improvements during the acute phase. This study considers what risk factors might contribute to delayed entry to hospital for treatment. Methods: We analyzed data from 2472 patients admitted to the Stroke Unit of the Royal Brisbane Women's Hospital, Australia, between 2002 to 2011. Results: Previous studies have reported that factors which can contribute to delay include the patient’s age, the time of day, physical location, visit the GP instead of going to the emergency, means of transport, severity of symptoms and type of stroke. Contrary to findings of other studies, we found a strong correlation between side of lesion and delay in admission: patients with right hemisphere lesions had an average delay of 3.78 days, while patients with left hemisphere lesions had an average delay of 1.49 days. Damage to the right hemisphere generally ends in motor impairment in the non-dominant hand and no speech impediment. In contrast, left hemisphere lesions can result in deficit to; dominant hand function and aphasia which will be noticed even if their impact on performance is relatively minor. A finding which goes against many previous studies, is the fact that women get to the hospital much sooner than men, showing an average delay of 0.92 days in women vs. 3.36 days in men. Conclusion: Acute surgical-pharmacological therapies are most effective if applied immediately after stroke. Hence delays to admission can be crucial to the degree of recovery. The tendency of patients to overlook symptoms of right hemisphere lesion should be the target of information campaigns both for the general public and GPs. Why do men go to hospital so late? We don't know yet! Nevertheless an awareness plan specifically direct to male population should be on the agenda of Health Departments.Keywords: gender, admission delay, stroke location, bioinformatics, biomedicine
Procedia PDF Downloads 2294075 GIS Mapping of Sheep Population and Distribution Pattern in the Derived Savannah of Nigeria
Authors: Sosina Adedayo O., Babyemi Olaniyi J.
Abstract:
The location, population, and distribution pattern of sheep are severe challenges to agribusiness investment and policy formulation in the livestock industry. There is a significant disconnect between farmers' needs and the policy framework towards ameliorating the sheep production constraints. Information on the population, production, and distribution pattern of sheep remains very scanty. A multi-stage sampling technique was used to elicit information from 180 purposively selected respondents from the study area comprised of Oluyole, Ona-ara, Akinyele, Egbeda, Ido and Ibarapa East LGA. The Global Positioning Systems (GPS) of the farmers' location (distribution), and average sheep herd size (Total Livestock Unit, TLU) (population) were recorded, taking the longitude and latitude of the locations in question. The recorded GPS data of the study area were transferred into the ARC-GIS. The ARC-GIS software processed the data using the ARC-GIS model 10.0. Sheep production and distribution (TLU) ranged from 4.1 (Oluyole) to 25.0 (Ibarapa East), with Oluyole, Akinyele, Ona-ara and Egbeda having TLU of 5, 7, 8 and 20, respectively. The herd sizes were classified as less than 8 (smallholders), 9-25 (medium), 26-50 (large), and above 50 (commercial). The majority (45%) of farmers were smallholders. The FR CP (%) ranged from 5.81±0.26 (cassava leaf) to 24.91±0.91 (Amaranthus spinosus), NDF (%) ranged from 22.38±4.43 (Amaranthus spinosus) to 67.96 ± 2.58 (Althemanthe dedentata) while ME ranged from 7.88±0.24 (Althemanthe dedentata) to 10.68±0.18 (cassava leaf). The smallholders’ sheep farmers were the majority, evenly distributed across rural areas due to the availability of abundant feed resources (crop residues, tree crops, shrubs, natural pastures, and feed ingredients) coupled with a large expanse of land in the study area. Most feed resources available were below sheep protein requirement level, hence supplementation is necessary for productivity. Bio-informatics can provide relevant information for sheep production for policy framework and intervention strategies.Keywords: sheep enterprise, agribusiness investment, policy, bio-informatics, ecological zone
Procedia PDF Downloads 814074 Smart Safari: Safari Guidance Mobile Application
Authors: D. P. Lawrence, T. M. M. D. Ariyarathna, W. N. K. De Silva, M. D. S. C. De Silva, Lasantha Abeysiri, Pradeep Abeygunawardhna
Abstract:
Safari traveling is one of the most famous hobbies all over the world. In Sri Lanka, 'Yala' is the second-largest national park, which is a better place to go for a safari. Many number of local and foreign travelers are coming to go for a safari in 'Yala'. But 'Yala' does not have a mobile application that is made to facilitate the traveler with some important features that the traveler wants to achieve in the safari experience. To overcome these difficulties, the proposed mobile application by adding those identified features to make travelers, guiders, and administration's works easier. The proposed safari traveling guidance mobile application is called 'SMART SAFARI' for the 'Yala' National Park in Sri Lanka. There are four facilities in this mobile application that provide for travelers as well as the guiders. As the first facility, the guider and traveler can view the created map of the park, and the guider can add temporary locations of animals and special locations on the map. This is a Geographic Information System (GIS) to capture, analyze, and display geographical data. And as the second facility is to generate optimal paths according to the travelers' requirements through the park by using machine learning techniques. In the third part, the traveler can get information about animals using an animal identification system by capturing the animal. As in the other facility, the traveler will be facilitated to add reviews and a rate and view those comments under categorized sections and pre-defined score range. With those facilities, this user-friendly mobile application provides the user to get a better experience in safari traveling, and it will probably help to develop tourism culture in Sri Lanka.Keywords: animal identification system, geographic information system, machine learning techniques, pre defined score range
Procedia PDF Downloads 1314073 Improving Fingerprinting-Based Localization (FPL) System Using Generative Artificial Intelligence (GAI)
Authors: Getaneh Berie Tarekegn, Li-Chia Tai
Abstract:
With the rapid advancement of artificial intelligence, low-power built-in sensors on Internet of Things devices, and communication technologies, location-aware services have become increasingly popular and have permeated every aspect of people’s lives. Global navigation satellite systems (GNSSs) are the default method of providing continuous positioning services for ground and aerial vehicles, as well as consumer devices (smartphones, watches, notepads, etc.). However, the environment affects satellite positioning systems, particularly indoors, in dense urban and suburban cities enclosed by skyscrapers, or when deep shadows obscure satellite signals. This is because (1) indoor environments are more complicated due to the presence of many objects surrounding them; (2) reflection within the building is highly dependent on the surrounding environment, including the positions of objects and human activity; and (3) satellite signals cannot be reached in an indoor environment, and GNSS doesn't have enough power to penetrate building walls. GPS is also highly power-hungry, which poses a severe challenge for battery-powered IoT devices. Due to these challenges, IoT applications are limited. Consequently, precise, seamless, and ubiquitous Positioning, Navigation and Timing (PNT) systems are crucial for many artificial intelligence Internet of Things (AI-IoT) applications in the era of smart cities. Their applications include traffic monitoring, emergency alarming, environmental monitoring, location-based advertising, intelligent transportation, and smart health care. This paper proposes a generative AI-based positioning scheme for large-scale wireless settings using fingerprinting techniques. In this article, we presented a novel semi-supervised deep convolutional generative adversarial network (S-DCGAN)-based radio map construction method for real-time device localization. We also employed a reliable signal fingerprint feature extraction method with t-distributed stochastic neighbor embedding (t-SNE), which extracts dominant features while eliminating noise from hybrid WLAN and long-term evolution (LTE) fingerprints. The proposed scheme reduced the workload of site surveying required to build the fingerprint database by up to 78.5% and significantly improved positioning accuracy. The results show that the average positioning error of GAILoc is less than 0.39 m, and more than 90% of the errors are less than 0.82 m. According to numerical results, SRCLoc improves positioning performance and reduces radio map construction costs significantly compared to traditional methods.Keywords: location-aware services, feature extraction technique, generative adversarial network, long short-term memory, support vector machine
Procedia PDF Downloads 464072 Effect of Packaging Material and Water-Based Solutions on Performance of Radio Frequency Identification for Food Packaging Applications
Authors: Amelia Frickey, Timothy (TJ) Sheridan, Angelica Rossi, Bahar Aliakbarian
Abstract:
The growth of large food supply chains demanded improved end-to-end traceability of food products, which has led to companies being increasingly interested in using smart technologies such as Radio Frequency Identification (RFID)-enabled packaging to track items. As technology is being widely used, there are several technological or economic issues that should be overcome to facilitate the adoption of this track-and-trace technology. One of the technological challenges of RFID technology is its sensitivity to different environmental form factors, including packaging materials and the content of the packaging. Although researchers have assessed the performance loss due to the proximity of water and aqueous solutions, there is still the need to further investigate the impacts of food products on the reading range of RFID tags. However, to the best of our knowledge, there are not enough studies to determine the correlation between RFID tag performance and food beverages properties. The goal of this project was to investigate the effect of the solution properties (pH and conductivity) and different packaging materials filled with food-like water-based solutions on the performance of an RFID tag. Three commercially available ultra high-frequency RFID tags were placed on three different bottles and filled with different concentrations of water-based solutions, including sodium chloride, citric acid, sucrose, and ethanol. Transparent glass, Polyethylneterephtalate (PET), and Tetrapak® were used as the packaging materials commonly used in the beverage industries. Tag readability (Theoretical Read Range, TRR) and sensitivity (Power on Tag Forward, PoF) were determined using an anechoic chamber. First, the best place to attach the tag for each packaging material was investigated using empty and water-filled bottles. Then, the bottles were filled with the food-like solutions and tested with the three different tags and the PoF and TRR at the fixed frequency of 915MHz. In parallel, the pH and conductivity of solutions were measured. The best-performing tag was then selected to test the bottles filled with wine, orange, and apple juice. Despite various solutions altering the performance of each tag, the change in tag performance had no correlation with the pH or conductivity of the solution. Additionally, packaging material played a significant role in tag performance. Each tag tested performed optimally under different conditions. This study is the first part of comprehensive research to determine the regression model for the prediction of tag performance behavior based on the packaging material and the content. More investigations, including more tags and food products, are needed to be able to develop a robust regression model. The results of this study can be used by RFID tag manufacturers to design suitable tags for specific products with similar properties.Keywords: smart food packaging, supply chain management, food waste, radio frequency identification
Procedia PDF Downloads 1124071 Efficient Storage in Cloud Computing by Using Index Replica
Authors: Bharat Singh Deora, Sushma Satpute
Abstract:
Cloud computing is based on resource sharing. Like other resources which can be shareable, storage is a resource which can be shared. We can use collective resources of storage from different locations and maintain a central index table for storage details. The storage combining of different places can form a suitable data storage which is operated from one location and is very economical. Proper storage of data should improve data reliability & availability and bandwidth utilization. Also, we are moving the contents of one storage to other according to our need.Keywords: cloud computing, cloud storage, Iaas, PaaS, SaaS
Procedia PDF Downloads 3404070 A Study of Binding Methods and Techniques in Safavid Era Emphasizing on Iran Shahnamehs (16-18th Century AD/10-12th Century AH)
Authors: Ashrafosadat Mousavi Laer, Elaheh Moravej
Abstract:
The art of binding was simple and elementary at the beginning of Islam. This art thrived gradually and continued its development as an independent art. Identification of the binding techniques and used materials in covers and investigation of the arrays give us indexes for the better identification of different doctrines and methods of that time. The catalogers of the manuscripts usually pay attention to four items: gender, color, art elegances, injury, and exquisiteness of the cover. The criterion for classification of the covers is their art nature and gender. 15th century AD (9th century AH) was the period of the binding art development in which the most beautiful covers were produced by the so-called method of ‘burning’. At 16th century AD (10th century AH), in Safavid era, art changed completely and a fundamental evolution occurred in the technique and method of binding. The greatest change in this art was the extensive use of stamp that was made mostly of steel and copper. Theses stamps were presses against leather. These covers were called ‘beat’. In this paper, writing and bookbinding of about 32 Shahnamehs of Safavid era available in the Iranian libraries and museums are studied. An analytical-statistical study shows that four methods have been used including beat, burning, mosaic, and oily. 69 percent of the covers of these copies are cardboards with a leathery coating (goatskin) and have been produced by burning and beat methods. Its reasons are that these two methods have been common methods in Safavid era and performing them was only feasible on leather and the most desirable and commonly used leather of that time was goatskin which was the best option for cover legend durability and preserving the book and it was more durable because it had been made of goat skin. In addition, it had prepared a suitable opportunity for the binding artist’s creativity and innovation.Keywords: Shahnameh, Safavid era, bookbinding, beat cover, burning cover
Procedia PDF Downloads 2374069 Measurement of Ionospheric Plasma Distribution over Myanmar Using Single Frequency Global Positioning System Receiver
Authors: Win Zaw Hein, Khin Sandar Linn, Su Su Yi Mon, Yoshitaka Goto
Abstract:
The Earth ionosphere is located at the altitude of about 70 km to several 100 km from the ground, and it is composed of ions and electrons called plasma. In the ionosphere, these plasma makes delay in GPS (Global Positioning System) signals and reflect in radio waves. The delay along the signal path from the satellite to the receiver is directly proportional to the total electron content (TEC) of plasma, and this delay is the largest error factor in satellite positioning and navigation. Sounding observation from the top and bottom of the ionosphere was popular to investigate such ionospheric plasma for a long time. Recently, continuous monitoring of the TEC using networks of GNSS (Global Navigation Satellite System) observation stations, which are basically built for land survey, has been conducted in several countries. However, in these stations, multi-frequency support receivers are installed to estimate the effect of plasma delay using their frequency dependence and the cost of multi-frequency support receivers are much higher than single frequency support GPS receiver. In this research, single frequency GPS receiver was used instead of expensive multi-frequency GNSS receivers to measure the ionospheric plasma variation such as vertical TEC distribution. In this measurement, single-frequency support ublox GPS receiver was used to probe ionospheric TEC. The location of observation was assigned at Mandalay Technological University in Myanmar. In the method, the ionospheric TEC distribution is represented by polynomial functions for latitude and longitude, and parameters of the functions are determined by least-squares fitting on pseudorange data obtained at a known location under an assumption of thin layer ionosphere. The validity of the method was evaluated by measurements obtained by the Japanese GNSS observation network called GEONET. The performance of measurement results using single-frequency of GPS receiver was compared with the results by dual-frequency measurement.Keywords: ionosphere, global positioning system, GPS, ionospheric delay, total electron content, TEC
Procedia PDF Downloads 1364068 The Feasibility of a Protected Launch Site near Melkbosstrand for a Public Transport Ferry across Table Bay, Cape Town
Authors: Mardi Falck, André Theron
Abstract:
Traffic congestion on the Northern side of Table Bay is a major problem. In Gauteng, the implementation of the Gautrain between Pretoria and Johannesburg, solved their traffic congestion. In 2002 two entrepreneurs endeavoured to implement a hovercraft ferry service across the bay from Table View to the Port of Cape Town. However, the EIA process proved that disgruntled residents from the area did not agree with their location for a launch site. 17 years later the traffic problem has not gone away, but instead the congestion has increased. While property prices in the City Bowl of Cape Town are ever increasing, people tend to live more on the outskirts of the CBD and commute to work. This means more vehicles on the road every day and the public transport services cannot keep up with the demand. For this reason, the study area of the previous hovercraft plans is being extended further North. The study’s aim is thus to determine the feasibility of a launch site North of Bloubergstrand to launch and receive a public transport ferry across Table Bay. The feasibility is being established by researching ferry services across the world and on what makes them successful. Different types of ferries and their operational capacities in terms of weather and waves are researched and by establishing the offshore and nearshore wind and wave climate for the area, an appropriate protected launch site is determined. It was concluded that travel time could potentially be halved. A hovercraft proved to be the most feasible ferry type, because it does not require a conventional harbour. Other types of vessels require a protected launch site because of the wave climate. This means large breakwaters that influence the cost substantially. The Melkbos Cultural Centre proved to be the most viable option for the location of the launch site, because it already has buildings and infrastructure. It is recommended that, if a harbour is chosen for the proposed ferry service, it could be used for more services like fishing, eco-tourism and leisure. Further studies are recommended to optimise the feasibility of such a harbour.Keywords: Cape Town, ferry, public, Table Bay
Procedia PDF Downloads 1524067 Development of Digital Twin Concept to Detect Abnormal Changes in Structural Behaviour
Authors: Shady Adib, Vladimir Vinogradov, Peter Gosling
Abstract:
Digital Twin (DT) technology is a new technology that appeared in the early 21st century. The DT is defined as the digital representation of living and non-living physical assets. By connecting the physical and virtual assets, data are transmitted smoothly, allowing the virtual asset to fully represent the physical asset. Although there are lots of studies conducted on the DT concept, there is still limited information about the ability of the DT models for monitoring and detecting unexpected changes in structural behaviour in real time. This is due to the large computational efforts required for the analysis and an excessively large amount of data transferred from sensors. This paper aims to develop the DT concept to be able to detect the abnormal changes in structural behaviour in real time using advanced modelling techniques, deep learning algorithms, and data acquisition systems, taking into consideration model uncertainties. finite element (FE) models were first developed offline to be used with a reduced basis (RB) model order reduction technique for the construction of low-dimensional space to speed the analysis during the online stage. The RB model was validated against experimental test results for the establishment of a DT model of a two-dimensional truss. The established DT model and deep learning algorithms were used to identify the location of damage once it has appeared during the online stage. Finally, the RB model was used again to identify the damage severity. It was found that using the RB model, constructed offline, speeds the FE analysis during the online stage. The constructed RB model showed higher accuracy for predicting the damage severity, while deep learning algorithms were found to be useful for estimating the location of damage with small severity.Keywords: data acquisition system, deep learning, digital twin, model uncertainties, reduced basis, reduced order model
Procedia PDF Downloads 974066 The Reasons behind Individuals to Join Terrorist Organizations: Recruitment from Outside
Authors: Murat Sözen
Abstract:
Today terrorism is gaining momentum again. Parallel to this, it hurts more than before because it has victims from not only its own locations but also remote places. As victims are from outside, militants are likewise from own location and outside. What made these individuals join the terrorist organizations and how these organizations recruit militants are still unanswered. The purpose of this work is to find reasons of joining and power of recruiting. In addition, the role of most popular tool of recruiting, ‘social media’ will be examined.Keywords: recruitment, social media, recruitment, militants
Procedia PDF Downloads 3464065 Monitoring of Serological Test of Blood Serum in Indicator Groups of the Population of Central Kazakhstan
Authors: Praskovya Britskaya, Fatima Shaizadina, Alua Omarova, Nessipkul Alysheva
Abstract:
Planned preventive vaccination, which is carried out in the Republic of Kazakhstan, promoted permanent decrease in the incidence of measles and viral hepatitis B. In the structure of VHB patients prevail people of young, working age. Monitoring of infectious incidence, monitoring of coverage of immunization of the population, random serological control over the immunity enable well-timed identification of distribution of the activator, effectiveness of the taken measures and forecasting. The serological blood analysis was conducted in indicator groups of the population of Central Kazakhstan for the purpose of identification of antibody titre for vaccine preventable infections (measles, viral hepatitis B). Measles antibodies were defined by method of enzyme-linked assay (ELA) with test-systems "VektoKor" – Ig G ('Vektor-Best' JSC). Antibodies for HBs-antigen of hepatitis B virus in blood serum was identified by method of enzyme-linked assay (ELA) with VektoHBsAg test systems – antibodies ('Vektor-Best' JSC). The result of the analysis is positive, the concentration of IgG to measles virus in the studied sample is equal to 0.18 IU/ml or more. Protective level of concentration of anti-HBsAg makes 10 mIU/ml. The results of the study of postvaccinal measles immunity showed that the share of seropositive people made 87.7% of total number of surveyed. The level of postvaccinal immunity to measles in age groups differs. So, among people older than 56 the percentage of seropositive made 95.2%. Among people aged 15-25 were registered 87.0% seropositive, at the age of 36-45 – 86.6%. In age groups of 25-35 and 36-45 the share of seropositive people was approximately at the same level – 88.5% and 88.8% respectively. The share of people seronegative to a measles virus made 12.3%. The biggest share of seronegative people was found among people aged 36-45 – 13.4% and 15-25 – 13.0%. The analysis of results of the examined people for the existence of postvaccinal immunity to viral hepatitis B showed that from all surveyed only 33.5% have the protective level of concentration of anti-HBsAg of 10 mIU/ml and more. The biggest share of people protected from VHB virus is observed in the age group of 36-45 and makes 60%. In the indicator group – above 56 – seropositive people made 4.8%. The high percentage of seronegative people has been observed in all studied age groups from 40.0% to 95.2%. The group of people which is least protected from getting VHB is people above 56 (95.2%). The probability to get VHB is also high among young people aged 25-35, the percentage of seronegative people made 80%. Thus, the results of the conducted research testify to the need for carrying out serological monitoring of postvaccinal immunity for the purpose of operational assessment of the epidemiological situation, early identification of its changes and prediction of the approaching danger.Keywords: antibodies, blood serum, immunity, immunoglobulin
Procedia PDF Downloads 2554064 The Common Location and the Intensity of Surface Electrical Stimulation on the Thorax and Abdomen Areas: A Systematic Review
Authors: Vu Hoang Thu Huong
Abstract:
Background: Surface electrical stimulation (SES) is a popular non-invasive approach that offers a wide range of treatments for many diseases of physical therapy. It involves applying electrical stimulation to the skin via surface electrodes to stimulate nerve fibers. SES was regularly used to treat the back and upper or lower extremities, but it was rarely used to treat the chest and abdomen. SES on the thorax and abdomen should be administered with more attention because crucial organs are under those areas (i.e., heart, lungs, liver, etc.). In these areas, safety precautions are suggested, and some SES applications might even be a contraindication. The fact that physical therapists have less experience with SES in these situations can also be attributed to these. Although a few earlier studies applied it to these settings and discovered hopeful results, none of them highlight the relationship between the intensity of SES and its depth of impact for safety considerations. Objective: To assure feasibility when using SES in these areas, the purpose of this study is to summarize the common location and intensity of those areas that have been conducted in previous studies. Method: A thorough systematic review was conducted to determine the common surface electrode position for the thorax and abdomen areas. The studies with the randomized controlled design were systematically searched using inclusion and exclusion criteria through nine electronic databases, including Pubmed, Scopus, etc., between 1975 and Dec 2021. Results: Thirty-three studies with over 1800 participants and 4 types of SES (TENS, IFC, NMES, and FES) with various categories of department hospitals were found. Following an anterior, lateral, and posterior observation, the particular SES positions found that it concentrated on 6 regions (the thoracic, abdomen, upper lateral, lower lateral, upper back, and lower back regions), and its intensity for each region was also summarized. Conclusion: This systematic review figured out the popular locations of SES in the thorax and abdominal areas as well as a summarized maximum of intensity that was found in previous studies with outstanding outcomes.Keywords: surface electrical stimulation, electrical stimulation, thoracic, abdomen, abdominal.
Procedia PDF Downloads 1064063 Unsupervised Learning and Similarity Comparison of Water Mass Characteristics with Gaussian Mixture Model for Visualizing Ocean Data
Authors: Jian-Heng Wu, Bor-Shen Lin
Abstract:
The temperature-salinity relationship is one of the most important characteristics used for identifying water masses in marine research. Temperature-salinity characteristics, however, may change dynamically with respect to the geographic location and is quite sensitive to the depth at the same location. When depth is taken into consideration, however, it is not easy to compare the characteristics of different water masses efficiently for a wide range of areas of the ocean. In this paper, the Gaussian mixture model was proposed to analyze the temperature-salinity-depth characteristics of water masses, based on which comparison between water masses may be conducted. Gaussian mixture model could model the distribution of a random vector and is formulated as the weighting sum for a set of multivariate normal distributions. The temperature-salinity-depth data for different locations are first used to train a set of Gaussian mixture models individually. The distance between two Gaussian mixture models can then be defined as the weighting sum of pairwise Bhattacharyya distances among the Gaussian distributions. Consequently, the distance between two water masses may be measured fast, which allows the automatic and efficient comparison of the water masses for a wide range area. The proposed approach not only can approximate the distribution of temperature, salinity, and depth directly without the prior knowledge for assuming the regression family, but may restrict the complexity by controlling the number of mixtures when the amounts of samples are unevenly distributed. In addition, it is critical for knowledge discovery in marine research to represent, manage and share the temperature-salinity-depth characteristics flexibly and responsively. The proposed approach has been applied to a real-time visualization system of ocean data, which may facilitate the comparison of water masses by aggregating the data without degrading the discriminating capabilities. This system provides an interface for querying geographic locations with similar temperature-salinity-depth characteristics interactively and for tracking specific patterns of water masses, such as the Kuroshio near Taiwan or those in the South China Sea.Keywords: water mass, Gaussian mixture model, data visualization, system framework
Procedia PDF Downloads 1424062 Isolation and Identification of Microorganisms from Marine-Associated Samples under Laboratory Conditions
Authors: Sameen Tariq, Saira Bano, Sayyada Ghufrana Nadeem
Abstract:
The Ocean, which covers over 70% of the world's surface, is wealthy in biodiversity as well as a rich wellspring of microorganisms with huge potential. The oceanic climate is home to an expansive scope of plants, creatures, and microorganisms. Marine microbial networks, which incorporate microscopic organisms, infections, and different microorganisms, enjoy different benefits in biotechnological processes. Samples were collected from marine environments, including soil and water samples, to cultivate the uncultured marine organisms by using Zobell’s medium, Sabouraud’s dextrose agar, and casein media for this purpose. Following isolation, we conduct microscopy and biochemical tests, including gelatin, starch, glucose, casein, catalase, and carbohydrate hydrolysis for further identification. The results show that more gram-positive and gram-negative bacteria. The isolation process of marine organisms is essential for understanding their ecological roles, unraveling their biological secrets, and harnessing their potential for various applications. Marine organisms exhibit remarkable adaptations to thrive in the diverse and challenging marine environment, offering vast potential for scientific, medical, and industrial applications. The isolation process plays a crucial role in unlocking the secrets of marine organisms, understanding their biological functions, and harnessing their valuable properties. They offer a rich source of bioactive compounds with pharmaceutical potential, including antibiotics, anticancer agents, and novel therapeutics. This study is an attempt to explore the diversity and dynamics related to marine microflora and their role in biofilm formation.Keywords: marine microorganisms, ecosystem, fungi, biofilm, gram-positive, gram-negative
Procedia PDF Downloads 444061 Getting to Know the Enemy: Utilization of Phone Record Analysis Simulations to Uncover a Target’s Personal Life Attributes
Authors: David S. Byrne
Abstract:
The purpose of this paper is to understand how phone record analysis can enable identification of subjects in communication with a target of a terrorist plot. This study also sought to understand the advantages of the implementation of simulations to develop the skills of future intelligence analysts to enhance national security. Through the examination of phone reports which in essence consist of the call traffic of incoming and outgoing numbers (and not by listening to calls or reading the content of text messages), patterns can be uncovered that point toward members of a criminal group and activities planned. Through temporal and frequency analysis, conclusions were drawn to offer insights into the identity of participants and the potential scheme being undertaken. The challenge lies in the accurate identification of the users of the phones in contact with the target. Often investigators rely on proprietary databases and open sources to accomplish this task, however it is difficult to ascertain the accuracy of the information found. Thus, this paper poses two research questions: how effective are freely available web sources of information at determining the actual identification of callers? Secondly, does the identity of the callers enable an understanding of the lifestyle and habits of the target? The methodology for this research consisted of the analysis of the call detail records of the author’s personal phone activity spanning the period of a year combined with a hypothetical theory that the owner of said phone was a leader of terrorist cell. The goal was to reveal the identity of his accomplices and understand how his personal attributes can further paint a picture of the target’s intentions. The results of the study were interesting, nearly 80% of the calls were identified with over a 75% accuracy rating via datamining of open sources. The suspected terrorist’s inner circle was recognized including relatives and potential collaborators as well as financial institutions [money laundering], restaurants [meetings], a sporting goods store [purchase of supplies], and airline and hotels [travel itinerary]. The outcome of this research showed the benefits of cellphone analysis without more intrusive and time-consuming methodologies though it may be instrumental for potential surveillance, interviews, and developing probable cause for wiretaps. Furthermore, this research highlights the importance of building upon the skills of future intelligence analysts through phone record analysis via simulations; that hands-on learning in this case study emphasizes the development of the competencies necessary to improve investigations overall.Keywords: hands-on learning, intelligence analysis, intelligence education, phone record analysis, simulations
Procedia PDF Downloads 144060 Establishing a Computational Screening Framework to Identify Environmental Exposures Using Untargeted Gas-Chromatography High-Resolution Mass Spectrometry
Authors: Juni C. Kim, Anna R. Robuck, Douglas I. Walker
Abstract:
The human exposome, which includes chemical exposures over the lifetime and their effects, is now recognized as an important measure for understanding human health; however, the complexity of the data makes the identification of environmental chemicals challenging. The goal of our project was to establish a computational workflow for the improved identification of environmental pollutants containing chlorine or bromine. Using the “pattern. search” function available in the R package NonTarget, we wrote a multifunctional script that searches mass spectral clusters from untargeted gas-chromatography high-resolution mass spectrometry (GC-HRMS) for the presence of spectra consistent with chlorine and bromine-containing organic compounds. The “pattern. search” function was incorporated into a different function that allows the evaluation of clusters containing multiple analyte fragments, has multi-core support, and provides a simplified output identifying listing compounds containing chlorine and/or bromine. The new function was able to process 46,000 spectral clusters in under 8 seconds and identified over 150 potential halogenated spectra. We next applied our function to a deidentified dataset from patients diagnosed with primary biliary cholangitis (PBC), primary sclerosing cholangitis (PSC), and healthy controls. Twenty-two spectra corresponded to potential halogenated compounds in the PSC and PBC dataset, including six significantly different in PBC patients, while four differed in PSC patients. We have developed an improved algorithm for detecting halogenated compounds in GC-HRMS data, providing a strategy for prioritizing exposures in the study of human disease.Keywords: exposome, metabolome, computational metabolomics, high-resolution mass spectrometry, exposure, pollutants
Procedia PDF Downloads 136