Search results for: big data in higher education
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 35171

Search results for: big data in higher education

22721 Modeling of Wind Loads on Heliostats Installed in South Algeria of Various Pylon Height

Authors: Hakim Merarda, Mounir Aksas, Toufik Arrif, Abd Elfateh Belaid, Amor Gama, Reski Khelifi

Abstract:

Knowledge of wind loads is important to develop a heliostat with good performance. These loads can be calculated by mathematical equations based on several parameters: the density, wind velocity, the aspect ratio of the mirror (height/width) and the coefficient of the height of the tower. Measurement data of the wind velocity and the density of the air are used in a numerical simulation of wind profile that was performed on heliostats with different pylon heights, with 1m^2 mirror areas and with aspect ratio of mirror equal to 1. These measurement data are taken from the meteorological station installed in Ghardaia, Algeria. The main aim of this work is to find a mathematical correlation between the wind loads and the height of the tower.

Keywords: heliostat, solar tower power, wind loads simulation, South Algeria

Procedia PDF Downloads 540
22720 The Use of Technology in Theatrical Performances as a Tool of Audience’S Engagement

Authors: Chrysoula Bousiouta

Abstract:

Throughout the history of theatre, technology has played an important role both in influencing the relationship between performance and audience and offering different kinds of experiences. The use of technology dates back in ancient times, when the introduction of artifacts, such as “Deus ex machine” in ancient Greek theatre, started. Taking into account the key techniques and experiences used throughout history, this paper investigates how technology, through new media, influences contemporary theatre. In the context of this research, technology is defined as projections, audio environments, video-projections, sensors, tele-connections, all alongside with the performance, challenging audience’s participation. The theoretical framework of the research covers, except for the history of theatre, the theory of “experience economy” that took over the service and goods economy. The research is based on the qualitative and comparative analysis of two case studies, Contact Theatre in Manchester (United Kingdom) and Bios in Athens (Greece). The data selection includes desk research and is complemented with semi structured interviews. Building on the results of the research one could claim that the intended experience of modern/contemporary theatre is that of engagement. In this context, technology -as defined above- plays a leading role in creating it. This experience passes through and exists in the middle of the realms of entertainment, education, estheticism and escapism. Furthermore, it is observed that nowadays, theatre is not only about acting but also about performing; it is that one where the performances are unfinished without the participation of the audience. Both case studies try to achieve the experience of engagement through practices that promote the attraction of attention, the increase of imagination, the interaction, the intimacy and the true activity. These practices are achieved through the script, the scenery, the language and the environment of a performance. Contact and Bios consider technology as an intimate tool in order to accomplish the above, and they make an extended use of it. The research completes a notable record of technological techniques that modern theatres use. The use of technology, inside or outside the limits of film technique’s, helps to rivet the attention of the audience, to make performances enjoyable, to give the sense of the “unfinished” or to be used for things that take place around the spectators and force them to take action, being spect-actors. The advantage of technology is that it can be used as a hook for interaction in all stages of a performance. Further research on the field could involve exploring alternative ways of binding technology and theatre or analyzing how the performance is perceived through the use of technological artifacts.

Keywords: experience of engagement, interactive theatre, modern theatre, performance, technology

Procedia PDF Downloads 236
22719 Experience Report about the Inclusion of People with Disabilities in the Process of Testing an Accessible System for Learning Management

Authors: Marcos Devaner, Marcela Alves, Cledson Braga, Fabiano Alves, Wilton Bezerra

Abstract:

This article discusses the inclusion of people with disabilities in the process of testing an accessible system solution for distance education. The accessible system, team profile, methodologies and techniques covered in the testing process are presented. The testing process shown in this paper was designed from the experience with user. The testing process emerged from lessons learned from past experiences and the end user is present at all stages of the tests. Also, lessons learned are reported and how it was possible the maturing of the team and the methods resulting in a simple, productive and effective process.

Keywords: experience report, accessible systems, software testing, testing process, systems, e-learning

Procedia PDF Downloads 375
22718 Improve B-Tree Index’s Performance Using Lock-Free Hash Table

Authors: Zhanfeng Ma, Zhiping Xiong, Hu Yin, Zhengwei She, Aditya P. Gurajada, Tianlun Chen, Ying Li

Abstract:

Many RDBMS vendors use B-tree index to achieve high performance for point queries and range queries, and some of them also employ hash index to further enhance the performance as hash table is more efficient for point queries. However, there are extra overheads to maintain a separate hash index, for example, hash mapping for all data records must always be maintained, which results in more memory space consumption; locking, logging and other mechanisms are needed to guarantee ACID, which affects the concurrency and scalability of the system. To relieve the overheads, Hash Cached B-tree (HCB) index is proposed in this paper, which consists of a standard disk-based B-tree index and an additional in-memory lock-free hash table. Initially, only the B-tree index is constructed for all data records, the hash table is built on the fly based on runtime workload, only data records accessed by point queries are indexed using hash table, this helps reduce the memory footprint. Changes to hash table are done using compare-and-swap (CAS) without performing locking and logging, this helps improve the concurrency and avoid contention. The hash table is also optimized to be cache conscious. HCB index is implemented in SAP ASE database, compared with the standard B-tree index, early experiments and customer adoptions show significant performance improvement. This paper provides an overview of the design of HCB index and reports the experimental results.

Keywords: B-tree, compare-and-swap, lock-free hash table, point queries, range queries, SAP ASE database

Procedia PDF Downloads 273
22717 Measurement and Modelling of HIV Epidemic among High Risk Groups and Migrants in Two Districts of Maharashtra, India: An Application of Forecasting Software-Spectrum

Authors: Sukhvinder Kaur, Ashok Agarwal

Abstract:

Background: For the first time in 2009, India was able to generate estimates of HIV incidence (the number of new HIV infections per year). Analysis of epidemic projections helped in revealing that the number of new annual HIV infections in India had declined by more than 50% during the last decade (GOI Ministry of Health and Family Welfare, 2010). Then, National AIDS Control Organisation (NACO) planned to scale up its efforts in generating projections through epidemiological analysis and modelling by taking recent available sources of evidence such as HIV Sentinel Surveillance (HSS), India Census data and other critical data sets. Recently, NACO generated current round of HIV estimates-2012 through globally recommended tool “Spectrum Software” and came out with the estimates for adult HIV prevalence, annual new infections, number of people living with HIV, AIDS-related deaths and treatment needs. State level prevalence and incidence projections produced were used to project consequences of the epidemic in spectrum. In presence of HIV estimates generated at state level in India by NACO, USIAD funded PIPPSE project under the leadership of NACO undertook the estimations and projections to district level using same Spectrum software. In 2011, adult HIV prevalence in one of the high prevalent States, Maharashtra was 0.42% ahead of the national average of 0.27%. Considering the heterogeneity of HIV epidemic between districts, two districts of Maharashtra – Thane and Mumbai were selected to estimate and project the number of People-Living-with-HIV/AIDS (PLHIV), HIV-prevalence among adults and annual new HIV infections till 2017. Methodology: Inputs in spectrum included demographic data from Census of India since 1980 and sample registration system, programmatic data on ‘Alive and on ART (adult and children)’,‘Mother-Baby pairs under PPTCT’ and ‘High Risk Group (HRG)-size mapping estimates’, surveillance data from various rounds of HSS, National Family Health Survey–III, Integrated Biological and Behavioural Assessment and Behavioural Sentinel Surveillance. Major Findings: Assuming current programmatic interventions in these districts, an estimated decrease of 12% points in Thane and 31% points in Mumbai among new infections in HRGs and migrants is observed from 2011 by 2017. Conclusions: Project also validated decrease in HIV new infection among one of the high risk groups-FSWs using program cohort data since 2012 to 2016. Though there is a decrease in HIV prevalence and new infections in Thane and Mumbai, further decrease is possible if appropriate programme response, strategies and interventions are envisaged for specific target groups based on this evidence. Moreover, evidence need to be validated by other estimation/modelling techniques; and evidence can be generated for other districts of the state, where HIV prevalence is high and reliable data sources are available, to understand the epidemic within the local context.

Keywords: HIV sentinel surveillance, high risk groups, projections, new infections

Procedia PDF Downloads 202
22716 Use of Chlorophyll Meters to Assess In-Season Wheat Nitrogen Fertilizer Requirements in the Southern San Joaquin Valley

Authors: Brian Marsh

Abstract:

Nitrogen fertilizer is the most used and often the most mismanaged nutrient input. Nitrogen management has tremendous implications on crop productivity, quality and environmental stewardship. Sufficient nitrogen is needed to optimum yield and quality. Soil and in-season plant tissue testing for nitrogen status are a time consuming and expensive process. Real time sensing of plant nitrogen status can be a useful tool in managing nitrogen inputs. The objectives of this project were to assess the reliability of remotely sensed non-destructive plant nitrogen measurements compared to wet chemistry data from sampled plant tissue, develop in-season nitrogen recommendations based on remotely sensed data for improved nitrogen use efficiency and assess the potential for determining yield and quality from remotely sensed data. Very good correlations were observed between early-season remotely sensed crop nitrogen status and plant nitrogen concentrations and subsequent in-season fertilizer recommendations. The transmittance/absorbance type meters gave the most accurate readings. Early in-season fertilizer recommendation would be to apply 40 kg nitrogen per hectare plus 16 kg nitrogen per hectare for each unit difference measured with the SPAD meter between the crop and reference area or 25 kg plus 13 kg per hectare for each unit difference measured with the CCM 200. Once the crop was sufficiently fertilized meter readings became inconclusive and were of no benefit for determining nitrogen status, silage yield and quality and grain yield and protein.

Keywords: wheat, nitrogen fertilization, chlorophyll meter

Procedia PDF Downloads 376
22715 A Comparative Study between Different Techniques of Off-Page and On-Page Search Engine Optimization

Authors: Ahmed Ishtiaq, Maeeda Khalid, Umair Sajjad

Abstract:

In the fast-moving world, information is the key to success. If information is easily available, then it makes work easy. The Internet is the biggest collection and source of information nowadays, and with every single day, the data on internet increases, and it becomes difficult to find required data. Everyone wants to make his/her website at the top of search results. This can be possible when you have applied some techniques of SEO inside your application or outside your application, which are two types of SEO, onsite and offsite SEO. SEO is an abbreviation of Search Engine Optimization, and it is a set of techniques, methods to increase users of a website on World Wide Web or to rank up your website in search engine indexing. In this paper, we have compared different techniques of Onpage and Offpage SEO, and we have suggested many things that should be changed inside webpage, outside web page and mentioned some most powerful and search engine considerable elements and techniques in both types of SEO in order to gain high ranking on Search Engine.

Keywords: auto-suggestion, search engine optimization, SEO, query, web mining, web crawler

Procedia PDF Downloads 133
22714 Adaptive Beamforming with Steering Error and Mutual Coupling between Antenna Sensors

Authors: Ju-Hong Lee, Ching-Wei Liao

Abstract:

Owing to close antenna spacing between antenna sensors within a compact space, a part of data in one antenna sensor would outflow to other antenna sensors when the antenna sensors in an antenna array operate simultaneously. This phenomenon is called mutual coupling effect (MCE). It has been shown that the performance of antenna array systems can be degraded when the antenna sensors are in close proximity. Especially, in a systems equipped with massive antenna sensors, the degradation of beamforming performance due to the MCE is significantly inevitable. Moreover, it has been shown that even a small angle error between the true direction angle of the desired signal and the steering angle deteriorates the effectiveness of an array beamforming system. However, the true direction vector of the desired signal may not be exactly known in some applications, e.g., the application in land mobile-cellular wireless systems. Therefore, it is worth developing robust techniques to deal with the problem due to the MCE and steering angle error for array beamforming systems. In this paper, we present an efficient technique for performing adaptive beamforming with robust capabilities against the MCE and the steering angle error. Only the data vector received by an antenna array is required by the proposed technique. By using the received array data vector, a correlation matrix is constructed to replace the original correlation matrix associated with the received array data vector. Then, the mutual coupling matrix due to the MCE on the antenna array is estimated through a recursive algorithm. An appropriate estimate of the direction angle of the desired signal can also be obtained during the recursive process. Based on the estimated mutual coupling matrix, the estimated direction angle, and the reconstructed correlation matrix, the proposed technique can effectively cure the performance degradation due to steering angle error and MCE. The novelty of the proposed technique is that the implementation procedure is very simple and the resulting adaptive beamforming performance is satisfactory. Simulation results show that the proposed technique provides much better beamforming performance without requiring complicated complexity as compared with the existing robust techniques.

Keywords: adaptive beamforming, mutual coupling effect, recursive algorithm, steering angle error

Procedia PDF Downloads 310
22713 Stature and Gender Estimation Using Foot Measurements in South Indian Population

Authors: Jagadish Rao Padubidri, Mehak Bhandary, Sowmya J. Rao

Abstract:

Introduction: The significance of the human foot and its measurements in identifying an individual has been proved a lot of times by different studies in different geographical areas and its association to the stature and gender of the individual has been justified by many researches. In our study we have used different foot measurements including the length, width, malleol height and navicular height for establishing its association to stature and gender and to find out its accuracy. The purpose of this study is to show the relation of foot measurements with stature and gender, and to derive Multiple and Logistic regression equations for stature and gender estimation in South Indian population. Materials and Methods: The subjects for this study were 200 South Indian students out of which 100 were females and 100 were males, aged between 18 to 24 years. The data for the present study included the stature, foot length, foot breath, foot malleol height, foot navicular height of both right and left foot. Descriptive statistics, T-test and Pearson correlation coefficients were derived between stature, gender and foot measurements. The stature was estimated from right and left foot measurements for both male and female South Indian population using multiple regression analysis and logistic regression analysis for gender estimation. Results: The means, standard deviation, stature, right and left foot measurements and T-test in male population were higher than in females. LFL (Left foot length) is more than RFL (Right Foot length) in male groups, but in female groups the length of both foot are almost equal [RFL=226.6, LFL=227.1]. There is not much of difference in means of RFW (Right foot width) and LFW (Left foot width) in both the genders. Significant difference were seen in mean values of malleol and navicular height of right and left feet in male gender. No such difference was seen in female subjects. Conclusions: The study has successfully demonstrated the correlation of foot length in stature estimation in all the three study groups in both right and left foot. Next in parameters are Foot width and malleol height in estimating stature among male and female groups. Navicular height of both right and left foot showed poor relationship with stature estimation in both male and female groups. Multiple regression equations for both right and left foot measurements to estimate stature were derived with standard error ranging from 11-12 cm in males and 10-11 cm in females. The SEE was 5.8 when both male and female groups were pooled together. The logistic regression model which was derived to determine gender showed 85% accuracy and 92.5% accuracy using right and left foot measurements respectively. We believe that stature and gender can be estimated with foot measurements in South Indian population.

Keywords: foot length, gender, stature, South Indian

Procedia PDF Downloads 322
22712 Towards an Environmental Knowledge System in Water Management

Authors: Mareike Dornhoefer, Madjid Fathi

Abstract:

Water supply and water quality are key problems of mankind at the moment and - due to increasing population - in the future. Management disciplines like water, environment and quality management therefore need to closely interact, to establish a high level of water quality and to guarantee water supply in all parts of the world. Groundwater remediation is one aspect in this process. From a knowledge management perspective it is only possible to solve complex ecological or environmental problems if different factors, expert knowledge of various stakeholders and formal regulations regarding water, waste or chemical management are interconnected in form of a knowledge base. In general knowledge management focuses the processes of gathering and representing existing and new knowledge in a way, which allows for inference or deduction of knowledge for e.g. a situation where a problem solution or decision support are required. A knowledge base is no sole data repository, but a key element in a knowledge based system, thus providing or allowing for inference mechanisms to deduct further knowledge from existing facts. In consequence this knowledge provides decision support. The given paper introduces an environmental knowledge system in water management. The proposed environmental knowledge system is part of a research concept called Green Knowledge Management. It applies semantic technologies or concepts such as ontology or linked open data to interconnect different data and information sources about environmental aspects, in this case, water quality, as well as background material enriching an established knowledge base. Examples for the aforementioned ecological or environmental factors threatening water quality are among others industrial pollution (e.g. leakage of chemicals), environmental changes (e.g. rise in temperature) or floods, where all kinds of waste are merged and transferred into natural water environments. Water quality is usually determined with the help of measuring different indicators (e.g. chemical or biological), which are gathered with the help of laboratory testing, continuous monitoring equipment or other measuring processes. During all of these processes data are gathered and stored in different databases. Meanwhile the knowledge base needs to be established through interconnecting data of these different data sources and enriching its semantics. Experts may add their knowledge or experiences of previous incidents or influencing factors. In consequence querying or inference mechanisms are applied for the deduction of coherence between indicators, predictive developments or environmental threats. Relevant processes or steps of action may be modeled in form of a rule based approach. Overall the environmental knowledge system supports the interconnection of information and adding semantics to create environmental knowledge about water environment, supply chain as well as quality. The proposed concept itself is a holistic approach, which links to associated disciplines like environmental and quality management. Quality indicators and quality management steps need to be considered e.g. for the process and inference layers of the environmental knowledge system, thus integrating the aforementioned management disciplines in one water management application.

Keywords: water quality, environmental knowledge system, green knowledge management, semantic technologies, quality management

Procedia PDF Downloads 208
22711 Performance Improvement of Photovoltaic Module at Different Tilt Angle in Kuwait

Authors: Hussain Bunyan, Wesam Ali

Abstract:

In this paper we will study the performance of a Silicon Photovoltaic (PV) system with different tilt angle arrangement in Kuwait (latitude 30˚ N). In this study the PV system is installed facing south, collecting maximum solar radiation at noon, and their angles are from 00 to 900 respectively, during full year at the Solstice and Equinox periods and aiming for a higher angle than 300 with competitive output power. The results show that the performance and the output power of the PV system with 50˚ tilt angle, is equivalent to the latitude tilt angle (30˚) during a full year.

Keywords: photovoltaic model, tilt angle, solar collector, PV system performance, State of Kuwait

Procedia PDF Downloads 501
22710 Private Coded Computation of Matrix Multiplication

Authors: Malihe Aliasgari, Yousef Nejatbakhsh

Abstract:

The era of Big Data and the immensity of real-life datasets compels computation tasks to be performed in a distributed fashion, where the data is dispersed among many servers that operate in parallel. However, massive parallelization leads to computational bottlenecks due to faulty servers and stragglers. Stragglers refer to a few slow or delay-prone processors that can bottleneck the entire computation because one has to wait for all the parallel nodes to finish. The problem of straggling processors, has been well studied in the context of distributed computing. Recently, it has been pointed out that, for the important case of linear functions, it is possible to improve over repetition strategies in terms of the tradeoff between performance and latency by carrying out linear precoding of the data prior to processing. The key idea is that, by employing suitable linear codes operating over fractions of the original data, a function may be completed as soon as enough number of processors, depending on the minimum distance of the code, have completed their operations. The problem of matrix-matrix multiplication in the presence of practically big sized of data sets faced with computational and memory related difficulties, which makes such operations are carried out using distributed computing platforms. In this work, we study the problem of distributed matrix-matrix multiplication W = XY under storage constraints, i.e., when each server is allowed to store a fixed fraction of each of the matrices X and Y, which is a fundamental building of many science and engineering fields such as machine learning, image and signal processing, wireless communication, optimization. Non-secure and secure matrix multiplication are studied. We want to study the setup, in which the identity of the matrix of interest should be kept private from the workers and then obtain the recovery threshold of the colluding model, that is, the number of workers that need to complete their task before the master server can recover the product W. The problem of secure and private distributed matrix multiplication W = XY which the matrix X is confidential, while matrix Y is selected in a private manner from a library of public matrices. We present the best currently known trade-off between communication load and recovery threshold. On the other words, we design an achievable PSGPD scheme for any arbitrary privacy level by trivially concatenating a robust PIR scheme for arbitrary colluding workers and private databases and the proposed SGPD code that provides a smaller computational complexity at the workers.

Keywords: coded distributed computation, private information retrieval, secret sharing, stragglers

Procedia PDF Downloads 106
22709 Meeting the Energy Balancing Needs in a Fully Renewable European Energy System: A Stochastic Portfolio Framework

Authors: Iulia E. Falcan

Abstract:

The transition of the European power sector towards a clean, renewable energy (RE) system faces the challenge of meeting power demand in times of low wind speed and low solar radiation, at a reasonable cost. This is likely to be achieved through a combination of 1) energy storage technologies, 2) development of the cross-border power grid, 3) installed overcapacity of RE and 4) dispatchable power sources – such as biomass. This paper uses NASA; derived hourly data on weather patterns of sixteen European countries for the past twenty-five years, and load data from the European Network of Transmission System Operators-Electricity (ENTSO-E), to develop a stochastic optimization model. This model aims to understand the synergies between the four classes of technologies mentioned above and to determine the optimal configuration of the energy technologies portfolio. While this issue has been addressed before, it was done so using deterministic models that extrapolated historic data on weather patterns and power demand, as well as ignoring the risk of an unbalanced grid-risk stemming from both the supply and the demand side. This paper aims to explicitly account for the inherent uncertainty in the energy system transition. It articulates two levels of uncertainty: a) the inherent uncertainty in future weather patterns and b) the uncertainty of fully meeting power demand. The first level of uncertainty is addressed by developing probability distributions for future weather data and thus expected power output from RE technologies, rather than known future power output. The latter level of uncertainty is operationalized by introducing a Conditional Value at Risk (CVaR) constraint in the portfolio optimization problem. By setting the risk threshold at different levels – 1%, 5% and 10%, important insights are revealed regarding the synergies of the different energy technologies, i.e., the circumstances under which they behave as either complements or substitutes to each other. The paper concludes that allowing for uncertainty in expected power output - rather than extrapolating historic data - paints a more realistic picture and reveals important departures from results of deterministic models. In addition, explicitly acknowledging the risk of an unbalanced grid - and assigning it different thresholds - reveals non-linearity in the cost functions of different technology portfolio configurations. This finding has significant implications for the design of the European energy mix.

Keywords: cross-border grid extension, energy storage technologies, energy system transition, stochastic portfolio optimization

Procedia PDF Downloads 152
22708 Simultaneous Detection of Cd⁺², Fe⁺², Co⁺², and Pb⁺² Heavy Metal Ions by Stripping Voltammetry Using Polyvinyl Chloride Modified Glassy Carbon Electrode

Authors: Sai Snehitha Yadavalli, K. Sruthi, Swati Ghosh Acharyya

Abstract:

Heavy metal ions are toxic to humans and all living species when exposed in large quantities or for long durations. Though Fe acts as a nutrient, when intake is in large quantities, it becomes toxic. These toxic heavy metal ions, when consumed through water, will cause many disorders and are harmful to all flora and fauna through biomagnification. Specifically, humans are prone to innumerable diseases ranging from skin to gastrointestinal, neurological, etc. In higher quantities, they even cause cancer in humans. Detection of these toxic heavy metal ions in water is thus important. Traditionally, the detection of heavy metal ions in water has been done by techniques like Inductively Coupled Plasma Mass Spectroscopy (ICPMS) and Atomic Absorption Spectroscopy (AAS). Though these methods offer accurate quantitative analysis, they require expensive equipment and cannot be used for on-site measurements. Anodic Stripping Voltammetry is a good alternative as the equipment is affordable, and measurements can be made at the river basins or lakes. In the current study, Square Wave Anodic Stripping Voltammetry (SWASV) was used to detect the heavy metal ions in water. Literature reports various electrodes on which deposition of heavy metal ions was carried out like Bismuth, Polymers, etc. The working electrode used in this study is a polyvinyl chloride (PVC) modified glassy carbon electrode (GCE). Ag/AgCl reference electrode and Platinum counter electrode were used. Biologic Potentiostat SP 300 was used for conducting the experiments. Through this work of simultaneous detection, four heavy metal ions were successfully detected at a time. The influence of modifying GCE with PVC was studied in comparison with unmodified GCE. The simultaneous detection of Cd⁺², Fe⁺², Co⁺², Pb⁺² heavy metal ions was done using PVC modified GCE by drop casting 1 wt.% of PVC dissolved in Tetra Hydro Furan (THF) solvent onto GCE. The concentration of all heavy metal ions was 0.2 mg/L, as shown in the figure. The scan rate was 0.1 V/s. Detection parameters like pH, scan rate, temperature, time of deposition, etc., were optimized. It was clearly understood that PVC helped in increasing the sensitivity and selectivity of detection as the current values are higher for PVC-modified GCE compared to unmodified GCE. The peaks were well defined when PVC-modified GCE was used.

Keywords: cadmium, cobalt, electrochemical sensing, glassy carbon electrodes, heavy metal Ions, Iron, lead, polyvinyl chloride, potentiostat, square wave anodic stripping voltammetry

Procedia PDF Downloads 85
22707 Efficacy of Single-Dose Azithromycin Therapy for the Treatment of Chlamydia trachomatis in Patients Evaluated for Child Sexual Abuse in an Urban Health Center 2006-16

Authors: Trenton Hubbard, Kenneth Soyemi, Emily Siffermann

Abstract:

Introduction: According to the American Academy of Pediatrics (AAP) there are different weight-based recommendations for the treatment of Chlamydia trachomatis (CT) in patients who are being evaluated for sexual assault. Current AAP Red Book guidelines recommend that uncomplicated C. trachomatis anogenital infection in prepubertal patients weighing less than =<45 kg be treated with oral erythromycin 50 mg/kg/day QID for 14 days with no alternative therapies, and for patients whose weight => 45 kg are Azithromycin 1 gm PO once. Our study objective was to determine the efficacy of single-dose Azithromycin therapy for the treatment of Chlamydia trachomatis in patients weighing less than 50 kg who were evaluated for child sexual abuse in an urban setting. Methods: We conducted a retrospective chart review of historical medical records (paper and electronic) patients weighing less than 50 kg who were evaluated for child sexual abuse and subsequently treated for C. trachomatis infection with Azithromycin (20 mg/kg PO once up to a maximum 1 gm) and received a Test of Cure (TOC) from 2006-2016. Qualitative variables were expressed as percentages. Quantitative variables were expressed as mean values (+/- standard deviation [SD]) if they followed a normal distribution or as median values (interquartile range[IQR]) if they did not. Wilcoxson two-sample test was used to compare means of Azithromycin Dose, mg/kg, and TOC timing between treatment responders and non-responders. Results: We reviewed records of 34 patients, average age (SD) was 5.4 (2.0) years, 33 (97%) were treated for CT and 1(3%) for both GC and CT. 25 (74%) were females. Urine PCR was the most commonly used test at evaluation and as TOC with 13 (38%) patients completing both tests. The average (SD) dose of Azithromycin at treatment was 470 (136) mg and average (SD) mg/kg dose of 20 (1.9) mg/kg for all patients. Median (IQR) timing for TOC testing was 19 (14-26) days. Of the 33 with complete data 25 (74%) had a negative TOC. When compared with treatment non-responders (TOC failures), treatment responders received higher doses (average dose (SD) received 495 (139) vs 401(110), P 0.06)); similar average (SD) weight base dosing received (20.8(2.0) vs 19.7 (1.5), P 0.15)), and earlier average (SD)TOC test timing (18.8 (5.6) vs 32 (28.6) P 0.02)). Conclusion: Azithromycin dosing appears to be efficacious in the treatment of CT post sexual assault as majority of patients responded. Although treatment responders and non-responders received similar weight based doses, there is need for additional studies to understand variances and predictors of response.

Keywords: child sexual abuse, chlmaydia trachmotis infection, single-dose azithromycin, weight less than or equal to 45 kilograms

Procedia PDF Downloads 277
22706 Effective Use of Visuals in Teaching Mathematics

Authors: Gohar Marikyan

Abstract:

This article is about investigating how to effectively use visuals in teaching introductory mathematics. The analysis showed the use of visuals in teaching introductory mathematics can be an effective tool for enhancing students’ learning and engagement in mathematics. The use of visuals was particularly effective for teaching concepts of numbers, operations with whole numbers, and properties of operations. The analysis also provides strong evidence that the effectiveness of visuals varied depending on the way the visuals are used. Furthermore, the analysis revealed that the use of visuals in mathematics instruction had a positive impact on student’s attitudes toward mathematics, with students showing higher levels of motivation and enjoyment in mathematics classes.

Keywords: analytical thinking skills, instructional strategies with visuals, introductory mathematics, student engagement and motivation

Procedia PDF Downloads 106
22705 Comparisons of Surveying with Terrestrial Laser Scanner and Total Station for Volume Determination of Overburden and Coal Excavations in Large Open-Pit Mine

Authors: B. Keawaram, P. Dumrongchai

Abstract:

The volume of overburden and coal excavations in open-pit mine is generally determined by conventional survey such as total station. This study aimed to evaluate the accuracy of terrestrial laser scanner (TLS) used to measure overburden and coal excavations, and to compare TLS survey data sets with the data of the total station. Results revealed that, the reference points measured with the total station showed 0.2 mm precision for both horizontal and vertical coordinates. When using TLS on the same points, the standard deviations of 4.93 cm and 0.53 cm for horizontal and vertical coordinates, respectively, were achieved. For volume measurements covering the mining areas of 79,844 m2, TLS yielded the mean difference of about 1% and the surface error margin of 6 cm at the 95% confidence level when compared to the volume obtained by total station.

Keywords: mine, survey, terrestrial laser scanner, total station

Procedia PDF Downloads 364
22704 Satellite Derived Snow Cover Status and Trends in the Indus Basin Reservoir

Authors: Muhammad Tayyab Afzal, Muhammad Arslan, Mirza Muhammad Waqar

Abstract:

Snow constitutes an important component of the cryosphere, characterized by high temporal and spatial variability. Because of the contribution of snow melt to water availability, snow is an important focus for research on climate change and adaptation. MODIS satellite data have been used to identify spatial-temporal trends in snow cover in the upper Indus basin. For this research MODIS satellite 8 day composite data of medium resolution (250m) have been analysed from 2001-2005.Pixel based supervised classification have been performed and extent of snow have been calculated of all the images. Results show large variation in snow cover between years while an increasing trend from west to east is observed. Temperature data for the Upper Indus Basin (UIB) have been analysed for seasonal and annual trends over the period 2001-2005 and calibrated with the results acquired by the research. From the analysis it is concluded that there are indications that regional warming is one of the factor that is affecting the hydrology of the upper Indus basin due to accelerated glacial melting during the simulation period, stream flow in the upper Indus basin can be predicted with a high degree of accuracy. This conclusion is also supported by the research of ICIMOD in which there is an observation that the average annual precipitation over a five year period is less than the observed stream flow and supported by positive temperature trends in all seasons.

Keywords: indus basin, MODIS, remote sensing, snow cover

Procedia PDF Downloads 370
22703 Performance of Photovoltaic Module at Different Tilt Angles

Authors: Hussain Bunyan, Wesam Ali

Abstract:

In this paper we will study the performance of a Silicon Photovoltaic (PV) system with different tilt angle arrangement in Kuwait (latitude 30˚ N). In the study the PV system is installed facing South, collecting maximum solar radiation at noon, and their angles are from 00 to 900 respectively, during full year at the Solstice and Equinox periods, aiming for a higher angle than 300 with competitive output power. The results show that the performance and the output power of the PV system with 50˚ tilt angle, is equivalent to the latitude tilt angle (30˚) during a full year.

Keywords: photovoltaic model, tilt angle, solar collector, PV system performance, State of Kuwait

Procedia PDF Downloads 474
22702 Enhancing Scalability in Ethereum Network Analysis: Methods and Techniques

Authors: Stefan K. Behfar

Abstract:

The rapid growth of the Ethereum network has brought forth the urgent need for scalable analysis methods to handle the increasing volume of blockchain data. In this research, we propose efficient methodologies for making Ethereum network analysis scalable. Our approach leverages a combination of graph-based data representation, probabilistic sampling, and parallel processing techniques to achieve unprecedented scalability while preserving critical network insights. Data Representation: We develop a graph-based data representation that captures the underlying structure of the Ethereum network. Each block transaction is represented as a node in the graph, while the edges signify temporal relationships. This representation ensures efficient querying and traversal of the blockchain data. Probabilistic Sampling: To cope with the vastness of the Ethereum blockchain, we introduce a probabilistic sampling technique. This method strategically selects a representative subset of transactions and blocks, allowing for concise yet statistically significant analysis. The sampling approach maintains the integrity of the network properties while significantly reducing the computational burden. Graph Convolutional Networks (GCNs): We incorporate GCNs to process the graph-based data representation efficiently. The GCN architecture enables the extraction of complex spatial and temporal patterns from the sampled data. This combination of graph representation and GCNs facilitates parallel processing and scalable analysis. Distributed Computing: To further enhance scalability, we adopt distributed computing frameworks such as Apache Hadoop and Apache Spark. By distributing computation across multiple nodes, we achieve a significant reduction in processing time and enhanced memory utilization. Our methodology harnesses the power of parallelism, making it well-suited for large-scale Ethereum network analysis. Evaluation and Results: We extensively evaluate our methodology on real-world Ethereum datasets covering diverse time periods and transaction volumes. The results demonstrate its superior scalability, outperforming traditional analysis methods. Our approach successfully handles the ever-growing Ethereum data, empowering researchers and developers with actionable insights from the blockchain. Case Studies: We apply our methodology to real-world Ethereum use cases, including detecting transaction patterns, analyzing smart contract interactions, and predicting network congestion. The results showcase the accuracy and efficiency of our approach, emphasizing its practical applicability in real-world scenarios. Security and Robustness: To ensure the reliability of our methodology, we conduct thorough security and robustness evaluations. Our approach demonstrates high resilience against adversarial attacks and perturbations, reaffirming its suitability for security-critical blockchain applications. Conclusion: By integrating graph-based data representation, GCNs, probabilistic sampling, and distributed computing, we achieve network scalability without compromising analytical precision. This approach addresses the pressing challenges posed by the expanding Ethereum network, opening new avenues for research and enabling real-time insights into decentralized ecosystems. Our work contributes to the development of scalable blockchain analytics, laying the foundation for sustainable growth and advancement in the domain of blockchain research and application.

Keywords: Ethereum, scalable network, GCN, probabilistic sampling, distributed computing

Procedia PDF Downloads 56
22701 Human Creativity through Dooyeweerd's Philosophy: The Case of Creative Diagramming

Authors: Kamaran Fathulla

Abstract:

Human creativity knows no bounds. More than a millennia ago humans have expressed their knowledge on cave walls and on clay artefacts. Using visuals such as diagrams and paintings have always provided us with a natural and intuitive medium for expressing such creativity. Making sense of human generated visualisation has been influenced by western scientific philosophies which are often reductionist in their nature. Theoretical frameworks such as those delivered by Peirce dominated our views of how to make sense of visualisation where a visual is seen as an emergent property of our thoughts. Others have reduced the richness of human-generated visuals to mere shapes drawn on a piece of paper or on a screen. This paper introduces an alternate framework where the centrality of human functioning is given explicit and richer consideration through the multi aspectual philosophical works of Herman Dooyeweerd. Dooyeweerd's framework of understanding reality was based on fifteen aspects of reality, each having a distinct core meaning. The totality of the aspects formed a ‘rainbow’ like spectrum of meaning. The thesis of this approach is that meaningful human functioning in most cases involves the diversity of all aspects working in synergy and harmony. Illustration of the foundations and applicability of this approach is underpinned in the case of humans use of diagramming for creative purposes, particularly within an educational context. Diagrams play an important role in education. Students and lecturers use diagrams as a powerful tool to aid their thinking. However, research into the role of diagrams used in education continues to reveal difficulties students encounter during both processes of interpretation and construction of diagrams. Their main problems shape up students difficulties with diagrams. The ever-increasing diversity of diagrams' types coupled with the fact that most real-world diagrams often contain a mix of these different types of diagrams such as boxes and lines, bar charts, surfaces, routes, shapes dotted around the drawing area, and so on with each type having its own distinct set of static and dynamic semantics. We argue that the persistence of these problems is grounded in our existing ways of understanding diagrams that are often reductionist in their underpinnings driven by a single perspective or formalism. In this paper, we demonstrate the limitations of these approaches in dealing with the three problems. Consequently, we propose, discuss, and demonstrate the potential of a nonreductionist framework for understanding diagrams based on Symbolic and Spatial Mappings (SySpM) underpinned by Dooyeweerd philosophy. The potential of the framework to account for the meaning of diagrams is demonstrated by applying it to a real-world case study physics diagram.

Keywords: SySpM, drawing style, mapping

Procedia PDF Downloads 227
22700 The Impact of the Method of Extraction on 'Chemchali' Olive Oil Composition in Terms of Oxidation Index, and Chemical Quality

Authors: Om Kalthoum Sallem, Saidakilani, Kamiliya Ounaissa, Abdelmajid Abid

Abstract:

Introduction and purposes: Olive oil is the main oil used in the Mediterranean diet. Virgin olive oil is valued for its organoleptic and nutritional characteristics and is resistant to oxidation due to its high monounsaturated fatty acid content (MUFAs), and low polyunsaturates (PUFAs) and the presence of natural antioxidants such as phenols, tocopherols and carotenoids. The fatty acid composition, especially the MUFA content, and the natural antioxidants provide advantages for health. The aim of the present study was to examine the impact of method of extraction on the chemical profiles of ‘Chemchali’ olive oil variety, which is cultivated in the city of Gafsa, and to compare it with chetoui and chemchali varieties. Methods: Our study is a qualitative prospective study that deals with ‘Chemchali’ olive oil variety. Analyses were conducted during three months (from December to February) in different oil mills in the city of Gafsa. We have compared ‘Chemchali’ olive oil obtained by continuous method to this obtained by superpress method. Then we have analyzed quality index parameters, including free fatty acid content (FFA), acidity, and UV spectrophotometric characteristics and other physico-chemical data [oxidative stability, ß-carotene, and chlorophyll pigment composition]. Results: Olive oil resulting from super press method compared with continuous method is less acid(0,6120 vs. 0,9760), less oxydazible(K232:2,478 vs. 2,592)(k270:0,216 vs. 0,228), more rich in oleic acid(61,61% vs. 66.99%), less rich in linoleic acid(13,38% vs. 13,98 %), more rich in total chlorophylls pigments (6,22 ppm vs. 3,18 ppm ) and ß-carotene (3,128 mg/kg vs. 1,73 mg/kg). ‘Chemchali’ olive oil showed more equilibrated total content in fatty acids compared with the varieties ’Chemleli’ and ‘Chetoui’. Gafsa’s variety ’Chemlali’ have significantly less saturated and polyunsaturated fatty acids. Whereas it has a higher content in monounsaturated fatty acid C18:2, compared with the two other varieties. Conclusion: The use of super press method had benefic effects on general chemical characteristics of ‘Chemchali’ olive oil, maintaining the highest quality according to the ecocert legal standards. In light of the results obtained in this study, a more detailed study is required to establish whether the differences in the chemical properties of oils are mainly due to agronomic and climate variables or, to the processing employed in oil mills.

Keywords: olive oil, extraction method, fatty acids, chemchali olive oil

Procedia PDF Downloads 367
22699 ACBM: Attention-Based CNN and Bi-LSTM Model for Continuous Identity Authentication

Authors: Rui Mao, Heming Ji, Xiaoyu Wang

Abstract:

Keystroke dynamics are widely used in identity recognition. It has the advantage that the individual typing rhythm is difficult to imitate. It also supports continuous authentication through the keyboard without extra devices. The existing keystroke dynamics authentication methods based on machine learning have a drawback in supporting relatively complex scenarios with massive data. There are drawbacks to both feature extraction and model optimization in these methods. To overcome the above weakness, an authentication model of keystroke dynamics based on deep learning is proposed. The model uses feature vectors formed by keystroke content and keystroke time. It ensures efficient continuous authentication by cooperating attention mechanisms with the combination of CNN and Bi-LSTM. The model has been tested with Open Data Buffalo dataset, and the result shows that the FRR is 3.09%, FAR is 3.03%, and EER is 4.23%. This proves that the model is efficient and accurate on continuous authentication.

Keywords: keystroke dynamics, identity authentication, deep learning, CNN, LSTM

Procedia PDF Downloads 137
22698 Prevalence Of Listeria And Salmonella Contamination In Fda Recalled Foods

Authors: Oluwatofunmi Musa-Ajakaiye, Paul Olorunfemi M.D MPH, John Obafaiye

Abstract:

Introduction: The U.S Food and Drug Administration (FDA) reports the public notices for recalled FDA-regulated products over periods of time. It study reviewed the primary reasons for recalls of products of various types over a period of 7 years. Methods: The study analyzed data provided in the FDA’s archived recalls for the years 2010-2017. It identified the various reasons for product recalls in the categories of foods, beverages, drugs, medical devices, animal and veterinary products, and dietary supplements. Using SPSS version 29, descriptive statistics and chi-square analysis of the data were performed. Results (numbers, percentages, p-values, chi-square): Over the period of analysis, a total of 931 recalls were reported. The most frequent reason for recalls was undeclared products (36.7%). The analysis showed that the most recalled product type in the data set was foods and beverages, representing 591 of all recalled products (63.5%).In addition, it was observed that foods and beverages represent 77.2% of products recalled due to the presence of microorganisms. Also, a sub-group analysis of recall reasons of food and beverages found that the most prevalent reason for such recalls was undeclared products (50.1%) followed by Listeria (17.3%) then Salmonella (13.2%). Conclusion: This analysis shows that foods and beverages have the greatest percentages of total recalls due to the presence of undeclared products listeria contamination and Salmonella contamination. The prevalence of Salmonella and Listeria contamination suggests that there is a high risk of microbial contamination in FDA-approved products and further studies on the effects of such contamination must be conducted to ensure consumer safety.

Keywords: food, beverages, listeria, salmonella, FDA, contamination, microbial

Procedia PDF Downloads 48
22697 Retail Managers’ Perception on Coca-Cola Company’s Success of Glass Package Recovery and Recycling in Nairobi, Kenya

Authors: Brigitte Wabuyabo-Okonga

Abstract:

Little research has been done to establish the level of success of Coca Cola Company in recycling and reusing their glass bottles. This paper attempts to establish retail managers’ perception of the company’s self acclaimed success. Retail managers of supermarkets in the CBD of Nairobi, Kenya were considered for the study. Data were collected through questionnaires and analyzed using descriptive (mean, frequencies and percentages) and inferential statistics (correlation analysis) were used to analyze the data. The study found out that there is relative success although a lot needs to be done. For example, improving in communicating policy issues and in practice enhance the actual collection of broken and/or non-broken Coca Cola Company glass bottles through providing drop-off points in open areas such as on the streets and in parks.

Keywords: Coca Cola Company glass bottles, Kenya, Nairobi, packaging, retail manager

Procedia PDF Downloads 302
22696 Science School Was Burned: A Case Study of Crisis Management in Thailand

Authors: Proud Arunrangsiwed

Abstract:

This study analyzes the crisis management and image repair strategies during the crisis of Mahidol Wittayanusorn School (MWIT) library burning. The library of this school was burned by a 16-year-old-male student on June 6th, 2010. This student blamed the school that the lesson was difficult, and other students were selfish. Although no one was in the building during the fire, it had caused damage to the building, books and electronic supplies around 130 million bahts (4.4 million USD). This event aroused many discourses arguing about the education system and morality. The strategies which were used during crisis were denial, shift the blame, bolstering, minimization, and uncertainty reduction. The results of using these strategies appeared after the crisis. That was the numbers of new students, who registered for the examination to get into this school in the later years, have remained the same.

Keywords: school, crisis management, violence, image repair strategies, uncertainty, burn

Procedia PDF Downloads 461
22695 Satellite LiDAR-Based Digital Terrain Model Correction using Gaussian Process Regression

Authors: Keisuke Takahata, Hiroshi Suetsugu

Abstract:

Forest height is an important parameter for forest biomass estimation, and precise elevation data is essential for accurate forest height estimation. There are several globally or nationally available digital elevation models (DEMs) like SRTM and ASTER. However, its accuracy is reported to be low particularly in mountainous areas where there are closed canopy or steep slope. Recently, space-borne LiDAR, such as the Global Ecosystem Dynamics Investigation (GEDI), have started to provide sparse but accurate ground elevation and canopy height estimates. Several studies have reported the high degree of accuracy in their elevation products on their exact footprints, while it is not clear how this sparse information can be used for wider area. In this study, we developed a digital terrain model correction algorithm by spatially interpolating the difference between existing DEMs and GEDI elevation products by using Gaussian Process (GP) regression model. The result shows that our GP-based methodology can reduce the mean bias of the elevation data from 3.7m to 0.3m when we use airborne LiDAR-derived elevation information as ground truth. Our algorithm is also capable of quantifying the elevation data uncertainty, which is critical requirement for biomass inventory. Upcoming satellite-LiDAR missions, like MOLI (Multi-footprint Observation Lidar and Imager), are expected to contribute to the more accurate digital terrain model generation.

Keywords: digital terrain model, satellite LiDAR, gaussian processes, uncertainty quantification

Procedia PDF Downloads 159
22694 Multiaxial Fatigue in Thermal Elastohydrodynamic Lubricated Contacts with Asperities and Slip

Authors: Carl-Magnus Everitt, Bo Alfredsson

Abstract:

Contact mechanics and tribology have been combined with fundamental fatigue and fracture mechanics to form the asperity mechanism which supplies an explanation for the surface-initiated rolling contact fatigue damage, called pitting or spalling. The cracks causing the pits initiates at one surface point and thereafter they slowly grow into the material before chipping of a material piece to form the pit. In the current study, the lubrication aspects on fatigue initiation are simulated by passing a single asperity through a thermal elastohydrodynamic lubricated, TEHL, contact. The physics of the lubricant was described with Reynolds equation and the lubricants pressure-viscosity relation was modeled by Roelands equation, formulated to include temperature dependence. A pressure dependent shear limit was incorporated. To capture the full phenomena of the sliding contact the temperature field was resolved through the incorporation of the energy flow. The heat was mainly generated due to shearing of the lubricant and from dry friction where metal contact occurred. The heat was then transported, and conducted, away by the solids and the lubricant. The fatigue damage caused by the asperities was evaluated through Findley’s fatigue criterion. The results show that asperities, in the size of surface roughness found in applications, may cause surface initiated fatigue damage and crack initiation. The simulations also show that the asperities broke through the lubricant in the inlet, causing metal to metal contact with high friction. When the asperities thereafter moved through the contact, the sliding provided the asperities with lubricant releasing the metal contact. The release of metal contact was possible due to the high viscosity the lubricant obtained from the high pressure. The metal contact in the inlet caused higher friction which increased the risk of fatigue damage. Since the metal contact occurred in the inlet it increased the fatigue risk more for asperities subjected to negative slip than positive slip. Therefore the fatigue evaluations showed that the asperities subjected to negative slip yielded higher fatigue stresses than the asperities subjected to positive slip of equal magnitude. This is one explanation for why pitting is more common in the dedendum than the addendum on pinion gear teeth. The simulations produced further validation for the asperity mechanism by showing that asperities cause surface initiated fatigue and crack initiation.

Keywords: fatigue, rolling, sliding, thermal elastohydrodynamic

Procedia PDF Downloads 107
22693 Comparison of the H-Index of Researchers of Google Scholar and Scopus

Authors: Adian Fatchur Rochim, Abdul Muis, Riri Fitri Sari

Abstract:

H-index has been widely used as a performance indicator of researchers around the world especially in Indonesia. The Government uses Scopus and Google scholar as indexing references in providing recognition and appreciation. However, those two indexing services yield to different H-index values. For that purpose, this paper evaluates the difference of the H-index from those services. Researchers indexed by Webometrics, are used as reference’s data in this paper. Currently, Webometrics only uses H-index from Google Scholar. This paper observed and compared corresponding researchers’ data from Scopus to get their H-index score. Subsequently, some researchers with huge differences in score are observed in more detail on their paper’s publisher. This paper shows that the H-index of researchers in Google Scholar is approximately 2.45 times of their Scopus H-Index. Most difference exists due to the existence of uncertified publishers, which is considered in Google Scholar but not in Scopus.

Keywords: Google Scholar, H-index, Scopus, performance indicator

Procedia PDF Downloads 258
22692 A Method against Obsolescence of Three-Dimensional Archaeological Collection. Two Cases of Study from Qubbet El-Hawa Necropolis, Aswan, Egypt

Authors: L. Serrano-Lara, J.M Alba-Gómez

Abstract:

Qubbet el–Hawa Project has been documented archaeological artifacts as 3d models by laser scanning technique since 2015. Currently, research has obtained the right methodology to develop a high accuracy photographic texture for each geometrical 3D model. Furthermore, the right methodology to attach the complete digital surrogate into a 3DPDF document has been obtained; it is used as a catalogue worksheet that brings archaeological data and, at the same time, allows us to obtain precise measurements, volume calculations and cross-section mapping of each scanned artifact. This validated archaeological documentation is the first step for dissemination, application as Qubbet el-Hawa Virtual Museum, and, moreover, multi-sensory experience through 3D print archaeological artifacts. Material culture from four funerary complexes constructed in West Aswan has become physical replicas opening the archaeological research process itself and offering creative possibilities on museology or educational projects. This paper shares a method of acquiring texture for scanning´s output product in order to achieve a 3DPDF archaeological cataloguing, and, on the other hand, to allow the colorfully 3D printing of singular archaeological artifacts. The proposed method has undergone two concrete cases, a polychrome wooden ushabti, and, a cartonnage mask belonging to a lady, bought recovered on intact tomb QH34aa. Both 3D model results have been implemented on three main applications, archaeological 3D catalogue, public dissemination activities, and the 3D artifact model in a bachelor education program. Due to those three already mentioned applications, productive interaction among spectator and three-dimensional artifact have been increased; moreover, functionality as archaeological documentation has been consolidated. Finding the right methodology to assign a specific color to each vector on the geometric 3D model, we had been achieved two essential archaeological applications. Firstly, 3DPDF as a display document for an archaeological catalogue, secondly, the possibility to obtain a colored 3d printed object to be displayed in public exhibitions. Obsolescences 3D models have become updated archaeological documentation of QH43aa tomb cultural material. Therefore, Qubbet el-Hawa Project has been actualized the educational potential of its results thanks to a multi-sensory experience that arose from 3d scanned´s archaeological artifacts.

Keywords: 3D printed, 3D scanner, Middle Kingdom, Qubbet el-Hawa necropolis, virtual archaeology

Procedia PDF Downloads 128