Search results for: software requirement engineering
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 8336

Search results for: software requirement engineering

6956 Centralizing the Teaching Process in Intelligent Tutoring System Architectures

Authors: Nikolaj Troels Graf Von Malotky, Robin Nicolay, Alke Martens

Abstract:

There exist a plethora of architectures for ITSs (Intelligent Tutoring Systems). A thorough analysis and comparison of the architectures revealed, that in most cases the architecture extensions are evolutionary grown, reflecting state of the art trends of each decade. However, from the perspective of software engineering, the main aspect of an ITS has not been reflected in any of these architectures, yet. From the perspective of cognitive research, the construction of the teaching process is what makes an ITS 'intelligent' regarding the spectrum of interaction with the students. Thus, in our approach, we focus on a behavior based architecture, which is based on the main teaching processes. To create a new general architecture for ITS, we have to define the prerequisites. This paper analyzes the current state of the existing architectures and derives rules for the behavior of ITS. It is presenting a teaching process for ITSs to be used together with the architecture.

Keywords: intelligent tutoring, ITS, tutoring process, system architecture, interaction process

Procedia PDF Downloads 380
6955 A Study on Multidimensional Locus of Control and the Procrastinating Behavior in Employees

Authors: Richa Mishra, Sonia Munjal

Abstract:

In this increasingly hectic and competitive climate, employees are expected to manage the resources available to them to perform their work. However, many are wasting the most precious and scarce resource at their disposal, time, by procrastinating on tasks and thereby costing themselves and their organizations. As timely performance is a requirement of most jobs, procrastination is particularly problematic in the workplace. Evidence suggests that procrastination and poor performance go hand-in-hand, as procrastinators miss more deadlines than non-procrastinators and make more errors and work at a slower speed than non-procrastinators when performing timed tasks. This research is hence an effort to add a little in the sparse knowledge base. It is an effort to throw light on the relationship of Levenson’s multi dimensions of locus of control and also an effort to identify if it is one of the causes and of employees procrastination which have not been explored earlier. The study also explores the effect and relationship of multidimensional locus of control and various levels of stress on procrastination. The results of the research have ascertained that there is significant impact of LOC dimensions on the procrastinating behavior of the employees. One of the major findings to emerge from the current research that managers with powerful others as their LOC dimensions were least procrastinating, contradicts the previous research results that external procrastinate more than internals.

Keywords: Multidimensional Locus of Control, workplace procrastination, employee behaviour, manufacturing industry

Procedia PDF Downloads 242
6954 Utilizing Topic Modelling for Assessing Mhealth App’s Risks to Users’ Health before and during the COVID-19 Pandemic

Authors: Pedro Augusto Da Silva E Souza Miranda, Niloofar Jalali, Shweta Mistry

Abstract:

BACKGROUND: Software developers utilize automated solutions to scrape users’ reviews to extract meaningful knowledge to identify problems (e.g., bugs, compatibility issues) and possible enhancements (e.g., users’ requests) to their solutions. However, most of these solutions do not consider the health risk aspects to users. Recent works have shed light on the importance of including health risk considerations in the development cycle of mHealth apps to prevent harm to its users. PROBLEM: The COVID-19 Pandemic in Canada (and World) is currently forcing physical distancing upon the general population. This new lifestyle made the usage of mHealth applications more essential than ever, with a projected market forecast of 332 billion dollars by 2025. However, this new insurgency in mHealth usage comes with possible risks to users’ health due to mHealth apps problems (e.g., wrong insulin dosage indication due to a UI error). OBJECTIVE: These works aim to raise awareness amongst mHealth developers of the importance of considering risks to users’ health within their development lifecycle. Moreover, this work also aims to help mHealth developers with a Proof-of-Concept (POC) solution to understand, process, and identify possible health risks to users of mHealth apps based on users’ reviews. METHODS: We conducted a mixed-method study design. We developed a crawler to mine the negative reviews from two samples of mHealth apps (my fitness, medisafe) from the Google Play store users. For each mHealth app, we performed the following steps: • The reviews are divided into two groups, before starting the COVID-19 (reviews’ submission date before 15 Feb 2019) and during the COVID-19 (reviews’ submission date starts from 16 Feb 2019 till Dec 2020). For each period, the Latent Dirichlet Allocation (LDA) topic model was used to identify the different clusters of reviews based on similar topics of review The topics before and during COVID-19 are compared, and the significant difference in frequency and severity of similar topics are identified. RESULTS: We successfully scraped, filtered, processed, and identified health-related topics in both qualitative and quantitative approaches. The results demonstrated the similarity between topics before and during the COVID-19.

Keywords: natural language processing (NLP), topic modeling, mHealth, COVID-19, software engineering, telemedicine, health risks

Procedia PDF Downloads 129
6953 Zonal and Sequential Extraction Design for Large Flat Space to Achieve Perpetual Tenability

Authors: Mingjun Xu, Man Pun Wan

Abstract:

This study proposed an effective smoke control strategy for the large flat space with a low ceiling to achieve the requirement of perpetual tenability. For the large flat space with a low ceiling, the depth of the smoke reservoir is very shallow, and it is difficult to perpetually constrain the smoke within a limited space. A series of numerical tests were conducted to determine the smoke strategy. A zonal design i.e., the fire zone and two adjacent zones was proposed and validated to be effective in controlling smoke. Once a fire happens in a compartment space, the Engineered Smoke Control (ESC) system will be activated in three zones i.e., the fire zone, in which the fire happened, and two adjacent zones. The smoke can be perpetually constrained within the three smoke zones. To further improve the extraction efficiency, sequential activation of the ESC system within the 3 zones turned out to be more efficient than simultaneous activation. Additionally, the proposed zonal and sequential extraction design can reduce the mechanical extraction flow rate by up to 40.7 % as compared to the conventional method, which is much more economical than that of the conventional method.

Keywords: performance-based design, perpetual tenability, smoke control, fire plume

Procedia PDF Downloads 72
6952 Interpreting Privacy Harms from a Non-Economic Perspective

Authors: Christopher Muhawe, Masooda Bashir

Abstract:

With increased Internet Communication Technology(ICT), the virtual world has become the new normal. At the same time, there is an unprecedented collection of massive amounts of data by both private and public entities. Unfortunately, this increase in data collection has been in tandem with an increase in data misuse and data breach. Regrettably, the majority of data breach and data misuse claims have been unsuccessful in the United States courts for the failure of proof of direct injury to physical or economic interests. The requirement to express data privacy harms from an economic or physical stance negates the fact that not all data harms are physical or economic in nature. The challenge is compounded by the fact that data breach harms and risks do not attach immediately. This research will use a descriptive and normative approach to show that not all data harms can be expressed in economic or physical terms. Expressing privacy harms purely from an economic or physical harm perspective negates the fact that data insecurity may result into harms which run counter the functions of privacy in our lives. The promotion of liberty, selfhood, autonomy, promotion of human social relations and the furtherance of the existence of a free society. There is no economic value that can be placed on these functions of privacy. The proposed approach addresses data harms from a psychological and social perspective.

Keywords: data breach and misuse, economic harms, privacy harms, psychological harms

Procedia PDF Downloads 195
6951 Experimental Investigation to Find Transition Temperature of VG 30 Binder

Authors: D. Latha, V. Sunitha, Samson Mathew

Abstract:

In India, most of the pavement is laid by bituminous road and the consumption of binder is high for pavement construction and also modified binders are used to satisfy any specific pavement requirement. Since the binders are visco-elastic material which is having the mechanical properties of binder transition from visco-elastic solid to visco-elastic fluid. In this paper, two different protocols were used to measure the viscosity property of binder using a Brookfield Viscometer and there is a need to find the appropriate mixing and compaction temperatures of various types of binders which can result in complete aggregate coating and adequate field density of HMA mixtures. The aim of this work is to find the transition temperature from Non-Newtonian behavior to Newtonian behavior of the binder by adopting a steady shear protocol and the shear rate ramp protocol. The transition from non-Newtonian to Newtonian can occur through an increase of temperature and shear of the material. The test has been conducted for unmodified binder VG 30. The transition temperature was found in the unmodified binder VG is 120oC. So the application of both modified binder and unmodified binder in the pavement construction needs to be studied properly by considering temperature and traffic loading factors of the respective project site.

Keywords: unmodified and modified binders, Brookfield viscometer, transition temperature, steady shear and shear rate protocol

Procedia PDF Downloads 213
6950 Low Probability of Intercept (LPI) Signal Detection and Analysis Using Choi-Williams Distribution

Authors: V. S. S. Kumar, V. Ramya

Abstract:

In the modern electronic warfare, the signal scenario is changing at a rapid pace with the introduction of Low Probability of Intercept (LPI) radars. In the modern battlefield, radar system faces serious threats from passive intercept receivers such as Electronic Attack (EA) and Anti-Radiation Missiles (ARMs). To perform necessary target detection and tracking and simultaneously hide themselves from enemy attack, radar systems should be LPI. These LPI radars use a variety of complex signal modulation schemes together with pulse compression with the aid of advancement in signal processing capabilities of the radar such that the radar performs target detection and tracking while simultaneously hiding enemy from attack such as EA etc., thus posing a major challenge to the ES/ELINT receivers. Today an increasing number of LPI radars are being introduced into the modern platforms and weapon systems so these LPI radars created a requirement for the armed forces to develop new techniques, strategies and equipment to counter them. This paper presents various modulation techniques used in generation of LPI signals and development of Time Frequency Algorithms to analyse those signals.

Keywords: anti-radiation missiles, cross terms, electronic attack, electronic intelligence, electronic warfare, intercept receiver, low probability of intercept

Procedia PDF Downloads 469
6949 The Validation of RadCalc for Clinical Use: An Independent Monitor Unit Verification Software

Authors: Junior Akunzi

Abstract:

In the matter of patient treatment planning quality assurance in 3D conformational therapy (3D-CRT) and volumetric arc therapy (VMAT or RapidArc), the independent monitor unit verification calculation (MUVC) is an indispensable part of the process. Concerning 3D-CRT treatment planning, the MUVC can be performed manually applying the standard ESTRO formalism. However, due to the complex shape and the amount of beams in advanced treatment planning technic such as RapidArc, the manual independent MUVC is inadequate. Therefore, commercially available software such as RadCalc can be used to perform the MUVC in complex treatment planning been. Indeed, RadCalc (version 6.3 LifeLine Inc.) uses a simplified Clarkson algorithm to compute the dose contribution for individual RapidArc fields to the isocenter. The purpose of this project is the validation of RadCalc in 3D-CRT and RapidArc for treatment planning dosimetry quality assurance at Antoine Lacassagne center (Nice, France). Firstly, the interfaces between RadCalc and our treatment planning systems (TPS) Isogray (version 4.2) and Eclipse (version13.6) were checked for data transfer accuracy. Secondly, we created test plans in both Isogray and Eclipse featuring open fields, wedges fields, and irregular MLC fields. These test plans were transferred from TPSs according to the radiotherapy protocol of DICOM RT to RadCalc and the linac via Mosaiq (version 2.5). Measurements were performed in water phantom using a PTW cylindrical semiflex ionisation chamber (0.3 cm³, 31010) and compared with the TPSs and RadCalc calculation. Finally, 30 3D-CRT plans and 40 RapidArc plans created with patients CT scan were recalculated using the CT scan of a solid PMMA water equivalent phantom for 3D-CRT and the Octavius II phantom (PTW) CT scan for RapidArc. Next, we measure the doses delivered into these phantoms for each plan with a 0.3 cm³ PTW 31010 cylindrical semiflex ionisation chamber (3D-CRT) and 0.015 cm³ PTW PinPoint ionisation chamber (Rapidarc). For our test plans, good agreements were found between calculation (RadCalc and TPSs) and measurement (mean: 1.3%; standard deviation: ± 0.8%). Regarding the patient plans, the measured doses were compared to the calculation in RadCalc and in our TPSs. Moreover, RadCalc calculations were compared to Isogray and Eclispse ones. Agreements better than (2.8%; ± 1.2%) were found between RadCalc and TPSs. As for the comparison between calculation and measurement the agreement for all of our plans was better than (2.3%; ± 1.1%). The independent MU verification calculation software RadCal has been validated for clinical use and for both 3D-CRT and RapidArc techniques. The perspective of this project includes the validation of RadCal for the Tomotherapy machine installed at centre Antoine Lacassagne.

Keywords: 3D conformational radiotherapy, intensity modulated radiotherapy, monitor unit calculation, dosimetry quality assurance

Procedia PDF Downloads 214
6948 A Method for False Alarm Recognition Based on Multi-Classification Support Vector Machine

Authors: Weiwei Cui, Dejian Lin, Leigang Zhang, Yao Wang, Zheng Sun, Lianfeng Li

Abstract:

Built-in test (BIT) is an important technology in testability field, and it is widely used in state monitoring and fault diagnosis. With the improvement of modern equipment performance and complexity, the scope of BIT becomes larger, and it leads to the emergence of false alarm problem. The false alarm makes the health assessment unstable, and it reduces the effectiveness of BIT. The conventional false alarm suppression methods such as repeated test and majority voting cannot meet the requirement for a complicated system, and the intelligence algorithms such as artificial neural networks (ANN) are widely studied and used. However, false alarm has a very low frequency and small sample, yet a method based on ANN requires a large size of training sample. To recognize the false alarm, we propose a method based on multi-classification support vector machine (SVM) in this paper. Firstly, we divide the state of a system into three states: healthy, false-alarm, and faulty. Then we use multi-classification with '1 vs 1' policy to train and recognize the state of a system. Finally, an example of fault injection system is taken to verify the effectiveness of the proposed method by comparing ANN. The result shows that the method is reasonable and effective.

Keywords: false alarm, fault diagnosis, SVM, k-means, BIT

Procedia PDF Downloads 155
6947 Molecular Characterization of Chicken B Cell Marker (ChB6) in Native Chicken of Poonch Region from International Borders of India and Pakistan

Authors: Mandeep Singh Azad.Dibyendu Chakraborty, Vikas Vohra

Abstract:

Introduction: Poonch is one of the remotest districts of the Jammu and Kashmir (UT) and situated on international borders. This native poultry population in these areas is quite hardy and thrives well in adverse climatic conditions. Till date, no local breed from this area (Jammu Province) has been characterized thus present study was undertaken with the main objectives of molecular characterization of ChB6 gene in local native chicken of Poonch region located at international borders between India and Pakistan. The chicken B-cell marker (ChB6) gene has been proposed as a candidate gene in regulating B-cell development. Material and Method: RNA was isolated by Blood RNA Purification Kit (HiPura) and Trizol method from whole blood samples. Positive PCR products with size 1110 bp were selected for further purification, sequencing and analysis. The amplified PCR product was sequenced by Sangers dideoxy chain termination method. The obtained sequence of ChB6 gene of Poonchi chicken were compared by MEGAX software. BioEdit software was used to construct phylogenic tree, and Neighbor Joining method was used to infer evolutionary history. In order to compute evolutionary distance Maximum Composite Likelihood method was used. Results: The positively amplified samples of ChB6 genes were then subjected to Sanger sequencing with “Primer Walking. The sequences were then analyzed using MEGA X and BioEdit software. The sequence results were compared with other reported sequence from different breed of chicken and with other species obtained from the NCBI (National Center for Biotechnology Information). ClustalW method using MEGA X software was used for multiple sequence alignment. The sequence results of ChB6 gene of Poonchi chicken was compared with Centrocercus urophasianus, G. gallus mRNA for B6.1 protein, G. gallus mRNA for B6.2, G. gallus mRNA for B6.3, Gallus gallus B6.1, Halichoeres bivittatus, Miniopterus fuliginosus Ferringtonia patagonica, Tympanuchus phasianellus. The genetic distances were 0.2720, 0.0000, 0.0245, 0.0212, 0.0147, 1.6461, 2.2394, 2.0070 and 0.2363 for ChB6 gene of Poonchi chicken sequence with other sequences in the present study respectively. Sequencing results showed variations between different species. It was observed that AT content were higher then GC content for ChB6 gene. The lower AT content suggests less thermostable. It was observed that there was no sequence difference within the Poonchi population for ChB6 gene. The high homology within chicken population indicates the conservation of ChB6 gene. The maximum difference was observed with Miniopterus fuliginosus (Eastern bent-wing bat) followed by Ferringtonia patagonica and Halichoeres bivittatus. Conclusion: Genetic variation is the essential component for genetic improvement. The results of immune related gene Chb6 shows between population genetic variability. Therefore, further association studies of this gene with some prevalent diseases in large population would be helpful to identify disease resistant/ susceptible genotypes in the indigenous chicken population.

Keywords: ChB6, sequencing, ClustalW, genetic distance, poonchi chicken, SNP

Procedia PDF Downloads 67
6946 Quantification of Leachate Potential of the Quezon City Controlled Dumping Facility Using Help Model

Authors: Paul Kenneth D. Luzon, Maria Antonia N. Tanchuling

Abstract:

The Quezon City Controlled Dumping facility also known as Payatas produces leachate which can contaminate soil and water environment in the area. The goal of this study is to quantify the leachate produced by the QCCDF using the Hydrologic Evaluation of Landfill Performance (HELP) model. Results could be used as input for groundwater contaminant transport studies. The HELP model is based on a simple water budget and is an essential “model requirement” used by the US Environmental Protection Agency (EPA). Annual waste profile of the QCCDF was calculated. Based on topographical maps and estimation of settlement due to overburden pressure and degradation, a total of 10M m^3 of waste is contained in the landfill. The input necessary for the HELP model are weather data, soil properties, and landfill design. Results showed that from 1988 to 2011, an average of 50% of the total precipitation percolates through the bottom layer. Validation of the results is still needed due to the assumptions made in the study. The decrease in porosity of the top soil cover showed the best mitigation for minimizing percolation rate. This study concludes that there is a need for better leachate management system in the QCCDF.

Keywords: help model, landfill, payatas trash slide, quezon city controlled dumping facility

Procedia PDF Downloads 289
6945 Applying Different Stenography Techniques in Cloud Computing Technology to Improve Cloud Data Privacy and Security Issues

Authors: Muhammad Muhammad Suleiman

Abstract:

Cloud Computing is a versatile concept that refers to a service that allows users to outsource their data without having to worry about local storage issues. However, the most pressing issues to be addressed are maintaining a secure and reliable data repository rather than relying on untrustworthy service providers. In this study, we look at how stenography approaches and collaboration with Digital Watermarking can greatly improve the system's effectiveness and data security when used for Cloud Computing. The main requirement of such frameworks, where data is transferred or exchanged between servers and users, is safe data management in cloud environments. Steganography is the cloud is among the most effective methods for safe communication. Steganography is a method of writing coded messages in such a way that only the sender and recipient can safely interpret and display the information hidden in the communication channel. This study presents a new text steganography method for hiding a loaded hidden English text file in a cover English text file to ensure data protection in cloud computing. Data protection, data hiding capability, and time were all improved using the proposed technique.

Keywords: cloud computing, steganography, information hiding, cloud storage, security

Procedia PDF Downloads 190
6944 Operation Cycle Model of ASz62IR Radial Aircraft Engine

Authors: M. Duk, L. Grabowski, P. Magryta

Abstract:

Today's very important element relating to air transport is the environment impact issues. Nowadays there are no emissions standards for turbine and piston engines used in air transport. However, it should be noticed that the environmental effect in the form of exhaust gases from aircraft engines should be as small as possible. For this purpose, R&D centers often use special software to simulate and to estimate the negative effect of engine working process. For cooperation between the Lublin University of Technology and the Polish aviation company WSK "PZL-KALISZ" S.A., to achieve more effective operation of the ASz62IR engine, one of such tools have been used. The AVL Boost software allows to perform 1D simulations of combustion process of piston engines. ASz62IR is a nine-cylinder aircraft engine in a radial configuration. In order to analyze the impact of its working process on the environment, the mathematical model in the AVL Boost software have been made. This model contains, among others, model of the operation cycle of the cylinders. This model was based on a volume change in combustion chamber according to the reciprocating movement of a piston. The simplifications that all of the pistons move identically was assumed. The changes in cylinder volume during an operating cycle were specified. Those changes were important to determine the energy balance of a cylinder in an internal combustion engine which is fundamental for a model of the operating cycle. The calculations for cylinder thermodynamic state were based on the first law of thermodynamics. The change in the mass in the cylinder was calculated from the sum of inflowing and outflowing masses including: cylinder internal energy, heat from the fuel, heat losses, mass in cylinder, cylinder pressure and volume, blowdown enthalpy, evaporation heat etc. The model assumed that the amount of heat released in combustion process was calculated from the pace of combustion, using Vibe model. For gas exchange, it was also important to consider heat transfer in inlet and outlet channels because of much higher values there than for flow in a straight pipe. This results from high values of heat exchange coefficients and temperature coefficients near valves and valve seats. A Zapf modified model of heat exchange was used. To use the model with the flight scenarios, the impact of flight altitude on engine performance has been analyze. It was assumed that the pressure and temperature at the inlet and outlet correspond to the values resulting from the model for International Standard Atmosphere (ISA). Comparing this model of operation cycle with the others submodels of the ASz62IR engine, it could be noticed, that a full analysis of the performance of the engine, according to the ISA conditions, can be made. This work has been financed by the Polish National Centre for Research and Development, INNOLOT, under

Keywords: aviation propulsion, AVL Boost, engine model, operation cycle, aircraft engine

Procedia PDF Downloads 291
6943 Initializing E-Classroom in a Multigrade School in the Philippines

Authors: Karl Erickson I. Ebora

Abstract:

Science and technology are two inseparable terms which bring wonders to all aspects of life such as education, medicine, food production and even the environment. In education, technology has become an integral part as it brings many benefits to the teaching-learning process. However, in the Philippines, being one of the developing countries resources are scarce and not all schools enjoy the fruits brought by technology. Much of this ordeal impacts that of multigrade instruction. These schools are often the last priority in resources allocation since these have limited number of students. In fact, it is not surprising that these schools do not have even a single computer unit much more a computer laboratory. This paper sought to present a plan on how public schools would receive its e-classroom. Specifically, this paper sought to answer questions like the level of the school readiness in terms of facilities and equipment; the attitude of the respondents towards the use of e-classroom; level of teacher’s familiarity in using different e-classroom software and the plans of interventions undertaken by the school to make it e-classroom ready. After gathering and analysing the necessary data, this paper came up with the following conclusions that in terms of facilities and equipment, Guisguis Talon Elementary School (Main), though a multigrade school, is ready to receive e-classroom.; that the respondents show positive disposition in technology utilization in teaching after they strongly agree that technology plays essential role in the teaching-learning process. Also, they strongly agree that technology is a good motivator; it makes the teaching and learning more interesting and effective; it makes teaching easy; and that technology enhances student’s learning. Additionally, Teacher-respondents in Guisguis Talon Elementary School (Main) show familiarity in using software. They are very familiar with MS Word; MS Excel; MS PowerPoint; and internet and email. Moreover, they are very familiar with basic e-classroom computer operations and basic application software. They are very familiar with MS office and can do simple editing and formatting; in accessing and saving information from CD/DVD, external hard drives, USB and the like; and in browsing effectively different search engines and educational sites, download and upload files. Likewise respondents strongly agree to the interventions undertaken by the school to make it e-classroom ready. They strongly agree that funding and support are needed by the school; that stakeholders should be encouraged to consider donating of equipment; and that school and community should try to mobilize their resources in order to help the school; that the teachers should be provided with trainings in order for them to be technologically competent; and that principals and administrators should motivate their teachers to undergo continuous professional development.

Keywords: e-classroom, multi-grade school, DCP, classroom computers

Procedia PDF Downloads 199
6942 Vitamin A Status and Its Correlation with the Dietary Intake of Young Females of Lahore, Pakistan

Authors: Sarah Fatima, Ahmad A. Malik, Saima Sadaf

Abstract:

This study was conducted in order to assess the dietary record and vitamin A status of young females of Lahore. A total sample of 376 consisted of 16 – 20 years of unmarried college going females. Three main tools were adopted: questionnaire, 3-day food diary and serum retinol test. The anthropometric measurements showed that a total of 32.6% of the sample was underweight (BMI < 18.5) and 54.5% had a healthy weight (BMI 18.5 – 22.9). The average Vitamin A intake of the sample was 257.95 µg/day while the RDA for the selected age group was 700 µg/day. The mean energy intake of the adolescents was 1153.64 kcal/ day, whereas the Estimated Energy Requirement (EER) for this age group was 2368 kcal/day. The mean serum Vitamin A level was 24.81µg/dL. 69.6% of the sample was deficient in serum Vitamin A i.e. serum retinol < 24 µg/dL. 30.4% had serum retinol in normal limit (24 – 84 µg/dL) from which 25.3% lied in lower limit (24 – 44 µg/dL) and only 5.1% had serum retinol in 44 – 64 µg/dL range. A slightly negative correlation (r = - 0.21, 95% confidence interval) was found between dietary intake of Vitamin A and serum Vitamin A It was concluded that the dietary intake of major nutrients and vitamin A is not adequate in the selected group. This is also confirmed by the lower serum retinol levels. Hence, vitamin An intake and status are generally inadequate, and vitamin deficiency is prevalent in the unmarried young females of Lahore.

Keywords: vitamin A, young Females, vitamin deficiency, Lahore

Procedia PDF Downloads 313
6941 Study υ_4 Fundamental Band of 12 CD4 Molecule

Authors: Kaarour Abdelkrim, Ouardi Okkacha, Meskine Mohamed

Abstract:

In this study, the υ_4 fundamental band of 12CD4 molecule has been studied by infrared spectroscopy with high resolution. Using XTDS and SPEVIEW software and the tensor formalism developed by ICB (laboratoire interdisciplinaire de Bourgogne) to several lines have been assigned and fitted with a standard deviation acceptable. This analysis allowed us to calculate several parameters of the molecule 12 CD4.

Keywords: XTDS, SPEVIEW, tetrahedral tensorial formalism, rovibrational band

Procedia PDF Downloads 325
6940 Reclaiming Properties of Bituminous Concrete Using Cold Mix Design Technology

Authors: Pradeep Kumar, Shalinee Shukla

Abstract:

Pavement plays a vital role in the socio-economic development of a country. Bituminous roads construction with conventional paving grade bitumen obtained from hot mix plant creates pollution and involves emission of greenhouse gases, also the construction of pavements at very high temperature is not feasible or desirable for high rainfall and snowfall areas. This problem of overheating can be eliminated by the construction of pavements with the usage of emulsified cold mixes which will eliminate emissions and help in the reduction of fuel requirement at mixing plant, which leads to energy conservation. Cold mix is a mixture of unheated aggregate and emulsion or cutback and filler. The primary objective of this research is to assess the volumetric mix design parameters of recycled aggregates with cold mixing technology and also to assess the impact of additives on volumetric mix characteristics. In this present study, bituminous pavement materials are reclaimed using cold mix technology, and Marshall specimens are prepared with the help of slow setting type 2 (SS-2) cationic bitumen emulsion as a binder for recycled aggregates. This technique of road construction is more environmentally friendly and can be done in adverse weather conditions.

Keywords: cold mixes, bitumen emulsion, recycled aggregates, volumetric properties

Procedia PDF Downloads 135
6939 Recent Advances in Data Warehouse

Authors: Fahad Hanash Alzahrani

Abstract:

This paper describes some recent advances in a quickly developing area of data storing and processing based on Data Warehouses and Data Mining techniques, which are associated with software, hardware, data mining algorithms and visualisation techniques having common features for any specific problems and tasks of their implementation.

Keywords: data warehouse, data mining, knowledge discovery in databases, on-line analytical processing

Procedia PDF Downloads 402
6938 Evaluating the ‘Assembled Educator’ of a Specialized Postgraduate Engineering Course Using Activity Theory and Genre Ecologies

Authors: Simon Winberg

Abstract:

The landscape of professional postgraduate education is changing: the focus of these programmes is moving from preparing candidates for a life in academia towards a focus of training in expert knowledge and skills to support industry. This is especially pronounced in engineering disciplines where increasingly more complex products are drawing on a depth of knowledge from multiple fields. This connects strongly with the broader notion of Industry 4.0 – where technology and society are being brought together to achieve more powerful and desirable products, but products whose inner workings also are more complex than before. The changes in what we do, and how we do it, has a profound impact on what industry would like universities to provide. One such change is the increased demand for taught doctoral and Masters programmes. These programmes aim to provide skills and training for professionals, to expand their knowledge of state-of-the-art tools and technologies. This paper investigates one such course, namely a Software Defined Radio (SDR) Master’s degree course. The teaching support for this course had to be drawn from an existing pool of academics, none of who were specialists in this field. The paper focuses on the kind of educator, a ‘hybrid academic’, assembled from available academic staff and bolstered by research. The conceptual framework for this paper combines Activity Theory and Genre Ecology. Activity Theory is used to reason about learning and interactions during the course, and Genre Ecology is used to model building and sharing of technical knowledge related to using tools and artifacts. Data were obtained from meetings with students and lecturers, logs, project reports, and course evaluations. The findings show how the course, which was initially academically-oriented, metamorphosed into a tool-dominant peer-learning structure, largely supported by the sharing of technical tool-based knowledge. While the academic staff could address gaps in the participants’ fundamental knowledge of radio systems, the participants brought with them extensive specialized knowledge and tool experience which they shared with the class. This created a complicated dynamic in the class, which centered largely on engagements with technology artifacts, such as simulators, from which knowledge was built. The course was characterized by a richness of ‘epistemic objects’, which is to say objects that had knowledge-generating qualities. A significant portion of the course curriculum had to be adapted, and the learning methods changed to accommodate the dynamic interactions that occurred during classes. This paper explains the SDR Masters course in terms of conflicts and innovations in its activity system, as well as the continually hybridizing genre ecology to show how the structuring and resource-dependence of the course transformed from its initial ‘traditional’ academic structure to a more entangled arrangement over time. It is hoped that insights from this paper would benefit other educators involved in the design and teaching of similar types of specialized professional postgraduate taught programmes.

Keywords: professional postgraduate education, taught masters, engineering education, software defined radio

Procedia PDF Downloads 92
6937 The Development of XML Resume System in Thailand

Authors: Jarumon Nookhong, Thanakorn Uiphanit

Abstract:

This study is a research and development project which aims to develop XML Resume System to be the standard system in Thailand as well as to measure the efficiency of the XML Resume System in Thailand. This research separates into 2 stages: 1) to develop XML Document System to be the standard in Thailand, and 2) to experiment the system performance. The sample in this research is committed by 50 specialists in the field of human resources by selecting specifically. The tool that uses in this research is XML Resume System in Thailand and the performance evaluation format of system while the analysis of the data is calculated by using average and standard deviation. The result of the research found that the development of the XML Resume System that aims to be the standard system in Thailand had the result 2.67 of the average which is in a good level. The evaluation in testing the performance of the system had been done by the specialists of human resources who use the XML Resume system. When analyzing each part, it found out that the abilities according to the user’s requirement from specialists in the field of human resources, the convenience and easiness in usages, and the functional competency are respectively in a good level. The average of the ability according to the user’s need from specialists of human resources is 2.92. The average of the convenience and easiness in usages is 2.56. The average of functional competency is 2.53. These can be used as the standard effectively.

Keywords: resume, XML, XML schema, computer science

Procedia PDF Downloads 408
6936 Diffusion MRI: Clinical Application in Radiotherapy Planning of Intracranial Pathology

Authors: Pomozova Kseniia, Gorlachev Gennadiy, Chernyaev Aleksandr, Golanov Andrey

Abstract:

In clinical practice, and especially in stereotactic radiosurgery planning, the significance of diffusion-weighted imaging (DWI) is growing. This makes the existence of software capable of quickly processing and reliably visualizing diffusion data, as well as equipped with tools for their analysis in terms of different tasks. We are developing the «MRDiffusionImaging» software on the standard C++ language. The subject part has been moved to separate class libraries and can be used on various platforms. The user interface is Windows WPF (Windows Presentation Foundation), which is a technology for managing Windows applications with access to all components of the .NET 5 or .NET Framework platform ecosystem. One of the important features is the use of a declarative markup language, XAML (eXtensible Application Markup Language), with which you can conveniently create, initialize and set properties of objects with hierarchical relationships. Graphics are generated using the DirectX environment. The MRDiffusionImaging software package has been implemented for processing diffusion magnetic resonance imaging (dMRI), which allows loading and viewing images sorted by series. An algorithm for "masking" dMRI series based on T2-weighted images was developed using a deformable surface model to exclude tissues that are not related to the area of interest from the analysis. An algorithm of distortion correction using deformable image registration based on autocorrelation of local structure has been developed. Maximum voxel dimension was 1,03 ± 0,12 mm. In an elementary brain's volume, the diffusion tensor is geometrically interpreted using an ellipsoid, which is an isosurface of the probability density of a molecule's diffusion. For the first time, non-parametric intensity distributions, neighborhood correlations, and inhomogeneities are combined in one segmentation of white matter (WM), grey matter (GM), and cerebrospinal fluid (CSF) algorithm. A tool for calculating the coefficient of average diffusion and fractional anisotropy has been created, on the basis of which it is possible to build quantitative maps for solving various clinical problems. Functionality has been created that allows clustering and segmenting images to individualize the clinical volume of radiation treatment and further assess the response (Median Dice Score = 0.963 ± 0,137). White matter tracts of the brain were visualized using two algorithms: deterministic (fiber assignment by continuous tracking) and probabilistic using the Hough transform. The proposed algorithms test candidate curves in the voxel, assigning to each one a score computed from the diffusion data, and then selects the curves with the highest scores as the potential anatomical connections. White matter fibers were visualized using a Hough transform tractography algorithm. In the context of functional radiosurgery, it is possible to reduce the irradiation volume of the internal capsule receiving 12 Gy from 0,402 cc to 0,254 cc. The «MRDiffusionImaging» will improve the efficiency and accuracy of diagnostics and stereotactic radiotherapy of intracranial pathology. We develop software with integrated, intuitive support for processing, analysis, and inclusion in the process of radiotherapy planning and evaluating its results.

Keywords: diffusion-weighted imaging, medical imaging, stereotactic radiosurgery, tractography

Procedia PDF Downloads 84
6935 The Effects of Sewage Sludge Usage and Manure on Some Heavy Metals Uptake in Savory (Satureja Hortensis L.)

Authors: Abbas Hani

Abstract:

In recent decades with the development of technology and lack of food sources, sewage sludge in production of human foods is inevitable. Various sources of municipal and industrial sewage sludge that is produced can provide the requirement of plant nutrients. Soils in arid, semi-arid climate of central Iran that most affected by water drainage, iron and zinc deficiencies, using of sewage sludge is helpful. Therefore, the aim of this study is investigation of sewage sludge and manure application on Ni and Zn uptake by Savory. An experiment in a randomized complete block design with three replications was performed. Sewage sludge treatments consisted of four levels, control, 15, 30, 80 tons per hectares, the manure was used in four levels of control, 20, 40 and 80 tons per hectare. Results showed that the wet and dry weights was not affected by sewage sludge using, while, manure has significant effect on them. The effect of sewage sludge on the cadmium and lead concentrations were significant. Interactions of sewage sludge and manure on dry weight values were not significant. Compare mean analysis showed that increasing the amount of sewage sludge had no significant effect on cadmium concentration and it reduced when sewage sludge usage increased. This is probably due to increased plant growth and reduced concentrations of these elements in the plant.

Keywords: savory, lead, cadmium, sewage sludge, manure

Procedia PDF Downloads 420
6934 Efficient Study of Substrate Integrated Waveguide Devices

Authors: J. Hajri, H. Hrizi, N. Sboui, H. Baudrand

Abstract:

This paper presents a study of SIW circuits (Substrate Integrated Waveguide) with a rigorous and fast original approach based on Iterative process (WCIP). The theoretical suggested study is validated by the simulation of two different examples of SIW circuits. The obtained results are in good agreement with those of measurement and with software HFSS.

Keywords: convergence study, HFSS, modal decomposition, SIW circuits, WCIP method

Procedia PDF Downloads 497
6933 Integration of EEG and Motion Tracking Sensors for Objective Measure of Attention-Deficit Hyperactivity Disorder in Pre-Schoolers

Authors: Neha Bhattacharyya, Soumendra Singh, Amrita Banerjee, Ria Ghosh, Oindrila Sinha, Nairit Das, Rajkumar Gayen, Somya Subhra Pal, Sahely Ganguly, Tanmoy Dasgupta, Tanusree Dasgupta, Pulak Mondal, Aniruddha Adhikari, Sharmila Sarkar, Debasish Bhattacharyya, Asim Kumar Mallick, Om Prakash Singh, Samir Kumar Pal

Abstract:

Background: We aim to develop an integrated device comprised of single-probe EEG and CCD-based motion sensors for a more objective measure of Attention-deficit Hyperactivity Disorder (ADHD). While the integrated device (MAHD) relies on the EEG signal (spectral density of beta wave) for the assessment of attention during a given structured task (painting three segments of a circle using three different colors, namely red, green and blue), the CCD sensor depicts movement pattern of the subjects engaged in a continuous performance task (CPT). A statistical analysis of the attention and movement patterns was performed, and the accuracy of the completed tasks was analysed using indigenously developed software. The device with the embedded software, called MAHD, is intended to improve certainty with criterion E (i.e. whether symptoms are better explained by another condition). Methods: We have used the EEG signal from a single-channel dry sensor placed on the frontal lobe of the head of the subjects (3-5 years old pre-schoolers). During the painting of three segments of a circle using three distinct colors (red, green, and blue), absolute power for delta and beta EEG waves from the subjects are found to be correlated with relaxation and attention/cognitive load conditions. While the relaxation condition of the subject hints at hyperactivity, a more direct CCD-based motion sensor is used to track the physical movement of the subject engaged in a continuous performance task (CPT) i.e., separation of the various colored balls from one table to another. We have used our indigenously developed software for the statistical analysis to derive a scale for the objective assessment of ADHD. We have also compared our scale with clinical ADHD evaluation. Results: In a limited clinical trial with preliminary statistical analysis, we have found a significant correlation between the objective assessment of the ADHD subjects with that of the clinician’s conventional evaluation. Conclusion: MAHD, the integrated device, is supposed to be an auxiliary tool to improve the accuracy of ADHD diagnosis by supporting greater criterion E certainty.

Keywords: ADHD, CPT, EEG signal, motion sensor, psychometric test

Procedia PDF Downloads 97
6932 Left to Right-Right Most Parsing Algorithm with Lookahead

Authors: Jamil Ahmed

Abstract:

Left to Right-Right Most (LR) parsing algorithm is a widely used algorithm of syntax analysis. It is contingent on a parsing table, whereas the parsing tables are extracted from the grammar. The parsing table specifies the actions to be taken during parsing. It requires that the parsing table should have no action conflicts for the same input symbol. This requirement imposes a condition on the class of grammars over which the LR algorithms work. However, there are grammars for which the parsing tables hold action conflicts. In such cases, the algorithm needs a capability of scanning (looking-ahead) next input symbols ahead of the current input symbol. In this paper, a ‘Left to Right’-‘Right Most’ parsing algorithm with lookahead capability is introduced. The 'look-ahead' capability in the LR parsing algorithm is the major contribution of this paper. The practicality of the proposed algorithm is substantiated by the parser implementation of the Context Free Grammar (CFG) of an already proposed programming language 'State Controlled Object Oriented Programming' (SCOOP). SCOOP’s Context Free Grammar has 125 productions and 192 item sets. This algorithm parses SCOOP while the grammar requires to ‘look ahead’ the input symbols due to action conflicts in its parsing table. Proposed LR parsing algorithm with lookahead capability can be viewed as an optimization of ‘Simple Left to Right’-‘Right Most’ (SLR) parsing algorithm.

Keywords: left to right-right most parsing, syntax analysis, bottom-up parsing algorithm

Procedia PDF Downloads 123
6931 The Old Traditional Structures in Iran: A Suitable Model for Today's Greenhouse

Authors: Behbood Maashkar

Abstract:

One of the principles for societies’ development is the requirement to consider past experiences. Man should always take advantage of the predecessor’s experiences and analyze their works and methods. The predecessors have had a more friendly relationship with nature and their lives less damaged the nature, and it is one of the elements of green building. One of the things the ancients have observed in regard to green building in their houses, stores, sacred places, etc. was using wind-catchers as an air conditioning and cooling system which can be considered as the first foundations of green building. In designing houses Iranian architects have paid a great attention to the factor of making use of more shaded area in hot season and insulation of wall and ceiling against influence of hot weather and also air circulation inside the building. In order to circulate the air inside closed spaces and decrease the temperature, they have considered different winds which blow in Iran and its effective power, and in order to make use of it they invented wind catcher. Direction of wind blow and its height from the earth as well as the time and duration of wind blow and other factors have been effective in making different types of wind catchers. Using wind catchers has been and is prevalent mainly in central and south regions of Iran, coastal areas of Persian Gulf, and Khorasan, especially in cities like Yazd, Kashan, Bam, Abarghoo, Jahrom, and Tabas.

Keywords: environment pollution, green building, Iran, wind catchers

Procedia PDF Downloads 248
6930 Real Time Monitoring and Control of Proton Exchange Membrane Fuel Cell in Cognitive Radio Environment

Authors: Prakash Thapa, Gye Choon Park, Sung Gi Kwon, Jin Lee

Abstract:

The generation of electric power from a proton exchange membrane (PEM) fuel cell is influenced by temperature, pressure, humidity, flow rate of reactant gaseous and partial flooding of membrane electrode assembly (MEA). Among these factors, temperature and cathode flooding are the most affecting parameters on the performance of fuel cell. This paper describes the detail design and effect of these parameters on PEM fuel cell. Performance of all parameters was monitored, analyzed and controlled by using 5KWatt PEM fuel cell. In the real-time data communication for remote monitoring and control of PEM fuel cell, a normalized least mean square algorithm in cognitive radio environment is used. By the use of this method, probability of energy signal detection will be maximum which solved the frequency shortage problem. So the monitoring system hanging out and slow speed problem will be solved. Also from the control unit, all parameters are controlled as per the system requirement. As a result, PEM fuel cell generates maximum electricity with better performance.

Keywords: proton exchange membrane (PEM) fuel cell, pressure, temperature and humidity sensor (PTH), efficiency curve, cognitive radio network (CRN)

Procedia PDF Downloads 457
6929 Using Printouts as Social Media Evidence and Its Authentication in the Courtroom

Authors: Chih-Ping Chang

Abstract:

Different from traditional objective evidence, social media evidence has its own characteristics with easily tampering, recoverability, and cannot be read without using other devices (such as a computer). Simply taking a screenshot from social network sites must be questioned its original identity. When the police search and seizure digital information, a common way they use is to directly print out digital data obtained and ask the signature of the parties at the presence, without taking original digital data back. In addition to the issue on its original identity, this conduct to obtain evidence may have another two results. First, it will easily allege that is tampering evidence because the police wanted to frame the suspect and falsified evidence. Second, it is not easy to discovery hidden information. The core evidence associated with crime may not appear in the contents of files. Through discovery the original file, data related to the file, such as the original producer, creation time, modification date, and even GPS location display can be revealed from hidden information. Therefore, how to show this kind of evidence in the courtroom will be arguably the most important task for ruling social media evidence. This article, first, will introduce forensic software, like EnCase, TCT, FTK, and analyze their function to prove the identity with another digital data. Then turning back to the court, the second part of this article will discuss legal standard for authentication of social media evidence and application of that forensic software in the courtroom. As the conclusion, this article will provide a rethinking, that is, what kind of authenticity is this rule of evidence chase for. Does legal system automatically operate the transcription of scientific knowledge? Or furthermore, it wants to better render justice, not only under scientific fact, but through multivariate debating.

Keywords: federal rule of evidence, internet forensic, printouts as evidence, social media evidence, United States v. Vayner

Procedia PDF Downloads 289
6928 Optimization of Element Type for FE Model and Verification of Analyses with Physical Tests

Authors: Mustafa Tufekci, Caner Guven

Abstract:

In Automotive Industry, sliding door systems that are also used as body closures, are safety members. Extreme product tests are realized to prevent failures in a design process, but these tests realized experimentally result in high costs. Finite element analysis is an effective tool used for the design process. These analyses are used before production of a prototype for validation of design according to customer requirement. In result of this, the substantial amount of time and cost is saved. Finite element model is created for geometries that are designed in 3D CAD programs. Different element types as bar, shell and solid, can be used for creating mesh model. The cheaper model can be created by the selection of element type, but combination of element type that was used in model, number and geometry of element and degrees of freedom affects the analysis result. Sliding door system is a good example which used these methods for this study. Structural analysis was realized for sliding door mechanism by using FE models. As well, physical tests that have same boundary conditions with FE models were realized. Comparison study for these element types, were done regarding test and analyses results then the optimum combination was achieved.

Keywords: finite element analysis, sliding door mechanism, element type, structural analysis

Procedia PDF Downloads 327
6927 A New Computational Package for Using in CFD and Other Problems (Third Edition)

Authors: Mohammad Reza Akhavan Khaleghi

Abstract:

This paper shows changes done to the Reduced Finite Element Method (RFEM) that its result will be the most powerful numerical method that has been proposed so far (some forms of this method are so powerful that they can approximate the most complex equations simply Laplace equation!). Finite Element Method (FEM) is a powerful numerical method that has been used successfully for the solution of the existing problems in various scientific and engineering fields such as its application in CFD. Many algorithms have been expressed based on FEM, but none have been used in popular CFD software. In this section, full monopoly is according to Finite Volume Method (FVM) due to better efficiency and adaptability with the physics of problems in comparison with FEM. It doesn't seem that FEM could compete with FVM unless it was fundamentally changed. This paper shows those changes and its result will be a powerful method that has much better performance in all subjects in comparison with FVM and another computational method. This method is not to compete with the finite volume method but to replace it.

Keywords: reduced finite element method, new computational package, new finite element formulation, new higher-order form, new isogeometric analysis

Procedia PDF Downloads 116