Search results for: information centric network
11870 Digital Literacy Transformation and Implications in Institutions of Higher Learning in Kenya
Authors: Emily Cherono Sawe, Elisha Ondieki Makori
Abstract:
Knowledge and digital economies have brought challenges and potential opportunities for universities to innovate and improve the quality of learning. Disruption technologies and information dynamics continue to transform and change the landscape in teaching, scholarship, and research activities across universities. Digital literacy is a fundamental and imperative element in higher education and training, as witnessed during the new norm. COVID-19 caused unprecedented disruption in universities, where teaching and learning depended on digital innovations and applications. Academic services and activities were provided online, including library information services. Information professionals were forced to adopt various digital platforms in order to provide information services to patrons. University libraries’ roles in fulfilling educational responsibilities continue to evolve in response to changes in pedagogy, technology, economy, society, policies, and strategies of parent institutions. Libraries are currently undergoing considerable transformational change as a result of the inclusion of a digital environment. Academic libraries have been at the forefront of providing online learning resources and online information services, as well as supporting students and staff to develop digital literacy skills via online courses, tutorials, and workshops. Digital literacy transformation and information staff are crucial elements reminiscent of the prioritization of skills and knowledge for lifelong learning. The purpose of this baseline research is to assess the implications of digital literacy transformation in institutions of higher learning in Kenya and share appropriate strategies to leverage and sustain teaching and research. Objectives include examining the leverage and preparedness of the digital literacy environment in streamlining learning in the universities, exploring and benchmarking imperative digital competence for information professionals, establishing the perception of information professionals towards digital literacy skills, and determining lessons, best practices, and strategies to accelerate digital literacy transformation for effective research and learning in the universities. The study will adopt a descriptive research design using questionnaires and document analysis as the instruments for data collection. The targeted population is librarians and information professionals, as well as academics in public and private universities teaching information literacy programmes. Data and information are to be collected through an online structured questionnaire and digital face-to-face interviews. Findings and results will provide promising lessons together with best practices and strategies to transform and change digital literacies in university libraries in Kenya.Keywords: digital literacy, digital innovations, information professionals, librarians, higher education, university libraries, digital information literacy
Procedia PDF Downloads 9611869 A Schema of Building an Efficient Quality Gate throughout the Software Development with Tools
Authors: Le Chen
Abstract:
This paper presents an efficient tool platform scheme to ensure quality protection throughout the software development process. The main principle is to manage the information of requirements, design, development, testing, operation and maintenance process with proper tools, and to set up the quality standards of each process. Through the tools’ display and summary of quality standards, the quality standards can be visualizad and ready for policy decision, which is called Quality Gate in this paper. In addition, the tools are also integrated to achieve the exchange and relation of information which highly improving operational efficiency. In this paper, the feasibility of the scheme is verified by practical application of development projects, and the overall information display and data mining are proposed to be further improved.Keywords: efficiency, quality gate, software process, tools
Procedia PDF Downloads 35911868 INCIPIT-CRIS: A Research Information System Combining Linked Data Ontologies and Persistent Identifiers
Authors: David Nogueiras Blanco, Amir Alwash, Arnaud Gaudinat, René Schneider
Abstract:
At a time when the access to and the sharing of information are crucial in the world of research, the use of technologies such as persistent identifiers (PIDs), Current Research Information Systems (CRIS), and ontologies may create platforms for information sharing if they respond to the need of disambiguation of their data by assuring interoperability inside and between other systems. INCIPIT-CRIS is a continuation of the former INCIPIT project, whose goal was to set up an infrastructure for a low-cost attribution of PIDs with high granularity based on Archival Resource Keys (ARKs). INCIPIT-CRIS can be interpreted as a logical consequence and propose a research information management system developed from scratch. The system has been created on and around the Schema.org ontology with a further articulation of the use of ARKs. It is thus built upon the infrastructure previously implemented (i.e., INCIPIT) in order to enhance the persistence of URIs. As a consequence, INCIPIT-CRIS aims to be the hinge between previously separated aspects such as CRIS, ontologies and PIDs in order to produce a powerful system allowing the resolution of disambiguation problems using a combination of an ontology such as Schema.org and unique persistent identifiers such as ARK, allowing the sharing of information through a dedicated platform, but also the interoperability of the system by representing the entirety of the data as RDF triplets. This paper aims to present the implemented solution as well as its simulation in real life. We will describe the underlying ideas and inspirations while going through the logic and the different functionalities implemented and their links with ARKs and Schema.org. Finally, we will discuss the tests performed with our project partner, the Swiss Institute of Bioinformatics (SIB), by the use of large and real-world data sets.Keywords: current research information systems, linked data, ontologies, persistent identifier, schema.org, semantic web
Procedia PDF Downloads 13511867 Vision-Based Collision Avoidance for Unmanned Aerial Vehicles by Recurrent Neural Networks
Authors: Yao-Hong Tsai
Abstract:
Due to the sensor technology, video surveillance has become the main way for security control in every big city in the world. Surveillance is usually used by governments for intelligence gathering, the prevention of crime, the protection of a process, person, group or object, or the investigation of crime. Many surveillance systems based on computer vision technology have been developed in recent years. Moving target tracking is the most common task for Unmanned Aerial Vehicle (UAV) to find and track objects of interest in mobile aerial surveillance for civilian applications. The paper is focused on vision-based collision avoidance for UAVs by recurrent neural networks. First, images from cameras on UAV were fused based on deep convolutional neural network. Then, a recurrent neural network was constructed to obtain high-level image features for object tracking and extracting low-level image features for noise reducing. The system distributed the calculation of the whole system to local and cloud platform to efficiently perform object detection, tracking and collision avoidance based on multiple UAVs. The experiments on several challenging datasets showed that the proposed algorithm outperforms the state-of-the-art methods.Keywords: unmanned aerial vehicle, object tracking, deep learning, collision avoidance
Procedia PDF Downloads 16011866 Reactive Transport Modeling in Carbonate Rocks: A Single Pore Model
Authors: Priyanka Agrawal, Janou Koskamp, Amir Raoof, Mariette Wolthers
Abstract:
Calcite is the main mineral found in carbonate rocks, which form significant hydrocarbon reservoirs and subsurface repositories for CO2 sequestration. The injected CO2 mixes with the reservoir fluid and disturbs the geochemical equilibrium, triggering calcite dissolution. Different combinations of fluid chemistry and injection rate may therefore result in different evolution of porosity, permeability and dissolution patterns. To model the changes in porosity and permeability Kozeny-Carman equation K∝〖(∅)〗^n is used, where K is permeability and ∅ is porosity. The value of n is mostly based on experimental data or pore network models. In pore network models, this derivation is based on accuracy of relation used for conductivity and pore volume change. In fact, at a single pore scale, this relationship is the result of the pore shape development due to dissolution. We have prepared a new reactive transport model for a single pore which simulates the complex chemical reaction of carbonic-acid induced calcite dissolution and subsequent pore-geometry evolution at a single pore scale. We use COMSOL Multiphysics package 5.3 for the simulation. COMSOL utilizes the arbitary-Lagrangian Eulerian (ALE) method for the free-moving domain boundary. We examined the effect of flow rate on the evolution of single pore shape profiles due to calcite dissolution. We used three flow rates to cover diffusion dominated and advection-dominated transport regimes. The fluid in diffusion dominated flow (Pe number 0.037 and 0.37) becomes less reactive along the pore length and thus produced non-uniform pore shapes. However, for the advection-dominated flow (Pe number 3.75), the fast velocity of the fluid keeps the fluid relatively more reactive towards the end of the pore length, thus yielding uniform pore shape. Different pore shapes in terms of inlet opening vs overall pore opening will have an impact on the relation between changing volumes and conductivity. We have related the shape of pore with the Pe number which controls the transport regimes. For every Pe number, we have derived the relation between conductivity and porosity. These relations will be used in the pore network model to get the porosity and permeability variation.Keywords: single pore, reactive transport, calcite system, moving boundary
Procedia PDF Downloads 37411865 Building Information Modeling-Based Information Exchange to Support Facilities Management Systems
Authors: Sandra T. Matarneh, Mark Danso-Amoako, Salam Al-Bizri, Mark Gaterell
Abstract:
Today’s facilities are ever more sophisticated and the need for available and reliable information for operation and maintenance activities is vital. The key challenge for facilities managers is to have real-time accurate and complete information to perform their day-to-day activities and to provide their senior management with accurate information for decision-making process. Currently, there are various technology platforms, data repositories, or database systems such as Computer-Aided Facility Management (CAFM) that are used for these purposes in different facilities. In most current practices, the data is extracted from paper construction documents and is re-entered manually in one of these computerized information systems. Construction Operations Building information exchange (COBie), is a non-proprietary data format that contains the asset non-geometric data which was captured and collected during the design and construction phases for owners and facility managers use. Recently software vendors developed add-in applications to generate COBie spreadsheet automatically. However, most of these add-in applications are capable of generating a limited amount of COBie data, in which considerable time is still required to enter the remaining data manually to complete the COBie spreadsheet. Some of the data which cannot be generated by these COBie add-ins is essential for facilities manager’s day-to-day activities such as job sheet which includes preventive maintenance schedules. To facilitate a seamless data transfer between BIM models and facilities management systems, we developed a framework that enables automated data generation using the data extracted directly from BIM models to external web database, and then enabling different stakeholders to access to the external web database to enter the required asset data directly to generate a rich COBie spreadsheet that contains most of the required asset data for efficient facilities management operations. The proposed framework is a part of ongoing research and will be demonstrated and validated on a typical university building. Moreover, the proposed framework supplements the existing body of knowledge in facilities management domain by providing a novel framework that facilitates seamless data transfer between BIM models and facilities management systems.Keywords: building information modeling, BIM, facilities management systems, interoperability, information management
Procedia PDF Downloads 11611864 Designing Mobile Application to Motivate Young People to Visit Cultural Heritage Sites
Authors: Yuko Hiramatsu, Fumihiro Sato, Atsushi Ito, Hiroyuki Hatano, Mie Sato, Yu Watanabe, Akira Sasaki
Abstract:
This paper presents a mobile phone application developed for sightseeing in Nikko, one of the cultural world heritages in Japan, using the BLE (Bluetooth Low Energy) beacon. Based on our pre-research, we decided to design our application for young people who walk around the area actively, but know little about the tradition and culture of Nikko. One solution is to construct many information boards to explain; however, it is difficult to construct new guide plates in cultural world heritage sites. The smartphone is a good solution to send such information to such visitors. This application was designed using a combination of the smartphone and beacons, set in the area, so that when a tourist passes near a beacon, the application displays information about the area including a map, historical or cultural information about the temples and shrines, and local shops nearby as well as a bus timetable. It is useful for foreigners, too. In addition, we developed quizzes relating to the culture and tradition of Nikko to provide information based on the Zeigarnik effect, a psychological effect. According to the results of our trials, tourists positively evaluated the basic information and young people who used the quiz function were able to learn the historical and cultural points. This application helped young visitors at Nikko to understand the cultural elements of the site. In addition, this application has a function to send notifications. This function is designed to provide information about the local community such as shops, local transportation companies and information office. The application hopes to also encourage people living in the area, and such cooperation from the local people will make this application vivid and inspire young visitors to feel that the cultural heritage site is still alive today. This is a gateway for young people to learn about a traditional place and understand the gravity of preserving such areas.Keywords: BLE beacon, smartphone application, Zeigarnik effect, world heritage site, school trip
Procedia PDF Downloads 32411863 Markov Characteristics of the Power Line Communication Channels in China
Authors: Ming-Yue Zhai
Abstract:
Due to the multipath and pulse noise nature, power line communications(PLC) channel can be modelled as a memory one with the finite states Markov model(FSMC). As the most important parameter modelling a Markov channel,the memory order in an FSMC is not solved in PLC systems yet. In the paper, the mutual information is used as a measure of the dependence between the different symbols, treated as the received SNA or amplitude of the current channel symbol or that of previous symbols. The joint distribution probabilities of the envelopes in PLC systems are computed based on the multi-path channel model, which is commonly used in PLC. we confirm that given the information of the symbol immediately preceding the current one, any other previous symbol is independent of the current one in PLC systems, which means the PLC channels is a Markov chain with the first-order. The field test is also performed to model the received OFDM signals with the help of AR model. The results show that the first-order AR model is enough to model the fading channel in PLC systems, which means the amount of uncertainty remaining in the current symbol should be negligible, given the information corresponding to the immediately preceding one.Keywords: power line communication, channel model, markovian, information theory, first-order
Procedia PDF Downloads 41211862 Design an Algorithm for Software Development in CBSE Envrionment Using Feed Forward Neural Network
Authors: Amit Verma, Pardeep Kaur
Abstract:
In software development organizations, Component based Software engineering (CBSE) is emerging paradigm for software development and gained wide acceptance as it often results in increase quality of software product within development time and budget. In component reusability, main challenges are the right component identification from large repositories at right time. The major objective of this work is to provide efficient algorithm for storage and effective retrieval of components using neural network and parameters based on user choice through clustering. This research paper aims to propose an algorithm that provides error free and automatic process (for retrieval of the components) while reuse of the component. In this algorithm, keywords (or components) are extracted from software document, after by applying k mean clustering algorithm. Then weights assigned to those keywords based on their frequency and after assigning weights, ANN predicts whether correct weight is assigned to keywords (or components) or not, otherwise it back propagates in to initial step (re-assign the weights). In last, store those all keywords into repositories for effective retrieval. Proposed algorithm is very effective in the error correction and detection with user base choice while choice of component for reusability for efficient retrieval is there.Keywords: component based development, clustering, back propagation algorithm, keyword based retrieval
Procedia PDF Downloads 37811861 An Automatic Speech Recognition of Conversational Telephone Speech in Malay Language
Authors: M. Draman, S. Z. Muhamad Yassin, M. S. Alias, Z. Lambak, M. I. Zulkifli, S. N. Padhi, K. N. Baharim, F. Maskuriy, A. I. A. Rahim
Abstract:
The performance of Malay automatic speech recognition (ASR) system for the call centre environment is presented. The system utilizes Kaldi toolkit as the platform to the entire library and algorithm used in performing the ASR task. The acoustic model implemented in this system uses a deep neural network (DNN) method to model the acoustic signal and the standard (n-gram) model for language modelling. With 80 hours of training data from the call centre recordings, the ASR system can achieve 72% of accuracy that corresponds to 28% of word error rate (WER). The testing was done using 20 hours of audio data. Despite the implementation of DNN, the system shows a low accuracy owing to the varieties of noises, accent and dialect that typically occurs in Malaysian call centre environment. This significant variation of speakers is reflected by the large standard deviation of the average word error rate (WERav) (i.e., ~ 10%). It is observed that the lowest WER (13.8%) was obtained from recording sample with a standard Malay dialect (central Malaysia) of native speaker as compared to 49% of the sample with the highest WER that contains conversation of the speaker that uses non-standard Malay dialect.Keywords: conversational speech recognition, deep neural network, Malay language, speech recognition
Procedia PDF Downloads 32211860 Energy-Efficient Contact Selection Method for CARD in Wireless Ad-Hoc Networks
Authors: Mehdi Assefi, Keihan Hataminezhad
Abstract:
One of the efficient architectures for exploring the resources in wireless ad-hoc networks is contact-based architecture. In this architecture, each node assigns a unique zone for itself and each node keeps all information from inside the zone, as well as some from outside the zone, which is called contact. Reducing the overlap between different zones of a node and its contacts increases its performance, therefore Edge Method (EM) is designed for this purpose. Contacts selected by EM do not have any overlap with their sources, but for choosing the contact a vast amount of information must be transmitted. In this article, we will offer a new protocol for contact selection, which is called PEM. The objective would be reducing the volume of transmitted information, using Non-Uniform Dissemination Probabilistic Protocols. Consumed energy for contact selection is a function of the size of transmitted information between nodes. Therefore, by reducing the content of contact selection message using the PEM will decrease the consumed energy. For evaluation of the PEM we applied the simulation method. Results indicated that PEM consumes less energy compared to EM, and by increasing the number of nodes (level of nodes), performance of PEM will improve in comparison with EM.Keywords: wireless ad-hoc networks, contact selection, method for CARD, energy-efficient
Procedia PDF Downloads 29011859 Utilizing Radio as a Resource Alternative for Disseminating Information to University Students in Ibadan, Nigeria: A Study of Lead City FM and Diamond FM Radio Stations
Authors: Olufemi Sunday Onabajo
Abstract:
Radio according to communication scholars is a veritable instrument of mass education. However, its full potentials in boosting higher education have not been realized because of the commercial nature of radio stations in Nigeria. The licensing of campus radio for disseminating information on university curricular is aimed at reinforcing information shared during face to face teaching. This study anchored on Agenda Setting and Technology determinism theories seeks to find out the extent to which university students in Lead City University and University of Ibadan, Nigeria have keyed-in to the philosophy of their campus radio – Lead City FM and Diamond FM in making information dissemination in their domiciled universities less cumbersome. The study employs both qualitative and quantitative methods though the use of depth interview for ten (10) academic staff and five (5) radio personnel of both radio stations; and a questionnaire addressed to 200 students of both institutions using the systematic random sampling technique. The data collected was analyzed using simple percentage and chi-square one tail test, and it was discovered that students of both universities and their radio personnel are yet to realize the potentials of campus radio as a resource alternative to effective learning, and recommends the coming together of all stakeholders to articulate the way forward.Keywords: disseminating information, effective learning, resource alternative, utilizing radio
Procedia PDF Downloads 29811858 Malware Beaconing Detection by Mining Large-scale DNS Logs for Targeted Attack Identification
Authors: Andrii Shalaginov, Katrin Franke, Xiongwei Huang
Abstract:
One of the leading problems in Cyber Security today is the emergence of targeted attacks conducted by adversaries with access to sophisticated tools. These attacks usually steal senior level employee system privileges, in order to gain unauthorized access to confidential knowledge and valuable intellectual property. Malware used for initial compromise of the systems are sophisticated and may target zero-day vulnerabilities. In this work we utilize common behaviour of malware called ”beacon”, which implies that infected hosts communicate to Command and Control servers at regular intervals that have relatively small time variations. By analysing such beacon activity through passive network monitoring, it is possible to detect potential malware infections. So, we focus on time gaps as indicators of possible C2 activity in targeted enterprise networks. We represent DNS log files as a graph, whose vertices are destination domains and edges are timestamps. Then by using four periodicity detection algorithms for each pair of internal-external communications, we check timestamp sequences to identify the beacon activities. Finally, based on the graph structure, we infer the existence of other infected hosts and malicious domains enrolled in the attack activities.Keywords: malware detection, network security, targeted attack, computational intelligence
Procedia PDF Downloads 26411857 Computer Aided Assembly Attributes Retrieval Methods for Automated Assembly Sequence Generation
Authors: M. V. A. Raju Bahubalendruni, Bibhuti Bhusan Biswal, B. B. V. L. Deepak
Abstract:
Achieving an appropriate assembly sequence needs deep verification for its physical feasibility. For this purpose, industrial engineers use several assembly predicates; namely, liaison, geometric feasibility, stability and mechanical feasibility. However, testing an assembly sequence for these predicates requires huge assembly information. Extracting such assembly information from an assembled product is a time consuming and highly skillful task with complex reasoning methods. In this paper, computer aided methods are proposed to extract all the necessary assembly information from computer aided design (CAD) environment in order to perform the assembly sequence planning efficiently. These methods use preliminary capabilities of three-dimensional solid modelling and assembly modelling methods used in CAD software considering equilibrium laws of physical bodies.Keywords: assembly automation, assembly attributes, assembly, CAD
Procedia PDF Downloads 30511856 The Representation of Migrants in the UK and Saudi Arabia Press: A Cross-Linguistic Discourse Analysis Study
Authors: Eman Alatawi
Abstract:
The world is currently experiencing an upsurge in the number of international migrants, which has reached 281 million worldwide; in particular, both the UK and Saudi Arabia have recently been faced with an unprecedented number of immigrants. As a result, the media in these two countries is constantly posting news about the issue, and newspapers, in particular, play a vital role in shaping the public’s view of immigration issues. Because the media is an influential tool in society, it has the ability to construct a specific image of migrants and influence public opinion concerning immigrant groups. However, most of the existing studies have addressed the plight of migrants in the UK, Europe, and the US, and few have considered the Middle East; specifically, there is a pressing need for studies that focus on the press in Saudi Arabia, which is one of the main countries that is experiencing immigration at a tremendous rate. This paper employs critical discourse analysis (CDA) to examine the depiction of migrants in the British and Saudi Arabian media in order to explore the involvement of three linguistic features in the media’s representation of migrant-related topics. These linguistic features are the names, metaphors, and collocations that the press in the UK and in Saudi Arabia uses to describe migrants; the impact of these depictions is also considered. This comparative study could create a better understanding of how the Saudi Arabian press presents the topic of migrants and immigration, which will assist in extending the understanding of migration discourses beyond an Anglo-centric viewpoint. The main finding of this study was that both British and Saudi Arabian newspapers tended to represent migrants’ issues by painting migrants in a negative light through the use of negative references or names, metaphors, and collocations; furthermore, the media’s negative stereotyping of migrants was found to be consistent, which could have an influence on the public’s opinion of these minority groups. Such observations show that the issue is not as simple as individuals, press systems, or political affiliations.Keywords: representation, migrants, the UK press, Saudi Arabia press, cross-linguistic, discourse analysis
Procedia PDF Downloads 8011855 Tourism in the Information Age
Authors: Suleyman Karacor
Abstract:
The main purpose of this study is to investigate tourism marketing in the information age because of the importance and sensitivity. In the twenty-first century as a result of today's the increasing competition and product diversification in the tourism sector, tourism businesses must take into account exogenous variables such as new technological developments, commercial experience and consumer demand. In the information age, tourist product consumers tend to reserve their leisure time and expenditure on more active opportunities for different experiences instead of living the same experience again. Increasing the number of agents in the tourism sector, travel opportunities offering different experiences and more intensive use of modern technology helps to present diversification of leisure activities for tourists. From the perspective of tourists, travel costs are still important for buying the touristic products but maintaining a high level of tourist satisfaction is also of increasing importance. Tourists tend to prefer activities that add value. A real tourist product must be able to create value and new priorities for tourists. Therefore this study aims to review recent significant developments in international tourism marketing research and practices. To this end, this study reviews tourism marketing-focused articles.Keywords: information age, tourism marketing, tourism marketing mix, management
Procedia PDF Downloads 42611854 Research on Evaluation Method of Urban Road Section Traffic Safety Status Based on Video Information
Authors: Qiang Zhang, Xiaojian Hu
Abstract:
Aiming at the problem of the existing real-time evaluation methods for traffic safety status, a video information-based urban road section traffic safety status evaluation method was established, and the rapid detection method of traffic flow parameters based on video information is analyzed. The concept of the speed dispersion of the road section that affects the traffic safety state of the urban road section is proposed, and the method of evaluating the traffic safety state of the urban road section based on the speed dispersion of the road section is established. Experiments show that the proposed method can reasonably evaluate the safety status of urban roads in real-time, and the evaluation results can provide a corresponding basis for the traffic management department to formulate an effective urban road section traffic safety improvement plan.Keywords: intelligent transportation system, road traffic safety, video information, vehicle speed dispersion
Procedia PDF Downloads 16411853 Use of Satellite Imaging to Understand Earth’s Surface Features: A Roadmap
Authors: Sabri Serkan Gulluoglu
Abstract:
It is possible with Geographic Information Systems (GIS) that the information about all natural and artificial resources on the earth is obtained taking advantage of satellite images are obtained by remote sensing techniques. However, determination of unknown sources, mapping of the distribution and efficient evaluation of resources are defined may not be possible with the original image. For this reasons, some process steps are needed like transformation, pre-processing, image enhancement and classification to provide the most accurate assessment numerically and visually. Many studies which present the phases of obtaining and processing of the satellite images have examined in the literature study. The research showed that the determination of the process steps may be followed at this subject with the existence of a common whole may provide to progress the process rapidly for the necessary and possible studies which will be.Keywords: remote sensing, satellite imaging, gis, computer science, information
Procedia PDF Downloads 31811852 Intelligent Transport System: Classification of Traffic Signs Using Deep Neural Networks in Real Time
Authors: Anukriti Kumar, Tanmay Singh, Dinesh Kumar Vishwakarma
Abstract:
Traffic control has been one of the most common and irritating problems since the time automobiles have hit the roads. Problems like traffic congestion have led to a significant time burden around the world and one significant solution to these problems can be the proper implementation of the Intelligent Transport System (ITS). It involves the integration of various tools like smart sensors, artificial intelligence, position technologies and mobile data services to manage traffic flow, reduce congestion and enhance driver's ability to avoid accidents during adverse weather. Road and traffic signs’ recognition is an emerging field of research in ITS. Classification problem of traffic signs needs to be solved as it is a major step in our journey towards building semi-autonomous/autonomous driving systems. The purpose of this work focuses on implementing an approach to solve the problem of traffic sign classification by developing a Convolutional Neural Network (CNN) classifier using the GTSRB (German Traffic Sign Recognition Benchmark) dataset. Rather than using hand-crafted features, our model addresses the concern of exploding huge parameters and data method augmentations. Our model achieved an accuracy of around 97.6% which is comparable to various state-of-the-art architectures.Keywords: multiclass classification, convolution neural network, OpenCV
Procedia PDF Downloads 17611851 Big Data Applications for Transportation Planning
Authors: Antonella Falanga, Armando Cartenì
Abstract:
"Big data" refers to extremely vast and complex sets of data, encompassing extraordinarily large and intricate datasets that require specific tools for meaningful analysis and processing. These datasets can stem from diverse origins like sensors, mobile devices, online transactions, social media platforms, and more. The utilization of big data is pivotal, offering the chance to leverage vast information for substantial advantages across diverse fields, thereby enhancing comprehension, decision-making, efficiency, and fostering innovation in various domains. Big data, distinguished by its remarkable attributes of enormous volume, high velocity, diverse variety, and significant value, represent a transformative force reshaping the industry worldwide. Their pervasive impact continues to unlock new possibilities, driving innovation and advancements in technology, decision-making processes, and societal progress in an increasingly data-centric world. The use of these technologies is becoming more widespread, facilitating and accelerating operations that were once much more complicated. In particular, big data impacts across multiple sectors such as business and commerce, healthcare and science, finance, education, geography, agriculture, media and entertainment and also mobility and logistics. Within the transportation sector, which is the focus of this study, big data applications encompass a wide variety, spanning across optimization in vehicle routing, real-time traffic management and monitoring, logistics efficiency, reduction of travel times and congestion, enhancement of the overall transportation systems, but also mitigation of pollutant emissions contributing to environmental sustainability. Meanwhile, in public administration and the development of smart cities, big data aids in improving public services, urban planning, and decision-making processes, leading to more efficient and sustainable urban environments. Access to vast data reservoirs enables deeper insights, revealing hidden patterns and facilitating more precise and timely decision-making. Additionally, advancements in cloud computing and artificial intelligence (AI) have further amplified the potential of big data, enabling more sophisticated and comprehensive analyses. Certainly, utilizing big data presents various advantages but also entails several challenges regarding data privacy and security, ensuring data quality, managing and storing large volumes of data effectively, integrating data from diverse sources, the need for specialized skills to interpret analysis results, ethical considerations in data use, and evaluating costs against benefits. Addressing these difficulties requires well-structured strategies and policies to balance the benefits of big data with privacy, security, and efficient data management concerns. Building upon these premises, the current research investigates the efficacy and influence of big data by conducting an overview of the primary and recent implementations of big data in transportation systems. Overall, this research allows us to conclude that big data better provide to enhance rational decision-making for mobility choices and is imperative for adeptly planning and allocating investments in transportation infrastructures and services.Keywords: big data, public transport, sustainable mobility, transport demand, transportation planning
Procedia PDF Downloads 6011850 Social Networking Sites: A Platform for Communication and Collaboration for Visually Impaired
Authors: Sufia Khowaja, Nishat Fatima
Abstract:
Social networking sites are significant for visually impaired to overcome the unique challenges they face and access the resources they need to succeed in their education and beyond which might be difficult to obtain through traditional means. It provides them an opportunity to build relationships, stay connected with their support network as well as to develop social skills which give them emotional support to fell less isolated. In this connection the study is conducted with the aim to determine the use of social networking sites, purpose of using and activities performed by visually impaired at Delhi University, Delhi, Jawaharlal Nehru University, Delhi and Jamia Milia Islamia, Delhi. The study followed survey technique in which structured interview is followed to collect data from 137 visually impaired students and analysed using ‘SPSS ver23’. The findings of the study revealed that mostly used social networking sites are whatsapp by 89.23% students of DU, 95.12% of JNU, 87.09% of JMI, followed by e-mail by 78.46% of DU, 78.04% of JNU, 64.51%; youtube by 73.84% DU, 90.24% JNU, 80.64% JMI. Purpose for using these sites is for academics mentioned by 96.92% DU, 100% JNU, 93.54% JMI. Activities performed on sites are sending and receiving messaging 96.92% DU, 92.68% JNU, 93.55% JMI, communicating with friends and family as well as getting academic information. Findings of the study will be helpful for libraries to disseminate their services and resources as well as latest updates to their visually impaired users with the help of most used tools.Keywords: social networking sites, visually impaired, Delhi University, Jawaharlal Nehru University, Jamia Milia Islamia
Procedia PDF Downloads 9011849 SCNet: A Vehicle Color Classification Network Based on Spatial Cluster Loss and Channel Attention Mechanism
Authors: Fei Gao, Xinyang Dong, Yisu Ge, Shufang Lu, Libo Weng
Abstract:
Vehicle color recognition plays an important role in traffic accident investigation. However, due to the influence of illumination, weather, and noise, vehicle color recognition still faces challenges. In this paper, a vehicle color classification network based on spatial cluster loss and channel attention mechanism (SCNet) is proposed for vehicle color recognition. A channel attention module is applied to extract the features of vehicle color representative regions and reduce the weight of nonrepresentative color regions in the channel. The proposed loss function, called spatial clustering loss (SC-loss), consists of two channel-specific components, such as a concentration component and a diversity component. The concentration component forces all feature channels belonging to the same class to be concentrated through the channel cluster. The diversity components impose additional constraints on the channels through the mean distance coefficient, making them mutually exclusive in spatial dimensions. In the comparison experiments, the proposed method can achieve state-of-the-art performance on the public datasets, VCD, and VeRi, which are 96.1% and 96.2%, respectively. In addition, the ablation experiment further proves that SC-loss can effectively improve the accuracy of vehicle color recognition.Keywords: feature extraction, convolutional neural networks, intelligent transportation, vehicle color recognition
Procedia PDF Downloads 18311848 Wet Processing of Algae for Protein and Carbohydrate Recovery as Co-Product of Algal Oil
Authors: Sahil Kumar, Rajaram Ghadge, Ramesh Bhujade
Abstract:
Historically, lipid extraction from dried algal biomass remained a focus area of the algal research. It has been realized over the past few years that the lipid-centric approach and conversion technologies that require dry algal biomass have several challenges. Algal culture in cultivation systems contains more than 99% water, with algal concentrations of just a few hundred milligrams per liter ( < 0.05 wt%), which makes harvesting and drying energy intensive. Drying the algal biomass followed by extraction also entails the loss of water and nutrients. In view of these challenges, focus has shifted toward developing processes that will enable oil production from wet algal biomass without drying. Hydrothermal liquefaction (HTL), an emerging technology, is a thermo-chemical conversion process that converts wet biomass to oil and gas using water as a solvent at high temperature and high pressure. HTL processes wet algal slurry containing more than 80% water and significantly reduces the adverse cost impact owing to drying the algal biomass. HTL, being inherently feedstock agnostic, i.e., can convert carbohydrates and proteins also to fuels and recovers water and nutrients. It is most effective with low-lipid (10--30%) algal biomass, and bio-crude yield is two to four times higher than the lipid content in the feedstock. In the early 2010s, research remained focused on increasing the oil yield by optimizing the process conditions of HTL. However, various techno-economic studies showed that simply converting algal biomass to only oil does not make economic sense, particularly in view of low crude oil prices. Making the best use of every component of algae is a key for economic viability of algal to oil process. On investigation of HTL reactions at the molecular level, it has been observed that sequential HTL has the potential to recover value-added products along with biocrude and improve the overall economics of the process. This potential of sequential HTL makes it a most promising technology for converting wet waste to wealth. In this presentation, we will share our experience on the techno-economic and engineering aspects of sequential HTL for conversion of algal biomass to algal bio-oil and co-products.Keywords: algae, biomass, lipid, protein
Procedia PDF Downloads 21411847 Integrated Information System on Human Resource Management in Project-Based Organizations
Authors: Akbar Farahani, Afsaneh Hassani, Peyman M. Farkhondeh
Abstract:
Human Resource Management as one of the core processes of the project-based companies, despite its key role in the success and competitive advantage, is relatively unknown. In the project-based companies, due to the accelerated movement of knowledge in the work activities and the temporary nature of the project, the need to develop mechanisms for achieving optimal management of this issues is very challenging. Approach to human resource management in these companies evolves with goals, strategies, and operational processes. Therefore, the need for appropriate tools to facilitate implementation of the optimized human resource management in the project is more than before,Which currently with the development of information technology and modern communication, appropriate to address the optimal approach for dynamic management of human resources in the project have been provided.This is done by using the referral system implemented in Mahab GCE that provides 1: the ability to use humans in projects without geographic limitation and 2:information on the activities and outcomes of referrals.Furthermore, by using this system, recording the lessons learned after any particular activity on projects,accessing quantitative information, procedures, documentation of learned practices that have been stored in the data base as well as using them in future projects is provided.Keywords: human resource management, project base company, ERP, referrals system
Procedia PDF Downloads 47711846 Elucidation of the Sequential Transcriptional Activity in Escherichia coli Using Time-Series RNA-Seq Data
Authors: Pui Shan Wong, Kosuke Tashiro, Satoru Kuhara, Sachiyo Aburatani
Abstract:
Functional genomics and gene regulation inference has readily expanded our knowledge and understanding of gene interactions with regards to expression regulation. With the advancement of transcriptome sequencing in time-series comes the ability to study the sequential changes of the transcriptome. This method presented here works to augment existing regulation networks accumulated in literature with transcriptome data gathered from time-series experiments to construct a sequential representation of transcription factor activity. This method is applied on a time-series RNA-Seq data set from Escherichia coli as it transitions from growth to stationary phase over five hours. Investigations are conducted on the various metabolic activities in gene regulation processes by taking advantage of the correlation between regulatory gene pairs to examine their activity on a dynamic network. Especially, the changes in metabolic activity during phase transition are analyzed with focus on the pagP gene as well as other associated transcription factors. The visualization of the sequential transcriptional activity is used to describe the change in metabolic pathway activity originating from the pagP transcription factor, phoP. The results show a shift from amino acid and nucleic acid metabolism, to energy metabolism during the transition to stationary phase in E. coli.Keywords: Escherichia coli, gene regulation, network, time-series
Procedia PDF Downloads 37211845 Convex Restrictions for Outage Constrained MU-MISO Downlink under Imperfect Channel State Information
Authors: A. Preetha Priyadharshini, S. B. M. Priya
Abstract:
In this paper, we consider the MU-MISO downlink scenario, under imperfect channel state information (CSI). The main issue in imperfect CSI is to keep the probability of each user achievable outage rate below the given threshold level. Such a rate outage constraints present significant and analytical challenges. There are many probabilistic methods are used to minimize the transmit optimization problem under imperfect CSI. Here, decomposition based large deviation inequality and Bernstein type inequality convex restriction methods are used to perform the optimization problem under imperfect CSI. These methods are used for achieving improved output quality and lower complexity. They provide a safe tractable approximation of the original rate outage constraints. Based on these method implementations, performance has been evaluated in the terms of feasible rate and average transmission power. The simulation results are shown that all the two methods offer significantly improved outage quality and lower computational complexity.Keywords: imperfect channel state information, outage probability, multiuser- multi input single output, channel state information
Procedia PDF Downloads 81411844 Parallelization of Random Accessible Progressive Streaming of Compressed 3D Models over Web
Authors: Aayushi Somani, Siba P. Samal
Abstract:
Three-dimensional (3D) meshes are data structures, which store geometric information of an object or scene, generally in the form of vertices and edges. Current technology in laser scanning and other geometric data acquisition technologies acquire high resolution sampling which leads to high resolution meshes. While high resolution meshes give better quality rendering and hence is used often, the processing, as well as storage of 3D meshes, is currently resource-intensive. At the same time, web applications for data processing have become ubiquitous owing to their accessibility. For 3D meshes, the advancement of 3D web technologies, such as WebGL, WebVR, has enabled high fidelity rendering of huge meshes. However, there exists a gap in ability to stream huge meshes to a native client and browser application due to high network latency. Also, there is an inherent delay of loading WebGL pages due to large and complex models. The focus of our work is to identify the challenges faced when such meshes are streamed into and processed on hand-held devices, owing to its limited resources. One of the solutions that are conventionally used in the graphics community to alleviate resource limitations is mesh compression. Our approach deals with a two-step approach for random accessible progressive compression and its parallel implementation. The first step includes partition of the original mesh to multiple sub-meshes, and then we invoke data parallelism on these sub-meshes for its compression. Subsequent threaded decompression logic is implemented inside the Web Browser Engine with modification of WebGL implementation in Chromium open source engine. This concept can be used to completely revolutionize the way e-commerce and Virtual Reality technology works for consumer electronic devices. These objects can be compressed in the server and can be transmitted over the network. The progressive decompression can be performed on the client device and rendered. Multiple views currently used in e-commerce sites for viewing the same product from different angles can be replaced by a single progressive model for better UX and smoother user experience. Can also be used in WebVR for commonly and most widely used activities like virtual reality shopping, watching movies and playing games. Our experiments and comparison with existing techniques show encouraging results in terms of latency (compressed size is ~10-15% of the original mesh), processing time (20-22% increase over serial implementation) and quality of user experience in web browser.Keywords: 3D compression, 3D mesh, 3D web, chromium, client-server architecture, e-commerce, level of details, parallelization, progressive compression, WebGL, WebVR
Procedia PDF Downloads 17011843 Profitability Assessment of Granite Aggregate Production and the Development of a Profit Assessment Model
Authors: Melodi Mbuyi Mata, Blessing Olamide Taiwo, Afolabi Ayodele David
Abstract:
The purpose of this research is to create empirical models for assessing the profitability of granite aggregate production in Akure, Ondo state aggregate quarries. In addition, an artificial neural network (ANN) model and multivariate predicting models for granite profitability were developed in the study. A formal survey questionnaire was used to collect data for the study. The data extracted from the case study mine for this study includes granite marketing operations, royalty, production costs, and mine production information. The following methods were used to achieve the goal of this study: descriptive statistics, MATLAB 2017, and SPSS16.0 software in analyzing and modeling the data collected from granite traders in the study areas. The ANN and Multi Variant Regression models' prediction accuracy was compared using a coefficient of determination (R²), Root mean square error (RMSE), and mean square error (MSE). Due to the high prediction error, the model evaluation indices revealed that the ANN model was suitable for predicting generated profit in a typical quarry. More quarries in Nigeria's southwest region and other geopolitical zones should be considered to improve ANN prediction accuracy.Keywords: national development, granite, profitability assessment, ANN models
Procedia PDF Downloads 10111842 Self-Efficacy Perceptions of Pre-Service Art and Music Teachers towards the Use of Information and Communication Technologies
Authors: Agah Tugrul Korucu
Abstract:
Information and communication technologies have become an important part of our daily lives with significant investments in technology in the 21st century. Individuals are more willing to design and implement computer-related activities, and they are the main component of computer self-efficacy and self-efficacy related to the fact that the increase in information technology, with operations in parallel with these activities more successful. The Self-efficacy level is a significant factor which determines how individuals act in events, situations and difficult processes. It is observed that individuals with higher self-efficacy perception of computers who encounter problems related to computer use overcome them more easily. Therefore, this study aimed to examine self-efficacy perceptions of pre-service art and music teachers towards the use of information and communication technologies in terms of different variables. Research group consists of 60 pre-service teachers who are studying at Necmettin Erbakan University Ahmet Keleşoğlu Faculty of Education Art and Music department. As data collection tool of the study; “personal information form” developed by the researcher and used to collect demographic data and "the perception scale related to self-efficacy of informational technology" are used. The scale is 5-point Likert-type scale. It consists of 27 items. The Kaiser-Meyer-Olkin (KMO) sample compliance value is found 0.959. The Cronbach alpha reliability coefficient of the scale is found to be 0.97. computer-based statistical software package (SPSS 21.0) is used in order to analyze the data collected by data collection tools; descriptive statistics, t-test, analysis of variance are used as statistical techniques.Keywords: self-efficacy perceptions, teacher candidate, information and communication technologies, art teacher
Procedia PDF Downloads 32611841 End to End Monitoring in Oracle Fusion Middleware for Data Verification
Authors: Syed Kashif Ali, Usman Javaid, Abdullah Chohan
Abstract:
In large enterprises multiple departments use different sort of information systems and databases according to their needs. These systems are independent and heterogeneous in nature and sharing information/data between these systems is not an easy task. The usage of middleware technologies have made data sharing between systems very easy. However, monitoring the exchange of data/information for verification purposes between target and source systems is often complex or impossible for maintenance department due to security/access privileges on target and source systems. In this paper, we are intended to present our experience of an end to end data monitoring approach at middle ware level implemented in Oracle BPEL for data verification without any help of monitoring tool.Keywords: service level agreement, SOA, BPEL, oracle fusion middleware, web service monitoring
Procedia PDF Downloads 481