Search results for: Information and data requirements
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 31910

Search results for: Information and data requirements

31610 Do Interventions for Increasing Minorities' Access to Higher Education Work? The Case of Ethiopians in Israel

Authors: F. Nasser-Abu Alhija

Abstract:

In many countries, much efforts and resources are devoted to empowering and integrating minorities within the mainstream population. Major ventures in this route are crafted in higher education institutions where different outreach programs and methods such as lenient entry requirements, monitory incentives, learning skills workshops, tutoring and mentoring, are utilized. Although there is some information regarding these programs, their effectiveness still needs to be thoroughly examined. The Ethiopian community In Israel is one of the minority groups that has been targeted by sponsoring foundations and higher education institutions with the aim to ease the access, persistence and success of its young people in higher education and later in the job market. The evaluation study we propose to present focuses on the implementation of a program designed for this purpose. This program offers relevant candidates for study at a prestigious university a variety of generous incentives that include tuitions, livening allowance, tutoring, mentoring, skills and empowerment workshops and cultural meetings. Ten students were selected for the program and they started their studies in different subject areas before three and half years. A longitudinal evaluation has been conducted since the implementation of the program. Data were collected from different sources: participating students, program coordinator, mentors, tutors, program documents and university records. Questionnaires and interviews were used for collecting data on the different components of the program and on participants' perception of their effectiveness. Participants indicate that the lenient entry requirements and the monitory incentives are critical for starting their studies. During the first year, skills and empowering workshops, torturing and mentoring were evaluated as very important for persistence and success in studies. Tutoring was perceived as very important also at the second year but less importance is attributed to mentoring. Mixed results regarding integration in the Israeli culture emerged. The results are discussed with reference to findings from different settings around the world.

Keywords: access to higher education, minority groups, monitory incentives, torturing, mentoring

Procedia PDF Downloads 369
31609 Refining Employee's Customer Service Performance through an Inter-Organizational Climate Study: A Way Forward

Authors: Zainal Abu Zatim, Hafizah Omar Zaki

Abstract:

Substantial research had been done on refining employee’s customer service performance. Thus, there were very limited empirical studies that are engage in an inter-organizational climate study in assessing employee’s customer service performance. With the current economic situation as well as emerging needs and requirements, all businesses either from public or private sector serving customers put greater attention on fulfilling those needs and requirements. In this state of affairs, the act of polishing its employees’ skills, knowledge, teamwork and passion is very important in ensuring better performance deliverance. A study conducted in one of the telecommunication service provider company in Malaysia had been done to test its inter-organizational climate study. The Internal Climate Study was done to benchmark opinions and perceptions of its employees. The study had provided baseline information about perceptions that exist in the internal environment and ways forward to improve customer service performance. The approach used is through the use of focus group and qualitative interview.

Keywords: employees, Customer Service Performance, inter-organizational climate study, public and private sector

Procedia PDF Downloads 395
31608 The Influence of Gender on Job-Competencies Requirements of Chemical-Based Industries and Undergraduate-Competencies Acquisition of Chemists in South West, Nigeria

Authors: Rachael Olatoun Okunuga

Abstract:

Developing young people’s employability is a key policy issue for ensuring their successful transition to the labour market and their access to career oriented employment. The youths of today irrespective of their gender need to acquire the knowledge, skills and attitudes that will enable them to create or find jobs as well as cope with unpredictable labour market changes throughout their working lives. In a study carried out to determine the influence of gender on job-competencies requirements of chemical-based industries and undergraduate-competencies acquisition by chemists working in the industries, all chemistry graduates working in twenty (20) chemical-based industries that were randomly selected from six sectors of chemical-based industries in Lagos and Ogun States of Nigeria were administered with Job-competencies required and undergraduate-competencies acquired assessment questionnaire. The data were analysed using means and independent sample t-test. The findings revealed that the population of female chemists working in chemical-based industries is low compared with the number of male chemists; furthermore, job-competencies requirements are found not to be gender sensitive while there is no significant difference in undergraduate-competencies acquisition of male and female chemists. This suggests that females should be given the same opportunity of employment in chemical-based industries as their male counterparts. The study also revealed the level of acquisition of undergraduate competencies as related to the needs of chemical-based industries.

Keywords: knowledge, skill, attitude, acquired, required, employability

Procedia PDF Downloads 371
31607 Information Retrieval for Kafficho Language

Authors: Mareye Zeleke Mekonen

Abstract:

The Kafficho language has distinct issues in information retrieval because of its restricted resources and dearth of standardized methods. In this endeavor, with the cooperation and support of linguists and native speakers, we investigate the creation of information retrieval systems specifically designed for the Kafficho language. The Kafficho information retrieval system allows Kafficho speakers to access information easily in an efficient and effective way. Our objective is to conduct an information retrieval experiment using 220 Kafficho text files, including fifteen sample questions. Tokenization, normalization, stop word removal, stemming, and other data pre-processing chores, together with additional tasks like term weighting, were prerequisites for the vector space model to represent each page and a particular query. The three well-known measurement metrics we used for our word were Precision, Recall, and and F-measure, with values of 87%, 28%, and 35%, respectively. This demonstrates how well the Kaffiho information retrieval system performed well while utilizing the vector space paradigm.

Keywords: Kafficho, information retrieval, stemming, vector space

Procedia PDF Downloads 50
31606 Partnership Oriented Innovation Alliance Strategy Based on Market Feedback

Authors: Victor Romanov, Daria Efimenko

Abstract:

The focus on innovation in modern economy is the main factor in surviving business in a competitive environment. The innovations are based on the search and use of knowledge in a global context. Nowadays consumers and market demand are the main innovation drivers. This leads to build a business as a system with feedback, promptly restructuring production and innovation implementation in response to market demands. In modern knowledge economy, because of speed of technical progress, the product's lifecycle became much shorter, what makes more stringent requirements for innovation implementation on the enterprises of and therefore the possibility for enterprise for receiving extra income is decreasing. This circumstance imposes additional requirements for the replacement of obsolete products and the prompt release of innovative products to the market. The development of information technologies has led to the fact that only in the conditions of partnership and knowledge sharing with partners it is possible to update products quickly for innovative products. Many companies pay attention to updating innovations through the search for new partners, but the task of finding new partners presents some difficulties. The search for a suitable one includes several stages such as: determining the moment of innovation-critical, introducing a search, identifying search criteria, justifying and deciding on the choice of a partner. No less important is the question of how to manage an innovative product in response to a changing market. The article considers the problems of information support for the search for the source of innovation and partnership to decrease the time for implementation of novelty products.

Keywords: partnership, novelty, market feedback, alliance

Procedia PDF Downloads 191
31605 Mitigating the Unwillingness of e-Forums Members to Engage in Information Exchange

Authors: Dora Triki, Irena Vida, Claude Obadia

Abstract:

Social networks such as e-Forums or dating sites often face the reluctance of key members to participate. Relying on the conation theory, this study investigates this phenomenon and proposes solutions to mitigate the issue. We show that highly experienced e-Forum members refuse to share business information in a peer to peer information exchange forums. However, forums managers can mitigate this behavior by developing a sentiment of belongingness to the network. Furthermore, by selecting only elite forum participants with ample experience, they can reduce the reluctance of key information providers to engage in information exchange. Our hypotheses are tested with PLS structural equations modeling using survey data from members of a French e-Forum dedicated to the exchange of business information about exporting.

Keywords: conation, e-Forum, information exchange, members participation

Procedia PDF Downloads 151
31604 Spatial Behavioral Model-Based Dynamic Data-Driven Diagram Information Model

Authors: Chiung-Hui Chen

Abstract:

Diagram and drawing are important ways to communicate and the reproduce of architectural design, Due to the development of information and communication technology, the professional thinking of architecture and interior design are also change rapidly. In development process of design, diagram always play very important role. This study is based on diagram theories, observe and record interaction between man and objects, objects and space, and space and time in a modern nuclear family. Construct a method for diagram to systematically and visualized describe the space plan of a modern nuclear family toward a intelligent design, to assist designer to retrieve information and check/review event pattern of past and present.

Keywords: digital diagram, information model, context aware, data analysis

Procedia PDF Downloads 329
31603 Data Augmentation for Automatic Graphical User Interface Generation Based on Generative Adversarial Network

Authors: Xulu Yao, Moi Hoon Yap, Yanlong Zhang

Abstract:

As a branch of artificial neural network, deep learning is widely used in the field of image recognition, but the lack of its dataset leads to imperfect model learning. By analysing the data scale requirements of deep learning and aiming at the application in GUI generation, it is found that the collection of GUI dataset is a time-consuming and labor-consuming project, which is difficult to meet the needs of current deep learning network. To solve this problem, this paper proposes a semi-supervised deep learning model that relies on the original small-scale datasets to produce a large number of reliable data sets. By combining the cyclic neural network with the generated countermeasure network, the cyclic neural network can learn the sequence relationship and characteristics of data, make the generated countermeasure network generate reasonable data, and then expand the Rico dataset. Relying on the network structure, the characteristics of collected data can be well analysed, and a large number of reasonable data can be generated according to these characteristics. After data processing, a reliable dataset for model training can be formed, which alleviates the problem of dataset shortage in deep learning.

Keywords: GUI, deep learning, GAN, data augmentation

Procedia PDF Downloads 178
31602 Nalanda ‘School of Joy’: Teaching Learning Strategies and Support System, for Implementing Child-Friendly Education in Bangladesh

Authors: Sufia Ferdousi

Abstract:

Child-friendly education (CFE) is very important for the children, especially the early year’s students, because it fosters the holistic development of a child. Teacher plays a key role in creating child-friendly education. This study intends to learn about child-friendly education in Bangladesh. The purpose of the study is to explore how CFE is being practiced in Bangladesh. The study attempted to fulfill the purpose through case study investigation. One school, named Nalanda, was selected for the study as it claims to run the school through CFE approach. The objective of the study was to identify, how this school is different from the other schools in Bangladesh, to explore overall teaching learning system like, curriculum, teaching strategies, assessments and to investigate the support system for Child Friendly Education provided to the teachers through training or mentoring. The nature of the case study was qualitative method to get maximum information from the students, parents, teachers and school authorities. The findings were based on 3 classroom observations, interviews with 1 teacher, 1 head teacher and 1 trainer, FGD with 10 students and 6 parents, were used to collect the data. It has been found that Nalanda is different than the other schools in Bangladesh in terms of, parents’ motivation about school curriculum, and sufficiency of teachers’ knowledge on joyful learning/child-friendly learning. The students took part in the extracurricular activities alongside the national curriculum. Teachers showed particular strength in the teaching learning strategies, using materials and assessment. And Nalanda gives strong support for teacher’s training. In conclusion, The Nalanda School in Dhaka was found appropriate for the requirements of Child-friendly education.

Keywords: child friendly education, overall teaching learning system, the requirements of child-friendly education, the alternative education approach

Procedia PDF Downloads 244
31601 Development of an Automatic Monitoring System Based on the Open Architecture Concept

Authors: Andrii Biloshchytskyi, Serik Omirbayev, Alexandr Neftissov, Sapar Toxanov, Svitlana Biloshchytska, Adil Faizullin

Abstract:

Kazakhstan has adopted a carbon neutrality strategy until 2060. In accordance with this strategy, it is necessary to introduce various tools to maintain the environmental safety of the environment. The use of IoT, in combination with the characteristics and requirements of Kazakhstan's environmental legislation, makes it possible to develop a modern environmental monitoring system. The article proposes a solution for developing an example of an automated system for the continuous collection of data on the concentration of pollutants in the atmosphere based on an open architecture. The Audino-based device acts as a microcontroller. It should be noted that the transmission of measured values is carried out via an open wireless communication protocol. The architecture of the system, which was used to build a prototype based on sensors, an Arduino microcontroller, and a wireless data transmission module, is presented. The selection of elementary components may change depending on the requirements of the system; the introduction of new units is limited by the number of ports. The openness of solutions allows you to change the configuration depending on the conditions. The advantages of the solutions are openness, low cost, versatility and mobility. However, there is no comparison of the working processes of the proposed solution with traditional ones.

Keywords: environmental monitoring, greenhouse gases emissions, environmental pollution, Industry 4.0, IoT, microcontroller, automated monitoring system.

Procedia PDF Downloads 40
31600 Is There a Group of "Digital Natives" at Secondary Schools?

Authors: L. Janská, J. Kubrický

Abstract:

The article describes a research focused on the influence of the information and communication technology (ICT) on the pupils' learning. The investigation deals with the influences that distinguish between the group of pupils influenced by ICT and the group of pupils not influenced by ICT. The group influenced by ICT should evince a different approach in number of areas (in managing of two and more activities at once, in a quick orientation and searching for information on the Internet, in an ability to quickly and effectively assess the data sources, in the assessment of attitudes and opinions of the other users of the network, in critical thinking, in the preference to work in teams, in the sharing of information and personal data via the virtual social networking, in insisting on the immediate reaction on their every action etc.).

Keywords: ICT influence, digital natives, pupil´s learning

Procedia PDF Downloads 287
31599 Does Stock Markets Asymmetric Information Affect Foreign Capital Flows?

Authors: Farid Habibi Tanha, Mojtaba Jahanbazi, Morteza Foroutan, Rasidah Mohd Rashid

Abstract:

This paper depicts the effects of asymmetric information in determining capital inflows to be captured through stock market microstructure. The model can explain several stylized facts regarding the capital immobility. The first phase of the research involves in collecting and refining 150,000,000 daily data of 11 stock markets over a period of one decade in an effort to minimize the impact of survivorship bias. Three micro techniques were used to measure information asymmetries. The final phase analyzes the model through panel data approach. As a unique contribution, this research will provide valuable information regarding negative effects of information asymmetries in stock markets on attracting foreign investments. The results of this study can be directly considered by policy makers to monitor and control changes of capital flow in order to keep market conditions in a healthy manner, by preventing and managing possible shocks to avoid sudden reversals and market failures.

Keywords: asymmetric information, capital inflow, market microstructure, investment

Procedia PDF Downloads 313
31598 Transmission Network Expansion Planning in Deregulated Power Systems to Facilitate Competition under Uncertainties

Authors: Hooshang Mohammad Alikhani, Javad Nikoukar

Abstract:

Restructuring and deregulation of power industry have changed the objectives of transmission expansion planning and increased the uncertainties. Due to these changes, new approaches and criteria are needed for transmission planning in deregulated power systems. The objective of this research work is to present a new approach for transmission expansion planning with considering new objectives and uncertainties in deregulated power systems. The approach must take into account the desires of all stakeholders in transmission expansion planning. Market based criteria must be defined to achieve the new objectives. Combination of market based criteria, technical criteria and economical criteria must be used for measuring the goodness of expansion plans to achieve market requirements, technical requirements, and economical requirements altogether.

Keywords: deregulated power systems, transmission network, stakeholder, energy systems

Procedia PDF Downloads 648
31597 Extending Theory of Planned Behavior to Modelling Chronic Patients’ Acceptance of Health Information: An Information Overload Perspective

Authors: Shu-Lien Chou, Chung-Feng Liu

Abstract:

Self-health management of chronic illnesses plays an important part in chronic illness treatments. However, various kinds of health information (health education materials) which government or healthcare institutions provide for patients may not achieve the expected outcome. One of the critical reasons affecting patients’ use intention could be patients’ perceived Information overload regarding the health information. This study proposed an extended model of Theory of Planned Behavior, which integrating perceived information overload as another construct to explore patients’ use intention of the health information for self-health management. The independent variables are attitude, subject norm, perceived behavior control and perceived information overload while the dependent variable is behavior intention to use the health information. The cross-sectional study used a structured questionnaire for data collection, focusing on the chronic patients with coronary artery disease (CAD), who are the potential users of the health information, in a medical center in Taiwan. Data were analyzed using descriptive statistics of the basic information distribution of the questionnaire respondents, and the Partial Least Squares (PLS) structural equation model to study the reliability and construct validity for testing our hypotheses. A total of 110 patients were enrolled in this study and 106 valid questionnaires were collected. The PLS analysis result indicates that the patients’ perceived information overload of health information contributes the most critical factor influencing the behavioral intention. Subjective norm and perceived behavioral control of TPB constructs had significant effects on patients’ intentions to use health information also, whereas the attitude construct did not. This study demonstrated a comprehensive framework, which extending perceived information overload into TPB model to predict patients’ behavioral intention of using heath information. We expect that the results of this study will provide useful insights for studying health information from the perspectives of academia, governments, and healthcare providers.

Keywords: chronic patients, health information, information overload, theory of planned behavior

Procedia PDF Downloads 432
31596 Optimal Pricing Based on Real Estate Demand Data

Authors: Vanessa Kummer, Maik Meusel

Abstract:

Real estate demand estimates are typically derived from transaction data. However, in regions with excess demand, transactions are driven by supply and therefore do not indicate what people are actually looking for. To estimate the demand for housing in Switzerland, search subscriptions from all important Swiss real estate platforms are used. These data do, however, suffer from missing information—for example, many users do not specify how many rooms they would like or what price they would be willing to pay. In economic analyses, it is often the case that only complete data is used. Usually, however, the proportion of complete data is rather small which leads to most information being neglected. Also, the data might have a strong distortion if it is complete. In addition, the reason that data is missing might itself also contain information, which is however ignored with that approach. An interesting issue is, therefore, if for economic analyses such as the one at hand, there is an added value by using the whole data set with the imputed missing values compared to using the usually small percentage of complete data (baseline). Also, it is interesting to see how different algorithms affect that result. The imputation of the missing data is done using unsupervised learning. Out of the numerous unsupervised learning approaches, the most common ones, such as clustering, principal component analysis, or neural networks techniques are applied. By training the model iteratively on the imputed data and, thereby, including the information of all data into the model, the distortion of the first training set—the complete data—vanishes. In a next step, the performances of the algorithms are measured. This is done by randomly creating missing values in subsets of the data, estimating those values with the relevant algorithms and several parameter combinations, and comparing the estimates to the actual data. After having found the optimal parameter set for each algorithm, the missing values are being imputed. Using the resulting data sets, the next step is to estimate the willingness to pay for real estate. This is done by fitting price distributions for real estate properties with certain characteristics, such as the region or the number of rooms. Based on these distributions, survival functions are computed to obtain the functional relationship between characteristics and selling probabilities. Comparing the survival functions shows that estimates which are based on imputed data sets do not differ significantly from each other; however, the demand estimate that is derived from the baseline data does. This indicates that the baseline data set does not include all available information and is therefore not representative for the entire sample. Also, demand estimates derived from the whole data set are much more accurate than the baseline estimation. Thus, in order to obtain optimal results, it is important to make use of all available data, even though it involves additional procedures such as data imputation.

Keywords: demand estimate, missing-data imputation, real estate, unsupervised learning

Procedia PDF Downloads 280
31595 Predicting Groundwater Areas Using Data Mining Techniques: Groundwater in Jordan as Case Study

Authors: Faisal Aburub, Wael Hadi

Abstract:

Data mining is the process of extracting useful or hidden information from a large database. Extracted information can be used to discover relationships among features, where data objects are grouped according to logical relationships; or to predict unseen objects to one of the predefined groups. In this paper, we aim to investigate four well-known data mining algorithms in order to predict groundwater areas in Jordan. These algorithms are Support Vector Machines (SVMs), Naïve Bayes (NB), K-Nearest Neighbor (kNN) and Classification Based on Association Rule (CBA). The experimental results indicate that the SVMs algorithm outperformed other algorithms in terms of classification accuracy, precision and F1 evaluation measures using the datasets of groundwater areas that were collected from Jordanian Ministry of Water and Irrigation.

Keywords: classification, data mining, evaluation measures, groundwater

Procedia PDF Downloads 273
31594 An Optimization Model for the Arrangement of Assembly Areas Considering Time Dynamic Area Requirements

Authors: Michael Zenker, Henrik Prinzhorn, Christian Böning, Tom Strating

Abstract:

Large-scale products are often assembled according to the job-site principle, meaning that during the assembly the product is located at a fixed position, while the area requirements are constantly changing. On one hand, the product itself is growing with each assembly step, whereas varying areas for storage, machines or working areas are temporarily required. This is an important factor when arranging products to be assembled within the factory. Currently, it is common to reserve a fixed area for each product to avoid overlaps or collisions with the other assemblies. Intending to be large enough to include the product and all adjacent areas, this reserved area corresponds to the superposition of the maximum extents of all required areas of the product. In this procedure, the reserved area is usually poorly utilized over the course of the entire assembly process; instead a large part of it remains unused. If the available area is a limited resource, a systematic arrangement of the products, which complies with the dynamic area requirements, will lead to an increased area utilization and productivity. This paper presents the results of a study on the arrangement of assembly objects assuming dynamic, competing area requirements. First, the problem situation is extensively explained, and existing research on associated topics is described and evaluated on the possibility of an adaptation. Then, a newly developed mathematical optimization model is introduced. This model allows an optimal arrangement of dynamic areas, considering logical and practical constraints. Finally, in order to quantify the potential of the developed method, some test series results are presented, showing the possible increase in area utilization.

Keywords: dynamic area requirements, facility layout problem, optimization model, product assembly

Procedia PDF Downloads 224
31593 A Review on Big Data Movement with Different Approaches

Authors: Nay Myo Sandar

Abstract:

With the growth of technologies and applications, a large amount of data has been producing at increasing rate from various resources such as social media networks, sensor devices, and other information serving devices. This large collection of massive, complex and exponential growth of dataset is called big data. The traditional database systems cannot store and process such data due to large and complexity. Consequently, cloud computing is a potential solution for data storage and processing since it can provide a pool of resources for servers and storage. However, moving large amount of data to and from is a challenging issue since it can encounter a high latency due to large data size. With respect to big data movement problem, this paper reviews the literature of previous works, discusses about research issues, finds out approaches for dealing with big data movement problem.

Keywords: Big Data, Cloud Computing, Big Data Movement, Network Techniques

Procedia PDF Downloads 78
31592 Enhancing Large Language Models' Data Analysis Capability with Planning-and-Execution and Code Generation Agents: A Use Case for Southeast Asia Real Estate Market Analytics

Authors: Kien Vu, Jien Min Soh, Mohamed Jahangir Abubacker, Piyawut Pattamanon, Soojin Lee, Suvro Banerjee

Abstract:

Recent advances in Generative Artificial Intelligence (GenAI), in particular Large Language Models (LLMs) have shown promise to disrupt multiple industries at scale. However, LLMs also present unique challenges, notably, these so-called "hallucination" which is the generation of outputs that are not grounded in the input data that hinders its adoption into production. Common practice to mitigate hallucination problem is utilizing Retrieval Agmented Generation (RAG) system to ground LLMs'response to ground truth. RAG converts the grounding documents into embeddings, retrieve the relevant parts with vector similarity between user's query and documents, then generates a response that is not only based on its pre-trained knowledge but also on the specific information from the retrieved documents. However, the RAG system is not suitable for tabular data and subsequent data analysis tasks due to multiple reasons such as information loss, data format, and retrieval mechanism. In this study, we have explored a novel methodology that combines planning-and-execution and code generation agents to enhance LLMs' data analysis capabilities. The approach enables LLMs to autonomously dissect a complex analytical task into simpler sub-tasks and requirements, then convert them into executable segments of code. In the final step, it generates the complete response from output of the executed code. When deployed beta version on DataSense, the property insight tool of PropertyGuru, the approach yielded promising results, as it was able to provide market insights and data visualization needs with high accuracy and extensive coverage by abstracting the complexities for real-estate agents and developers from non-programming background. In essence, the methodology not only refines the analytical process but also serves as a strategic tool for real estate professionals, aiding in market understanding and enhancement without the need for programming skills. The implication extends beyond immediate analytics, paving the way for a new era in the real estate industry characterized by efficiency and advanced data utilization.

Keywords: large language model, reasoning, planning and execution, code generation, natural language processing, prompt engineering, data analysis, real estate, data sense, PropertyGuru

Procedia PDF Downloads 80
31591 Knowledge Based Automated Software Engineering Platform Used for the Development of Bulgarian E-Customs

Authors: Ivan Stanev, Maria Koleva

Abstract:

Described are challenges to the Bulgarian e-Customs (BeC) related to low level of interoperability and standardization, inefficient use of available infrastructure, lack of centralized identification and authorization, extremely low level of software process automation, and insufficient quality of data stored in official registers. The technical requirements for BeC are prepared with a focus on domain independent common platform, specialized customs and excise components, high scalability, flexibility, and reusability. The Knowledge Based Automated Software Engineering (KBASE) Common Platform for Automated Programming (CPAP) is selected as an instrument covering BeC requirements for standardization, programming automation, knowledge interpretation and cloud computing. BeC stage 3 results are presented and analyzed. BeC.S3 development trends are identified.

Keywords: service oriented architecture, cloud computing, knowledge based automated software engineering, common platform for automated programming, e-customs

Procedia PDF Downloads 366
31590 Development of Database for Risk Assessment Appling to Ballast Water Managements

Authors: Eun-Chan Kim, Jeong-Hwan Oh, Seung-Guk Lee

Abstract:

Billions of tones of ballast water including various aquatic organisms are being carried around the world by ships. When the ballast water is discharged into new environments, some aquatic organisms discharged with ballast water may become invasive and severely disrupt the native ecology. Thus, International Maritime Organization (IMO) adopted the Ballast Water Management Convention in 2004. Regulation A-4 of the convention states that a government in waters under their jurisdiction may grant exemptions to any requirements to ballast water management, but only when they are granted to a ship or ships on a voyage or voyages between specified ports or locations, or to a ship which operates exclusively between specified ports or locations. In order to grant exemptions, risk assessment should be conducted based on the guidelines for risk assessment developed by the IMO. For the risk assessment, it is essential to collect the relevant information and establish a database system. This paper studies the database system for ballast water risk assessment. This database consists of the shipping database, ballast water database, port environment database and species database. The shipping database has been established based on the data collected from the port management information system of Korea Government. For the ballast water database, ballast water discharge has only been estimated by the loading/unloading of the cargoes as the convention has not come into effect yet. The port environment database and species database are being established based on the reference documents, and existing and newly collected monitoring data. This database system has been approved to be a useful system, capable of appropriately analyzing the risk assessment in the all ports of Korea.

Keywords: ballast water, IMO, risk assessment, shipping, environment, species

Procedia PDF Downloads 515
31589 BigCrypt: A Probable Approach of Big Data Encryption to Protect Personal and Business Privacy

Authors: Abdullah Al Mamun, Talal Alkharobi

Abstract:

As data size is growing up, people are became more familiar to store big amount of secret information into cloud storage. Companies are always required to need transfer massive business files from one end to another. We are going to lose privacy if we transmit it as it is and continuing same scenario repeatedly without securing the communication mechanism means proper encryption. Although asymmetric key encryption solves the main problem of symmetric key encryption but it can only encrypt limited size of data which is inapplicable for large data encryption. In this paper we propose a probable approach of pretty good privacy for encrypt big data using both symmetric and asymmetric keys. Our goal is to achieve encrypt huge collection information and transmit it through a secure communication channel for committing the business and personal privacy. To justify our method an experimental dataset from three different platform is provided. We would like to show that our approach is working for massive size of various data efficiently and reliably.

Keywords: big data, cloud computing, cryptography, hadoop, public key

Procedia PDF Downloads 314
31588 Analysis of Secondary School Students' Perceptions about Information Technologies through a Word Association Test

Authors: Fetah Eren, Ismail Sahin, Ismail Celik, Ahmet Oguz Akturk

Abstract:

The aim of this study is to discover secondary school students’ perceptions related to information technologies and the connections between concepts in their cognitive structures. A word association test consisting of six concepts related to information technologies is used to collect data from 244 secondary school students. Concept maps that present students’ cognitive structures are drawn with the help of frequency data. Data are analyzed and interpreted according to the connections obtained as a result of the concept maps. It is determined students associate most with these concepts—computer, Internet, and communication of the given concepts, and associate least with these concepts—computer-assisted education and information technologies. These results show the concepts, Internet, communication, and computer, are an important part of students’ cognitive structures. In addition, students mostly answer computer, phone, game, Internet and Facebook as the key concepts. These answers show students regard information technologies as a means for entertainment and free time activity, not as a means for education.

Keywords: word association test, cognitive structure, information technology, secondary school

Procedia PDF Downloads 408
31587 Novel Recommender Systems Using Hybrid CF and Social Network Information

Authors: Kyoung-Jae Kim

Abstract:

Collaborative Filtering (CF) is a popular technique for the personalization in the E-commerce domain to reduce information overload. In general, CF provides recommending items list based on other similar users’ preferences from the user-item matrix and predicts the focal user’s preference for particular items by using them. Many recommender systems in real-world use CF techniques because it’s excellent accuracy and robustness. However, it has some limitations including sparsity problems and complex dimensionality in a user-item matrix. In addition, traditional CF does not consider the emotional interaction between users. In this study, we propose recommender systems using social network and singular value decomposition (SVD) to alleviate some limitations. The purpose of this study is to reduce the dimensionality of data set using SVD and to improve the performance of CF by using emotional information from social network data of the focal user. In this study, we test the usability of hybrid CF, SVD and social network information model using the real-world data. The experimental results show that the proposed model outperforms conventional CF models.

Keywords: recommender systems, collaborative filtering, social network information, singular value decomposition

Procedia PDF Downloads 282
31586 The Impact of the Information Technologies on the Accounting Department of the Romanian Companies

Authors: Dumitru Valentin Florentin

Abstract:

The need to use high volumes of data and the high competition are only two reasons which make necessary the use of information technologies. The objective of our research is to establish the impact of information technologies on the accounting department of the Romanian companies. In order to achieve it, starting from the literature review we made an empirical research based on a questionnaire. We investigated the types of technologies used, the reasons which led to the implementation of certain technologies, the benefits brought by the use of the information technologies, the difficulties brought by the implementation and the future effects of the applications. The conclusions show that there is an evolution in the degree of implementation of the information technologies in the Romanian companies, compared with the results of other studies conducted a few years before.

Keywords: information technologies, impact, company, Romania, empirical study

Procedia PDF Downloads 419
31585 Legal Means for Access to Information Management

Authors: Sameut Bouhaik Mostafa

Abstract:

Information Act is the Canadian law gives the right of access to information for the institution of government. It declares the availability of government information to the public, but that exceptions should be limited and the necessary right of access to be specific, and also states the need to constantly re-examine the decisions on the disclosure of any government information independently from the government. By 1982, it enacted a dozen countries, including France, Denmark, Finland, Sweden, the Netherlands and the United States (1966) newly legally to access the information. It entered access to Canadian information into force of the Act of 1983, under the government of Pierre Trudeau, allowing Canadians to recover information from government files, and the development of what can be accessed from the information, and the imposition of timetables to respond. It has been applied by the Information Commissioner in Canada.

Keywords: law, information, management, legal

Procedia PDF Downloads 409
31584 Sourcing and Compiling a Maltese Traffic Dataset MalTra

Authors: Gabriele Borg, Alexei De Bono, Charlie Abela

Abstract:

There on a constant rise in the availability of high volumes of data gathered from multiple sources, resulting in an abundance of unprocessed information that can be used to monitor patterns and trends in user behaviour. Similarly, year after year, Malta is also constantly experiencing ongoing population growth and an increase in mobilization demand. This research takes advantage of data which is continuously being sourced and converting it into useful information related to the traffic problem on the Maltese roads. The scope of this paper is to provide a methodology to create a custom dataset (MalTra - Malta Traffic) compiled from multiple participants from various locations across the island to identify the most common routes taken to expose the main areas of activity. This use of big data is seen being used in various technologies and is referred to as ITSs (Intelligent Transportation Systems), which has been concluded that there is significant potential in utilising such sources of data on a nationwide scale.

Keywords: Big Data, vehicular traffic, traffic management, mobile data patterns

Procedia PDF Downloads 103
31583 An As-Is Analysis and Approach for Updating Building Information Models and Laser Scans

Authors: Rene Hellmuth

Abstract:

Factory planning has the task of designing products, plants, processes, organization, areas, and the construction of a factory. The requirements for factory planning and the building of a factory have changed in recent years. Regular restructuring of the factory building is becoming more important in order to maintain the competitiveness of a factory. Restrictions in new areas, shorter life cycles of product and production technology as well as a VUCA world (Volatility, Uncertainty, Complexity & Ambiguity) lead to more frequent restructuring measures within a factory. A building information model (BIM) is the planning basis for rebuilding measures and becomes an indispensable data repository to be able to react quickly to changes. Use as a planning basis for restructuring measures in factories only succeeds if the BIM model has adequate data quality. Under this aspect and the industrial requirement, three data quality factors are particularly important for this paper regarding the BIM model: up-to-dateness, completeness, and correctness. The research question is: how can a BIM model be kept up to date with required data quality and which visualization techniques can be applied in a short period of time on the construction site during conversion measures? An as-is analysis is made of how BIM models and digital factory models (including laser scans) are currently being kept up to date. Industrial companies are interviewed, and expert interviews are conducted. Subsequently, the results are evaluated, and a procedure conceived how cost-effective and timesaving updating processes can be carried out. The availability of low-cost hardware and the simplicity of the process are of importance to enable service personnel from facility mnagement to keep digital factory models (BIM models and laser scans) up to date. The approach includes the detection of changes to the building, the recording of the changing area, and the insertion into the overall digital twin. Finally, an overview of the possibilities for visualizations suitable for construction sites is compiled. An augmented reality application is created based on an updated BIM model of a factory and installed on a tablet. Conversion scenarios with costs and time expenditure are displayed. A user interface is designed in such a way that all relevant conversion information is available at a glance for the respective conversion scenario. A total of three essential research results are achieved: As-is analysis of current update processes for BIM models and laser scans, development of a time-saving and cost-effective update process and the conception and implementation of an augmented reality solution for BIM models suitable for construction sites.

Keywords: building information modeling, digital factory model, factory planning, restructuring

Procedia PDF Downloads 108
31582 Modeling Methodologies for Optimization and Decision Support on Coastal Transport Information System (Co.Tr.I.S.)

Authors: Vassilios Moussas, Dimos N. Pantazis, Panagioths Stratakis

Abstract:

The aim of this paper is to present the optimization methodology developed in the frame of a Coastal Transport Information System. The system will be used for the effective design of coastal transportation lines and incorporates subsystems that implement models, tools and techniques that may support the design of improved networks. The role of the optimization and decision subsystem is to provide the user with better and optimal scenarios that will best fulfill any constrains, goals or requirements posed. The complexity of the problem and the large number of parameters and objectives involved led to the adoption of an evolutionary method (Genetic Algorithms). The problem model and the subsystem structure are presented in detail, and, its support for simulation is also discussed.

Keywords: coastal transport, modeling, optimization

Procedia PDF Downloads 495
31581 Workflow Based Inspection of Geometrical Adaptability from 3D CAD Models Considering Production Requirements

Authors: Tobias Huwer, Thomas Bobek, Gunter Spöcker

Abstract:

Driving forces for enhancements in production are trends like digitalization and individualized production. Currently, such developments are restricted to assembly parts. Thus, complex freeform surfaces are not addressed in this context. The need for efficient use of resources and near-net-shape production will require individualized production of complex shaped workpieces. Due to variations between nominal model and actual geometry, this can lead to changes in operations in Computer-aided process planning (CAPP) to make CAPP manageable for an adaptive serial production. In this context, 3D CAD data can be a key to realizing that objective. Along with developments in the geometrical adaptation, a preceding inspection method based on CAD data is required to support the process planner by finding objective criteria to make decisions about the adaptive manufacturability of workpieces. Nowadays, this kind of decisions is depending on the experience-based knowledge of humans (e.g. process planners) and results in subjective decisions – leading to a variability of workpiece quality and potential failure in production. In this paper, we present an automatic part inspection method, based on design and measurement data, which evaluates actual geometries of single workpiece preforms. The aim is to automatically determine the suitability of the current shape for further machining, and to provide a basis for an objective decision about subsequent adaptive manufacturability. The proposed method is realized by a workflow-based approach, keeping in mind the requirements of industrial applications. Workflows are a well-known design method of standardized processes. Especially in applications like aerospace industry standardization and certification of processes are an important aspect. Function blocks, providing a standardized, event-driven abstraction to algorithms and data exchange, will be used for modeling and execution of inspection workflows. Each analysis step of the inspection, such as positioning of measurement data or checking of geometrical criteria, will be carried out by function blocks. One advantage of this approach is its flexibility to design workflows and to adapt algorithms specific to the application domain. In general, within the specified tolerance range it will be checked if a geometrical adaption is possible. The development of particular function blocks is predicated on workpiece specific information e.g. design data. Furthermore, for different product lifecycle phases, appropriate logics and decision criteria have to be considered. For example, tolerances for geometric deviations are different in type and size for new-part production compared to repair processes. In addition to function blocks, appropriate referencing systems are important. They need to support exact determination of position and orientation of the actual geometries to provide a basis for precise analysis. The presented approach provides an inspection methodology for adaptive and part-individual process chains. The analysis of each workpiece results in an inspection protocol and an objective decision about further manufacturability. A representative application domain is the product lifecycle of turbine blades containing a new-part production and a maintenance process. In both cases, a geometrical adaptation is required to calculate individual production data. In contrast to existing approaches, the proposed initial inspection method provides information to decide between different potential adaptive machining processes.

Keywords: adaptive, CAx, function blocks, turbomachinery

Procedia PDF Downloads 294