Search results for: cloud computing framework
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 6338

Search results for: cloud computing framework

5828 Learning to Teach on the Cloud: Preservice EFL Teachers’ Online Project-Based Practicum Experience

Authors: Mei-Hui Liu

Abstract:

This paper reports 20 preservice EFL teachers’ learning-to-teach experience when they were engaged in an online project-based practicum implemented on a Cloud Platform. This 10-month study filled in the literature gap by documenting the impact of online project-based instruction on preservice EFL teachers’ professional development. Data analysis showed that the online practicum was regarded as a flexible mechanism offering chances of teaching practices without geographical barriers. Additionally, this project-based practice helped the participants integrate the theories they had learned and further foster them how to create a self-directed online learning environment. Furthermore, these preservice teachers with experiences of technology-enabled practicum showed their motivation to apply technology and online platforms into future instructional practices. Yet, this study uncovered several concerns encountered by these participants during this online field experience. The findings of this study rendered meaning and lessons for teacher educators intending to integrate online practicum into preservice training courses.

Keywords: online teaching practicum, project-based learning, teacher preparation, English language education

Procedia PDF Downloads 371
5827 Project Management Framework and Influencing Factors

Authors: Mehrnoosh Askarizadeh

Abstract:

The increasing variations of the business world correspond with a high diversity of theoretical perspectives used in project management research. This diversity is reflected by a variety of influencing factors, which have been the subject of empirical studies. This article aims to systemize the different streams of research on the basis of a literature review and at developing a research framework influencing factors. We will identify fundamental elements of a project management theory. The framework consists of three dimensions: design, context, and goal. Its purpose is to support the combination of different perspectives and the development of strategies for further research.

Keywords: project, goal, project management, influencing factors

Procedia PDF Downloads 543
5826 A Cloud-Based Federated Identity Management in Europe

Authors: Jesus Carretero, Mario Vasile, Guillermo Izquierdo, Javier Garcia-Blas

Abstract:

Currently, there is a so called ‘identity crisis’ in cybersecurity caused by the substantial security, privacy and usability shortcomings encountered in existing systems for identity management. Federated Identity Management (FIM) could be solution for this crisis, as it is a method that facilitates management of identity processes and policies among collaborating entities without enforcing a global consistency, that is difficult to achieve when there are ID legacy systems. To cope with this problem, the Connecting Europe Facility (CEF) initiative proposed in 2014 a federated solution in anticipation of the adoption of the Regulation (EU) N°910/2014, the so-called eIDAS Regulation. At present, a network of eIDAS Nodes is being deployed at European level to allow that every citizen recognized by a member state is to be recognized within the trust network at European level, enabling the consumption of services in other member states that, until now were not allowed, or whose concession was tedious. This is a very ambitious approach, since it tends to enable cross-border authentication of Member States citizens without the need to unify the authentication method (eID Scheme) of the member state in question. However, this federation is currently managed by member states and it is initially applied only to citizens and public organizations. The goal of this paper is to present the results of a European Project, named eID@Cloud, that focuses on the integration of eID in 5 cloud platforms belonging to authentication service providers of different EU Member States to act as Service Providers (SP) for private entities. We propose an initiative based on a private eID Scheme both for natural and legal persons. The methodology followed in the eID@Cloud project is that each Identity Provider (IdP) is subscribed to an eIDAS Node Connector, requesting for authentication, that is subscribed to an eIDAS Node Proxy Service, issuing authentication assertions. To cope with high loads, load balancing is supported in the eIDAS Node. The eID@Cloud project is still going on, but we already have some important outcomes. First, we have deployed the federation identity nodes and tested it from the security and performance point of view. The pilot prototype has shown the feasibility of deploying this kind of systems, ensuring good performance due to the replication of the eIDAS nodes and the load balance mechanism. Second, our solution avoids the propagation of identity data out of the native domain of the user or entity being identified, which avoids problems well known in cybersecurity due to network interception, man in the middle attack, etc. Last, but not least, this system allows to connect any country or collectivity easily, providing incremental development of the network and avoiding difficult political negotiations to agree on a single authentication format (which would be a major stopper).

Keywords: cybersecurity, identity federation, trust, user authentication

Procedia PDF Downloads 166
5825 Addressing Supply Chain Data Risk with Data Security Assurance

Authors: Anna Fowler

Abstract:

When considering assets that may need protection, the mind begins to contemplate homes, cars, and investment funds. In most cases, the protection of those assets can be covered through security systems and insurance. Data is not the first thought that comes to mind that would need protection, even though data is at the core of most supply chain operations. It includes trade secrets, management of personal identifiable information (PII), and consumer data that can be used to enhance the overall experience. Data is considered a critical element of success for supply chains and should be one of the most critical areas to protect. In the supply chain industry, there are two major misconceptions about protecting data: (i) We do not manage or store confidential/personally identifiable information (PII). (ii) Reliance on Third-Party vendor security. These misconceptions can significantly derail organizational efforts to adequately protect data across environments. These statistics can be exciting yet overwhelming at the same time. The first misconception, “We do not manage or store confidential/personally identifiable information (PII)” is dangerous as it implies the organization does not have proper data literacy. Enterprise employees will zero in on the aspect of PII while neglecting trade secret theft and the complete breakdown of information sharing. To circumvent the first bullet point, the second bullet point forges an ideology that “Reliance on Third-Party vendor security” will absolve the company from security risk. Instead, third-party risk has grown over the last two years and is one of the major causes of data security breaches. It is important to understand that a holistic approach should be considered when protecting data which should not involve purchasing a Data Loss Prevention (DLP) tool. A tool is not a solution. To protect supply chain data, start by providing data literacy training to all employees and negotiating the security component of contracts with vendors to highlight data literacy training for individuals/teams that may access company data. It is also important to understand the origin of the data and its movement to include risk identification. Ensure processes effectively incorporate data security principles. Evaluate and select DLP solutions to address specific concerns/use cases in conjunction with data visibility. These approaches are part of a broader solutions framework called Data Security Assurance (DSA). The DSA Framework looks at all of the processes across the supply chain, including their corresponding architecture and workflows, employee data literacy, governance and controls, integration between third and fourth-party vendors, DLP as a solution concept, and policies related to data residency. Within cloud environments, this framework is crucial for the supply chain industry to avoid regulatory implications and third/fourth party risk.

Keywords: security by design, data security architecture, cybersecurity framework, data security assurance

Procedia PDF Downloads 88
5824 Terrestrial Laser Scans to Assess Aerial LiDAR Data

Authors: J. F. Reinoso-Gordo, F. J. Ariza-López, A. Mozas-Calvache, J. L. García-Balboa, S. Eddargani

Abstract:

The DEMs quality may depend on several factors such as data source, capture method, processing type used to derive them, or the cell size of the DEM. The two most important capture methods to produce regional-sized DEMs are photogrammetry and LiDAR; DEMs covering entire countries have been obtained with these methods. The quality of these DEMs has traditionally been evaluated by the national cartographic agencies through punctual sampling that focused on its vertical component. For this type of evaluation there are standards such as NMAS and ASPRS Positional Accuracy Standards for Digital Geospatial Data. However, it seems more appropriate to carry out this evaluation by means of a method that takes into account the superficial nature of the DEM and, therefore, its sampling is superficial and not punctual. This work is part of the Research Project "Functional Quality of Digital Elevation Models in Engineering" where it is necessary to control the quality of a DEM whose data source is an experimental LiDAR flight with a density of 14 points per square meter to which we call Point Cloud Product (PCpro). In the present work it is described the capture data on the ground and the postprocessing tasks until getting the point cloud that will be used as reference (PCref) to evaluate the PCpro quality. Each PCref consists of a patch 50x50 m size coming from a registration of 4 different scan stations. The area studied was the Spanish region of Navarra that covers an area of 10,391 km2; 30 patches homogeneously distributed were necessary to sample the entire surface. The patches have been captured using a Leica BLK360 terrestrial laser scanner mounted on a pole that reached heights of up to 7 meters; the position of the scanner was inverted so that the characteristic shadow circle does not exist when the scanner is in direct position. To ensure that the accuracy of the PCref is greater than that of the PCpro, the georeferencing of the PCref has been carried out with real-time GNSS, and its accuracy positioning was better than 4 cm; this accuracy is much better than the altimetric mean square error estimated for the PCpro (<15 cm); The kind of DEM of interest is the corresponding to the bare earth, so that it was necessary to apply a filter to eliminate vegetation and auxiliary elements such as poles, tripods, etc. After the postprocessing tasks the PCref is ready to be compared with the PCpro using different techniques: cloud to cloud or after a resampling process DEM to DEM.

Keywords: data quality, DEM, LiDAR, terrestrial laser scanner, accuracy

Procedia PDF Downloads 100
5823 Design of a Professional Development Framework in Teaching and Learning for Engineering Educators

Authors: Orla McConnell, Cormac MacMahon, Jen Harvey

Abstract:

Ireland’s national professional development framework for those who teach in higher education, aims to provide guidance and leadership in the planning, developing and engaging in professional development practices. A series of pilot projects have been initiated to help explore the framework’s likely utility and acceptance by educators and their institutions. These projects require engagement with staff in the interpretation and adaption of the framework within their working contexts. The purpose of this paper is to outline the development of one such project with engineering educators at three Institutes of Technology seeking designation as a technological university. The initiative aims to gain traction in the acceptance of the framework with the engineering education community by linking core and discipline-specific teaching and learning competencies with professional development activities most valued by engineering educators. Informed by three strands of literature: professional development in higher education; engineering education; and teaching and learning training provisions, the project begins with a survey of all those involved in teaching and learning in engineering across the three institutes. Based on engagement with key stakeholders, subsequent qualitative research informs the contextualization of the national framework for discipline-specific and institutional piloting. The paper concludes by exploring engineering educator perceptions of the national framework’s utility based on their engagement with the pilot process. Feedback from the pilot indicates that there is a significant gap between the professional development needs of engineering educators and the current professional development provision in teaching and learning.

Keywords: engineering education, pilot, professional development, teaching and learning

Procedia PDF Downloads 329
5822 Regression Approach for Optimal Purchase of Hosts Cluster in Fixed Fund for Hadoop Big Data Platform

Authors: Haitao Yang, Jianming Lv, Fei Xu, Xintong Wang, Yilin Huang, Lanting Xia, Xuewu Zhu

Abstract:

Given a fixed fund, purchasing fewer hosts of higher capability or inversely more of lower capability is a must-be-made trade-off in practices for building a Hadoop big data platform. An exploratory study is presented for a Housing Big Data Platform project (HBDP), where typical big data computing is with SQL queries of aggregate, join, and space-time condition selections executed upon massive data from more than 10 million housing units. In HBDP, an empirical formula was introduced to predict the performance of host clusters potential for the intended typical big data computing, and it was shaped via a regression approach. With this empirical formula, it is easy to suggest an optimal cluster configuration. The investigation was based on a typical Hadoop computing ecosystem HDFS+Hive+Spark. A proper metric was raised to measure the performance of Hadoop clusters in HBDP, which was tested and compared with its predicted counterpart, on executing three kinds of typical SQL query tasks. Tests were conducted with respect to factors of CPU benchmark, memory size, virtual host division, and the number of element physical host in cluster. The research has been applied to practical cluster procurement for housing big data computing.

Keywords: Hadoop platform planning, optimal cluster scheme at fixed-fund, performance predicting formula, typical SQL query tasks

Procedia PDF Downloads 232
5821 Motivational Interviewing as a Framework for Coaching Physicians through ACGME Milestones

Authors: Michael Olson

Abstract:

The Accreditation Council for Graduate Medical Education (ACGME) in the U.S. has established core competencies and milestones for family physicians in residency training programs. These competencies are intended to guide preceptors as they work with physician trainees toward independent practice. This conceptual paper describes a framework for coaching trainees toward these milestones using motivational interviewing as an evidence-based approach. The main objective of applying the motivational interviewing framework to the residency training setting is to facilitate clinical behavior change that meets higher level competencies/rubric. This is a work in progress and there is no manuscript/paper prepared to date. A conceptual paper/framework will be completed by the conference deadline. This is based on a separate but related development of work we have completed and published elsewhere.

Keywords: coaching, motivational interviewing, physicians, competencies

Procedia PDF Downloads 175
5820 mm-Wave Wearable Edge Computing Module Hosted by Printed Ridge Gap Waveguide Structures: A Physical Layer Study

Authors: Matthew Kostawich, Mohammed Elmorsy, Mohamed Sayed Sifat, Shoukry Shams, Mahmoud Elsaadany

Abstract:

6G communication systems represent the nominal future extension of current wireless technology, where its impact is extended to touch upon all human activities, including medical, security, and entertainment applications. As a result, human needs are allocated among the highest priority aspects of the system design and requirements. 6G communications is expected to replace all the current video conferencing with interactive virtual reality meetings involving high data-rate transmission merged with massive distributed computing resources. In addition, the current expansion of IoT applications must be mitigated with significant network changes to provide a reasonable Quality of Service (QoS). This directly implies a high demand for Human-Computer Interaction (HCI) through mobile computing modules in future wireless communication systems. This article proposes the utilization of a Printed Ridge Gap Waveguide (PRGW) to host the wearable nodes. To the best of our knowledge, we propose for the first time a physical layer analysis within the context of a complete architecture. A thorough study is provided on the impact of the distortion of the guiding structure on the overall system performance. The proposed structure shows small latency and small losses, highlighting its compatibility with future applications.

Keywords: ridge gap waveguide, edge computing module, 6G, multimedia IoT applications

Procedia PDF Downloads 71
5819 An Exploratory Study Applied to Search Relationship between Humans and Universe

Authors: Mohamed Hashelaf, Ahmed Al-Osdody

Abstract:

In this paper, we focused our efforts on one of the vaguest subjects in astrophysics that is the formation and evolution of the universe until the arrival of humans. Through an in-depth exploration of the origins of the universe, understanding what has happened since the Big Bang until now and checking the history of creation, we can answer questions about the future of life, the possibility of its existence elsewhere in the universe and to be able to understand how we came, what our role in the circle of life is and what the future of our development will be. Here is where we used systematic steps that allowed us first and foremost to identify the reason behind the big bang itself that formed a large cloud of cosmic dust. Then after a period of time from the expansion of the universe and its coolness, the initial molecules of gases from the cosmic cloud began to condense, forming a very dense field of gravity that after millions of years led to the formation of stars, galaxies, even earth and the else planets. Finally, it became clear before us that after the earth has formed, the existence of liquid water made it possible for life to form, starting from the bacteria all the way until the appearance of the humans that we know today. But it does not stop here. If we look and contemplate in ourselves as humans, we will understand that the universe is inside us and that’s what makes us exceptional. All of this means that just as life on earth was created, it could have been on other planets as well. It also means that we are the universe’s key to understand itself.

Keywords: Big Bang, cosmic dust, primary elements, universe

Procedia PDF Downloads 134
5818 Employers' Occupational Health and Safety Training Obligations in Framework Directive and Training Procedure and Rules in Turkey

Authors: Nuray Gökçek Karaca, Berrin Gökçek

Abstract:

Employers occupational safety and health training obligations are regulated in 89/391/EEC Framework Directive and also in 6331 numbered Occupational Health and Safety Law in Turkey. The main objective of this research is to determine and evaluate the employers’ occupational health and safety training obligations in Framework Directive in comparison with the 6331 numbered Occupational Health and Safety Law and to examine training principles in Turkey. For this purpose, employers’ occupational health and safety training obligations examined in Framework Directive and Occupational Health and Safety Law. This study carried out through comparative scanning model and literature model. The research data were collected through European Agency and ministry legislations. As a result, employers’ occupational health and safety training obligations in the 6331 numbered Occupational Health and Safety Law are compatible with the 89/391/EEC numbered Framework Directive and training principles are determined by in different ways like the trained workers, training issues, training period, training time, and trainers. In this study, employers’ training obligations are evaluated in detail.

Keywords: directive, occupational health and safety, training, work accidences

Procedia PDF Downloads 344
5817 Quality Standards for Emergency Response: A Methodological Framework

Authors: Jennifer E. Lynette

Abstract:

This study describes the development process of a methodological framework for quality standards used to measure the efficiency and quality of response efforts of trained personnel at emergency events. This paper describes the techniques used to develop the initial framework and its potential application to professions under the broader field of emergency management. The example described in detail in this paper applies the framework specifically to fire response activities by firefighters. Within the quality standards framework, the fire response process is chronologically mapped. Individual variables within the sequence of events are identified. Through in-person data collection, questionnaires, interviews, and the expansion of the incident reporting system, this study identifies and categorizes previously unrecorded variables involved in the response phase of a fire. Following a data analysis of each variable using a quantitative or qualitative assessment, the variables are ranked pertaining to the magnitude of their impact to the event outcome. Among others, key indicators of quality performance in the analysis involve decision communication, resource utilization, response techniques, and response time. Through the application of this framework and subsequent utilization of quality standards indicators, there is potential to increase efficiency in the response phase of an emergency event; thereby saving additional lives, property, and resources.

Keywords: emergency management, fire, quality standards, response

Procedia PDF Downloads 318
5816 Artificial Intelligence and Distributed System Computing: Application and Practice in Real Life

Authors: Lai Junzhe, Wang Lihao, Burra Venkata Durga Kumar

Abstract:

In recent years, due to today's global technological advances, big data and artificial intelligence technologies have been widely used in various industries and fields, playing an important role in reducing costs and increasing efficiency. Among them, artificial intelligence has derived another branch in its own continuous progress and the continuous development of computer personnel, namely distributed artificial intelligence computing systems. Distributed AI is a method for solving complex learning, decision-making, and planning problems, characterized by the ability to take advantage of large-scale computation and the spatial distribution of resources, and accordingly, it can handle problems with large data sets. Nowadays, distributed AI is widely used in military, medical, and human daily life and brings great convenience and efficient operation to life. In this paper, we will discuss three areas of distributed AI computing systems in vision processing, blockchain, and smart home to introduce the performance of distributed systems and the role of AI in distributed systems.

Keywords: distributed system, artificial intelligence, blockchain, IoT, visual information processing, smart home

Procedia PDF Downloads 113
5815 Image Encryption Using Eureqa to Generate an Automated Mathematical Key

Authors: Halima Adel Halim Shnishah, David Mulvaney

Abstract:

Applying traditional symmetric cryptography algorithms while computing encryption and decryption provides immunity to secret keys against different attacks. One of the popular techniques generating automated secret keys is evolutionary computing by using Eureqa API tool, which got attention in 2013. In this paper, we are generating automated secret keys for image encryption and decryption using Eureqa API (tool which is used in evolutionary computing technique). Eureqa API models pseudo-random input data obtained from a suitable source to generate secret keys. The validation of generated secret keys is investigated by performing various statistical tests (histogram, chi-square, correlation of two adjacent pixels, correlation between original and encrypted images, entropy and key sensitivity). Experimental results obtained from methods including histogram analysis, correlation coefficient, entropy and key sensitivity, show that the proposed image encryption algorithms are secure and reliable, with the potential to be adapted for secure image communication applications.

Keywords: image encryption algorithms, Eureqa, statistical measurements, automated key generation

Procedia PDF Downloads 482
5814 Intelligent Computing with Bayesian Regularization Artificial Neural Networks for a Nonlinear System of COVID-19 Epidemic Model for Future Generation Disease Control

Authors: Tahir Nawaz Cheema, Dumitru Baleanu, Ali Raza

Abstract:

In this research work, we design intelligent computing through Bayesian Regularization artificial neural networks (BRANNs) introduced to solve the mathematical modeling of infectious diseases (Covid-19). The dynamical transmission is due to the interaction of people and its mathematical representation based on the system's nonlinear differential equations. The generation of the dataset of the Covid-19 model is exploited by the power of the explicit Runge Kutta method for different countries of the world like India, Pakistan, Italy, and many more. The generated dataset is approximately used for training, testing, and validation processes for every frequent update in Bayesian Regularization backpropagation for numerical behavior of the dynamics of the Covid-19 model. The performance and effectiveness of designed methodology BRANNs are checked through mean squared error, error histograms, numerical solutions, absolute error, and regression analysis.

Keywords: mathematical models, beysian regularization, bayesian-regularization backpropagation networks, regression analysis, numerical computing

Procedia PDF Downloads 147
5813 Developing a Group Guidance Framework: A Review of Literature

Authors: Abdul Rawuf Hussein, Rusnani Abdul Kadir, Mona Adlina Binti Adanan

Abstract:

Guidance program has been an essential approach in helping professions from many institutions of learning as well as communities, organizations, and clinical settings. Although the term varies depending on the approaches, objectives, and theories, the core and central element is typically developmental in nature. In this conceptual paper, the researcher will review literature on the concept of group guidance, its impact on students’ and individual’s development, developing a guidance module and proposing a synthesised framework for group guidance program.

Keywords: concept, framework, group guidance, module development

Procedia PDF Downloads 528
5812 Exploring the Quest for Centralized Identity in Mohsin Hamid's "The Last White Man": Post-Apocalyptic Transformations and Societal Reconfigurations

Authors: Kashifa Khalid, Eesham Fatima

Abstract:

This study aims to analyze the loss of identity and its impact on one’s life in ‘The Last White Man’ by Mohsin Hamid. The theory of Alienation Effect by Bertolt Brecht has been applied to the text as Hamid offers the readers a unique perspective, alluding to significant themes like identity, race, and death. The aspects of defamiliarization align impeccably with the plot, as existence and the corresponding concept of identity seem to have dissolved into utter chaos. This extends from the unexplained transformation to the way the entire world unravels from its general norm into a dystopian mayhem. The characters, starting with the protagonist Anders, have lost their center. One’s own self transforms into the ‘other,’ and the struggle is to get refamiliarized with one’s own self. Alienation and isolation only rise as the construct of race and identity is taken apart brick by brick, ironically at its own pace as many new realities are blown to bits. The inseparable relationship between identity and grief under the ever-looming cloud of ‘death’ is studied in detail. The theoretical framework and thematic aspects harmonize in accordance with the writing style put forth by Hamid, tying all the loose ends together.

Keywords: alienation, chaos, identity, transformation

Procedia PDF Downloads 44
5811 Impact of Changes of the Conceptual Framework for Financial Reporting on the Indicators of the Financial Statement

Authors: Nadezhda Kvatashidze

Abstract:

The International Accounting Standards Board updated the conceptual framework for financial reporting. The main reason behind it is to resolve the tasks of the accounting, which are caused by the market development and business-transactions of a new economic content. Also, the investors call for higher transparency of information and responsibility for the results in order to make a more accurate risk assessment and forecast. All these make it necessary to further develop the conceptual framework for financial reporting so that the users get useful information. The market development and certain shortcomings of the conceptual framework revealed in practice require its reconsideration and finding new solutions. Some issues and concepts, such as disclosure and supply of information, its qualitative characteristics, assessment, and measurement uncertainty had to be supplemented and perfected. The criteria of recognition of certain elements (assets and liabilities) of reporting had to be updated, too and all this is set out in the updated edition of the conceptual framework for financial reporting, a comprehensive collection of concepts underlying preparation of the financial statement. The main objective of conceptual framework revision is to improve financial reporting and development of clear concepts package. This will support International Accounting Standards Board (IASB) to set common “Approach & Reflection” for similar transactions on the basis of mutually accepted concepts. As a result, companies will be able to develop coherent accounting policies for those transactions or events that are occurred from particular deals to which no standard is used or when standard allows choice of accounting policy.

Keywords: conceptual framework, measurement basis, measurement uncertainty, neutrality, prudence, stewardship

Procedia PDF Downloads 126
5810 Knowledge and Organisational Success: Developing a Scale of Knowledge Framework

Authors: Mohammed Almohammedali, David Edgar, Duncan Peter

Abstract:

The aim of this exploratory research is to further understand how organisations can evaluate their activities, which generate knowledge creation, to meet changing stakeholder expectations. A Scale of Knowledge (SoK) Framework is proposed which links knowledge management and organisational activities to changing stakeholder expectations. The framework was informed by the knowledge management literature, as well as empirical work conducted via a single case study of a multi-site hospital organisation in Saudi Arabia. Eight in-depth semi-structured interviews were conducted with managers from across the organisation regarding current and future stakeholder expectations, organisational strategy/activities and knowledge management. Data were analysed using thematic analysis and a hierarchical value map technique to identify activities that can produce further knowledge and consequently impact on how stakeholder expectations are met. The SoK Framework developed may be useful to practitioners as an analytical aid to determine if current organisational activities produce organisational knowledge which helps them meet (increasingly higher levels of) stakeholder expectations. The limitations of the research and avenues for future development of the proposed framework are discussed.

Keywords: knowledge creation, knowledge management, organisational knowledge, analytical aid, stakeholders

Procedia PDF Downloads 433
5809 Large Eddy Simulation of Particle Clouds Using Open-Source CFD

Authors: Ruo-Qian Wang

Abstract:

Open-source CFD has become increasingly popular and promising. The recent progress in multiphase flow enables new CFD applications, which provides an economic and flexible research tool for complex flow problems. Our numerical study using four-way coupling Euler-Lagrangian Large-Eddy Simulations to resolve particle cloud dynamics with OpenFOAM and CFDEM will be introduced: The fractioned Navier-Stokes equations are numerically solved for fluid phase motion, solid phase motion is addressed by Lagrangian tracking for every single particle, and total momentum is conserved by fluid-solid inter-phase coupling. The grid convergence test was performed, which proves the current resolution of the mesh is appropriate. Then, we validated the code by comparing numerical results with experiments in terms of particle cloud settlement and growth. A good comparison was obtained showing reliability of the present numerical schemes. The time and height at phase separations were defined and analyzed for a variety of initial release conditions. Empirical formulas were drawn to fit the results.

Keywords: four-way coupling, dredging, land reclamation, multiphase flows, oil spill

Procedia PDF Downloads 429
5808 Computing the Similarity and the Diversity in the Species Based on Cronobacter Genome

Authors: E. Al Daoud

Abstract:

The purpose of computing the similarity and the diversity in the species is to trace the process of evolution and to find the relationship between the species and discover the unique, the special, the common and the universal proteins. The proteins of the whole genome of 40 species are compared with the cronobacter genome which is used as reference genome. More than 3 billion pairwise alignments are performed using blastp. Several findings are introduced in this study, for example, we found 172 proteins in cronobacter genome which have insignificant hits in other species, 116 significant proteins in the all tested species with very high score value and 129 common proteins in the plants but have insignificant hits in mammals, birds, fishes, and insects.

Keywords: genome, species, blastp, conserved genes, Cronobacter

Procedia PDF Downloads 496
5807 Enhanced Model for Risk-Based Assessment of Employee Security with Bring Your Own Device Using Cyber Hygiene

Authors: Saidu I. R., Shittu S. S.

Abstract:

As the trend of personal devices accessing corporate data continues to rise through Bring Your Own Device (BYOD) practices, organizations recognize the potential cost reduction and productivity gains. However, the associated security risks pose a significant threat to these benefits. Often, organizations adopt BYOD environments without fully considering the vulnerabilities introduced by human factors in this context. This study presents an enhanced assessment model that evaluates the security posture of employees in BYOD environments using cyber hygiene principles. The framework assesses users' adherence to best practices and guidelines for maintaining a secure computing environment, employing scales and the Euclidean distance formula. By utilizing this algorithm, the study measures the distance between users' security practices and the organization's optimal security policies. To facilitate user evaluation, a simple and intuitive interface for automated assessment is developed. To validate the effectiveness of the proposed framework, design science research methods are employed, and empirical assessments are conducted using five artifacts to analyze user suitability in BYOD environments. By addressing the human factor vulnerabilities through the assessment of cyber hygiene practices, this study aims to enhance the overall security of BYOD environments and enable organizations to leverage the advantages of this evolving trend while mitigating potential risks.

Keywords: security, BYOD, vulnerability, risk, cyber hygiene

Procedia PDF Downloads 76
5806 3D Classification Optimization of Low-Density Airborne Light Detection and Ranging Point Cloud by Parameters Selection

Authors: Baha Eddine Aissou, Aichouche Belhadj Aissa

Abstract:

Light detection and ranging (LiDAR) is an active remote sensing technology used for several applications. Airborne LiDAR is becoming an important technology for the acquisition of a highly accurate dense point cloud. A classification of airborne laser scanning (ALS) point cloud is a very important task that still remains a real challenge for many scientists. Support vector machine (SVM) is one of the most used statistical learning algorithms based on kernels. SVM is a non-parametric method, and it is recommended to be used in cases where the data distribution cannot be well modeled by a standard parametric probability density function. Using a kernel, it performs a robust non-linear classification of samples. Often, the data are rarely linearly separable. SVMs are able to map the data into a higher-dimensional space to become linearly separable, which allows performing all the computations in the original space. This is one of the main reasons that SVMs are well suited for high-dimensional classification problems. Only a few training samples, called support vectors, are required. SVM has also shown its potential to cope with uncertainty in data caused by noise and fluctuation, and it is computationally efficient as compared to several other methods. Such properties are particularly suited for remote sensing classification problems and explain their recent adoption. In this poster, the SVM classification of ALS LiDAR data is proposed. Firstly, connected component analysis is applied for clustering the point cloud. Secondly, the resulting clusters are incorporated in the SVM classifier. Radial basic function (RFB) kernel is used due to the few numbers of parameters (C and γ) that needs to be chosen, which decreases the computation time. In order to optimize the classification rates, the parameters selection is explored. It consists to find the parameters (C and γ) leading to the best overall accuracy using grid search and 5-fold cross-validation. The exploited LiDAR point cloud is provided by the German Society for Photogrammetry, Remote Sensing, and Geoinformation. The ALS data used is characterized by a low density (4-6 points/m²) and is covering an urban area located in residential parts of the city Vaihingen in southern Germany. The class ground and three other classes belonging to roof superstructures are considered, i.e., a total of 4 classes. The training and test sets are selected randomly several times. The obtained results demonstrated that a parameters selection can orient the selection in a restricted interval of (C and γ) that can be further explored but does not systematically lead to the optimal rates. The SVM classifier with hyper-parameters is compared with the most used classifiers in literature for LiDAR data, random forest, AdaBoost, and decision tree. The comparison showed the superiority of the SVM classifier using parameters selection for LiDAR data compared to other classifiers.

Keywords: classification, airborne LiDAR, parameters selection, support vector machine

Procedia PDF Downloads 147
5805 Optimizing Production Yield Through Process Parameter Tuning Using Deep Learning Models: A Case Study in Precision Manufacturing

Authors: Tolulope Aremu

Abstract:

This paper is based on the idea of using deep learning methodology for optimizing production yield by tuning a few key process parameters in a manufacturing environment. The study was explicitly on how to maximize production yield and minimize operational costs by utilizing advanced neural network models, specifically Long Short-Term Memory and Convolutional Neural Networks. These models were implemented using Python-based frameworks—TensorFlow and Keras. The targets of the research are the precision molding processes in which temperature ranges between 150°C and 220°C, the pressure ranges between 5 and 15 bar, and the material flow rate ranges between 10 and 50 kg/h, which are critical parameters that have a great effect on yield. A dataset of 1 million production cycles has been considered for five continuous years, where detailed logs are present showing the exact setting of parameters and yield output. The LSTM model would model time-dependent trends in production data, while CNN analyzed the spatial correlations between parameters. Models are designed in a supervised learning manner. For the model's loss, an MSE loss function is used, optimized through the Adam optimizer. After running a total of 100 training epochs, 95% accuracy was achieved by the models recommending optimal parameter configurations. Results indicated that with the use of RSM and DOE traditional methods, there was an increase in production yield of 12%. Besides, the error margin was reduced by 8%, hence consistent quality products from the deep learning models. The monetary value was annually around $2.5 million, the cost saved from material waste, energy consumption, and equipment wear resulting from the implementation of optimized process parameters. This system was deployed in an industrial production environment with the help of a hybrid cloud system: Microsoft Azure, for data storage, and the training and deployment of their models were performed on Google Cloud AI. The functionality of real-time monitoring of the process and automatic tuning of parameters depends on cloud infrastructure. To put it into perspective, deep learning models, especially those employing LSTM and CNN, optimize the production yield by fine-tuning process parameters. Future research will consider reinforcement learning with a view to achieving further enhancement of system autonomy and scalability across various manufacturing sectors.

Keywords: production yield optimization, deep learning, tuning of process parameters, LSTM, CNN, precision manufacturing, TensorFlow, Keras, cloud infrastructure, cost saving

Procedia PDF Downloads 29
5804 Innovation in PhD Training in the Interdisciplinary Research Institute

Authors: B. Shaw, K. Doherty

Abstract:

The Cultural Communication and Computing Research Institute (C3RI) is a diverse multidisciplinary research institute including art, design, media production, communication studies, computing and engineering. Across these disciplines it can seem like there are enormous differences of research practice and convention, including differing positions on objectivity and subjectivity, certainty and evidence, and different political and ethical parameters. These differences sit within, often unacknowledged, histories, codes, and communication styles of specific disciplines, and it is all these aspects that can make understanding of research practice across disciplines difficult. To explore this, a one day event was orchestrated, testing how a PhD community might communicate and share research in progress in a multi-disciplinary context. Instead of presenting results at a conference, research students were tasked to articulate their method of inquiry. A working party of students from across disciplines had to design a conference call, visual identity and an event framework that would work for students across all disciplines. The process of establishing the shape and identity of the conference was revealing. Even finding a linguistic frame that would meet the expectations of different disciplines for the conference call was challenging. The first abstracts submitted either resorted to reporting findings, or only described method briefly. It took several weeks of supported intervention for research students to get ‘inside’ their method and to understand their research practice as a process rich with philosophical and practical decisions and implications. In response to the abstracts the conference committee generated key methodological categories for conference sessions, including sampling, capturing ‘experience’, ‘making models’, researcher identities, and ‘constructing data’. Each session involved presentations by visual artists, communications students and computing researchers with inter-disciplinary dialogue, facilitated by alumni Chairs. The apparently simple focus on method illuminated research process as a site of creativity, innovation and discovery, and also built epistemological awareness, drawing attention to what is being researched and how it can be known. It was surprisingly difficult to limit students to discussing method, and it was apparent that the vocabulary available for method is sometimes limited. However, by focusing on method rather than results, the genuine process of research, rather than one constructed for approval, could be captured. In unlocking the twists and turns of planning and implementing research, and the impact of circumstance and contingency, students had to reflect frankly on successes and failures. This level of self – and public- critique emphasised the degree of critical thinking and rigour required in executing research and demonstrated that honest reportage of research, faults and all, is good valid research. The process also revealed the degree that disciplines can learn from each other- the computing students gained insights from the sensitive social contextualizing generated by communications and art and design students, and art and design students gained understanding from the greater ‘distance’ and emphasis on application that computing students applied to their subjects. Finding the means to develop dialogue across disciplines makes researchers better equipped to devise and tackle research problems across disciplines, potentially laying the ground for more effective collaboration.

Keywords: interdisciplinary, method, research student, training

Procedia PDF Downloads 206
5803 Developing Cyber Security Asset Mangement Framework for UK Rail

Authors: Shruti Kohli

Abstract:

The sophistication and pervasiveness of cyber-attacks are constantly growing, driven partly by technological progress, profitable applications in organized crime and state-sponsored innovation. The modernization of rail control systems has resulted in an increasing reliance on digital technology and increased the potential for security breaches and cyber-attacks. This research track showcases the need for developing a secure reusable scalable framework for enhancing cyber security of rail assets. A cyber security framework has been proposed that is being developed to detect the tell-tale signs of cyber-attacks against industrial assets.

Keywords: cyber security, rail asset, security threat, cyber ontology

Procedia PDF Downloads 430
5802 Safe and Scalable Framework for Participation of Nodes in Smart Grid Networks in a P2P Exchange of Short-Term Products

Authors: Maciej Jedrzejczyk, Karolina Marzantowicz

Abstract:

Traditional utility value chain is being transformed during last few years into unbundled markets. Increased distributed generation of energy is one of considerable challenges faced by Smart Grid networks. New sources of energy introduce volatile demand response which has a considerable impact on traditional middlemen in E&U market. The purpose of this research is to search for ways to allow near-real-time electricity markets to transact with surplus energy based on accurate time synchronous measurements. A proposed framework evaluates the use of secure peer-2-peer (P2P) communication and distributed transaction ledgers to provide flat hierarchy, and allow real-time insights into present and forecasted grid operations, as well as state and health of the network. An objective is to achieve dynamic grid operations with more efficient resource usage, higher security of supply and longer grid infrastructure life cycle. Methods used for this study are based on comparative analysis of different distributed ledger technologies in terms of scalability, transaction performance, pluggability with external data sources, data transparency, privacy, end-to-end security and adaptability to various market topologies. An intended output of this research is a design of a framework for safer, more efficient and scalable Smart Grid network which is bridging a gap between traditional components of the energy network and individual energy producers. Results of this study are ready for detailed measurement testing, a likely follow-up in separate studies. New platforms for Smart Grid achieving measurable efficiencies will allow for development of new types of Grid KPI, multi-smart grid branches, markets, and businesses.

Keywords: autonomous agents, Distributed computing, distributed ledger technologies, large scale systems, micro grids, peer-to-peer networks, Self-organization, self-stabilization, smart grids

Procedia PDF Downloads 300
5801 Students' Ability to Solve Complex Accounting Problems Using a Framework-Based Approach

Authors: Karen Odendaal

Abstract:

Accounting transactions are becoming more complex, and more extensive accounting guidance is provided on a continuous basis. It is widely perceived that conceptual teaching of accounting contributes to lifelong learning. Such a conceptual teaching approach also contributes to effective accounting problem-solving. This framework-based approach is rooted in educational psychologies such as constructivism and Ausubel’s subsumption theory. This study aimed at investigating the ability of students to solve complex accounting problems by using only concepts underlying the Conceptual Framework. An assignment was administered to pre-graduate students at a South African university and this study made use of an interpretative research design which implemented multiple research instruments to investigate the ability of students to solve complex accounting problems using only concepts underlying the Conceptual Framework. Student perceptions were analysed and were aided by a related reflective questionnaire. The importance of the study indicates the necessity of Accounting educators to enhance a conceptual understanding among students as a mechanism for problem-solving of accounting issues. The results indicate that the ability of students to solve accounting problems effectively using only the Conceptual Framework depends on the complexity of the scenario and the students’ familiarity with the problem. The study promotes a balanced and more conceptual (rather than only technical) preference to the problem-solving of complex accounting problems. The study indubitably promotes considerable emphasis on the importance of the Conceptual Framework in accounting education and the promotion of life-long learning in the subject field.

Keywords: accounting education, conceptual teaching, constructivism, framework-based, problem-solving

Procedia PDF Downloads 233
5800 Computing Continuous Skyline Queries without Discriminating between Static and Dynamic Attributes

Authors: Ibrahim Gomaa, Hoda M. O. Mokhtar

Abstract:

Although most of the existing skyline queries algorithms focused basically on querying static points through static databases; with the expanding number of sensors, wireless communications and mobile applications, the demand for continuous skyline queries has increased. Unlike traditional skyline queries which only consider static attributes, continuous skyline queries include dynamic attributes, as well as the static ones. However, as skyline queries computation is based on checking the domination of skyline points over all dimensions, considering both the static and dynamic attributes without separation is required. In this paper, we present an efficient algorithm for computing continuous skyline queries without discriminating between static and dynamic attributes. Our algorithm in brief proceeds as follows: First, it excludes the points which will not be in the initial skyline result; this pruning phase reduces the required number of comparisons. Second, the association between the spatial positions of data points is examined; this phase gives an idea of where changes in the result might occur and consequently enables us to efficiently update the skyline result (continuous update) rather than computing the skyline from scratch. Finally, experimental evaluation is provided which demonstrates the accuracy, performance and efficiency of our algorithm over other existing approaches.

Keywords: continuous query processing, dynamic database, moving object, skyline queries

Procedia PDF Downloads 210
5799 A Geographical Framework for Studying the Territorial Sustainability Based on Land Use Change

Authors: Miguel Ramirez, Ivan Lizarazo

Abstract:

The emergence of various interpretations of sustainability, including weak and strong paradigms, can be traced back to the definition of sustainable development provided in the 1987 Brundtland report and the subsequent evolution of the sustainability concept. However, there has been limited scholarly attention given to clarifying the concept of sustainability within the theoretical and conceptual framework of geography. The discipline has predominantly been focused on understanding the diverse conceptions of sustainability within its epistemological boundaries, resulting in tensions between sustainability paradigms and their associated dimensions, including the incorporation of political perspectives, with particular emphasis on environmental geography's epistemology. In response to this gap, a conceptual framework for sustainability is proposed, effectively integrating spatial and territorial concepts. This framework aims to enhance geography's role in contributing to sustainability by utilizing the land system theory, which is based on the dynamics of land use change. Such an integrated conceptual framework enables incorporating methodological tools such as remote sensing, encompassing various earth observations and fusion methods, and supervised classification techniques. Additionally, it looks for better integration of socioecological information, thereby capturing essential population-related features.

Keywords: geography, sustainability, land change science, territorial sustainability

Procedia PDF Downloads 80