Search results for: cloud service models
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 10214

Search results for: cloud service models

7694 Barriers to Entry: The Pitfall of Charter School Accountability

Authors: Ian Kingsbury

Abstract:

The rapid expansion of charter schools (public schools that receive government but do not face the same regulations as traditional public schools) over the preceding two decades has raised concerns over the potential for graft and fraud. These concerns are largely justified: Incidents of financial crime and mismanagement are not unheard of, and the charter sector has become a darling of hedge fund managers. In response, several states have strengthened their charter school regulatory regimes. Imposing regulations and attempting to increase accountability seem like sensible measures, and perhaps they are necessary. However, increased regulation may come at the cost of imposing barriers to entry. Specifically, increased regulation often entails evidence for a high likelihood of fiscal solvency. That should theoretically entail access to capital in the short-term, which may systematically preclude Black or Hispanic applicants from opening charter schools. Moreover, increased regulation necessarily entails more red tape. The institutional wherewithal and the number of hours required to complete an application to open a charter school might favor those who have partnered with an education service provider, specifically a charter management organization (CMO) or education management organization (EMO). These potential barriers to entry pose a significant policy concern. Just as policymakers hope to increase the share of minority teachers and principals, they should sensibly care whether individuals who open charter schools look like the students in that school. Moreover, they might be concerned if successful applications in states with stringent regulations are overwhelmingly affiliated with education service providers. One of the original missions of charter schools was to serve as a laboratory of innovation. Approving only those applications affiliated with education service providers (and in effect establishing a parallel network of schools rather than a diverse marketplace of schools) undermines that mission. Data and methods: The analysis examines more than 2,000 charter school applications from 15 states. It compares the outcomes of applications from states with a strong regulatory environment (those with high scores) from NACSA-the National Association of Charter School Authorizers- to applications from states with a weak regulatory environment (those with a low NACSA score). If the hypothesis is correct, applicants not affiliated with an ESP are more likely to be rejected in high-regulation states compared to those affiliated with an ESP, and minority candidates not affiliated with an education service provider (ESP) are particularly likely to be rejected. Initial returns indicate that the hypothesis holds. More applications in low NASCA-scoring Arizona come from individuals not associated with an ESP, and those individuals are as likely to be accepted as those affiliated with an ESP. On the other hand, applicants in high-NACSA scoring Indiana and Ohio are more than 20 percentage points more likely to be accepted if they are affiliated with an ESP, and the effect is particularly pronounced for minority candidates. These findings should spur policymakers to consider the drawbacks of charter school accountability and consider accountability regimes that do not impose barriers to entry.

Keywords: accountability, barriers to entry, charter schools, choice

Procedia PDF Downloads 133
7693 Evaluation of Short-Term Load Forecasting Techniques Applied for Smart Micro-Grids

Authors: Xiaolei Hu, Enrico Ferrera, Riccardo Tomasi, Claudio Pastrone

Abstract:

Load Forecasting plays a key role in making today's and future's Smart Energy Grids sustainable and reliable. Accurate power consumption prediction allows utilities to organize in advance their resources or to execute Demand Response strategies more effectively, which enables several features such as higher sustainability, better quality of service, and affordable electricity tariffs. It is easy yet effective to apply Load Forecasting at larger geographic scale, i.e. Smart Micro Grids, wherein the lower available grid flexibility makes accurate prediction more critical in Demand Response applications. This paper analyses the application of short-term load forecasting in a concrete scenario, proposed within the EU-funded GreenCom project, which collect load data from single loads and households belonging to a Smart Micro Grid. Three short-term load forecasting techniques, i.e. linear regression, artificial neural networks, and radial basis function network, are considered, compared, and evaluated through absolute forecast errors and training time. The influence of weather conditions in Load Forecasting is also evaluated. A new definition of Gain is introduced in this paper, which innovatively serves as an indicator of short-term prediction capabilities of time spam consistency. Two models, 24- and 1-hour-ahead forecasting, are built to comprehensively compare these three techniques.

Keywords: short-term load forecasting, smart micro grid, linear regression, artificial neural networks, radial basis function network, gain

Procedia PDF Downloads 444
7692 On the End-of-Life Inventory Problem

Authors: Hans Frenk, Sonya Javadi, Semih Onur Sezer

Abstract:

We consider the so-called end of life inventory problem for the supplier of a product in its final phase of the service life cycle. This phase starts when the production of the items stops and continues until the warranty of the last sold item expires. At the beginning of this phase, the supplier places a final order for spare parts to serve customers coming with defective items. At any time during the final phase, the supplier may also decide to switch to an alternative and more cost-effective policy. This alternative policy may be in the form of replacing a defective item with a substitutable product or offering discounts / rebates on new generation products. In this setup, the objective is to find a final order quantity and also a switching time which will minimize the total expected discounted cost. We study this problem under a general cost structure in a continuous-time framework where arrivals of defective items are given by a non-homogeneous Poisson process. We consider four formulations which differ by the nature of the switching time. These formulations are studied in detail and properties of the objective function are derived in each case. Using these properties, we provide exact algorithms for efficient numerical implementations. Numerical examples are provided illustrating the application of these algorithms. In these examples, we also compare the costs associated with these different formulations.

Keywords: End-of-life inventory control, martingales, optimization, service parts

Procedia PDF Downloads 312
7691 Federal Bureau of Investigation Opposition to German Nationalist Organizations in the United States (1941-45)

Authors: Yaroslav Alexandrovich Levin

Abstract:

In modern research on the history of the United States in World War II, it is quite popular to study the opposition of the American special services and, in particular, the Federal Bureau of Investigation (FBI) to various organizations of the German diasporas in new historical conditions. The appeal to traditional methods of historical research, comparative studies, and the principles of historicism will make it possible to more accurately trace the process of tightening the counterintelligence work of the Bureau and the close connection of concerns about the involvement of public organizations in the intelligence activities of the enemy. The broadcast of nationalist ideas by various communities of Germans under the auspices of their governments quickly attracted the attention of the FBI, which is in the process of consolidating its powers as the main US counterintelligence service. At the same time, the investigations and trials conducted by the John Edgar Hoover Department following these investigations often had an openly political color and increasingly consolidated the beginning of a political investigation in this service. This practice and its implementation ran into a tough contradiction between the legal norms of America, which proclaimed "democratic values," the right to freedom of speech, and the need to strengthen the internal security of the state and society in wartime. All these processes and the associated nuances and complexities are considered in specific examples of the work of federal agents against various pro-German organizations in the period 1941-45.

Keywords: World War II, internal security, countering extremism, counterintelligence, political investigation, FBI

Procedia PDF Downloads 65
7690 Influence of Telkom Membership Card Customer Perceived Value on Retaining PT. Telkom Indonesia's Customer in 2013-2014

Authors: Eka Yuliana, Siska Shabrina Julyan

Abstract:

The competitive environment and high customer’s churn rate in telecommunication industries lead Indonesian telecommunication companies become strive to offer products with more value. Offering product with more value can encourage customers to keep using the companies product. One of way to retain customer is give a membership card to the customers as practiced by PT. Telkom by giving Telkom Membership Card to PT. Telkom loyal customer. This study aims to determine the influence of Telkom Membership Card customer perceived value on retaining PT. Telkom Indonesia’s customer in 2013-2014 by using quantitative method with causal study. Analythical technique used in this study is Structural Equation Modelling (SEM) to test the causal relationship with 216 owner of Telkom Membership Card in Indonesia. This study conclude that: (i) Customer perceived value on Telkom Membership Card is located in fair value zone, (ii) PT. Telkom efforts in order to retain the customers is classified as good, (iii) Customer perceived value is influencing the effort to retain the customer with the probability value less than 0.05 and level of influence 69%. Based on result of this study, PT. Telkom should (i) Improve Telkom Membership Card’s promotion because not all customer of PT. Telkom have the membership card. (iia) Adding Telkom Membership Card’s benefit such as discount at various merchant (iib) Making call center for member of Telkom Membership Card (iii) PT. Telkom should be ensure availability of their service. (iv) PT. Telkom should make a priority to customer who have telkom membership card and offers a better service.For future research should be use different variables.

Keywords: customer perceived value, customer retention, marketing, relationship marketing

Procedia PDF Downloads 297
7689 A Control Model for the Dismantling of Industrial Plants

Authors: Florian Mach, Eric Hund, Malte Stonis

Abstract:

The dismantling of disused industrial facilities such as nuclear power plants or refineries is an enormous challenge for the planning and control of the logistic processes. Existing control models do not meet the requirements for a proper dismantling of industrial plants. Therefore, the paper presents an approach for the control of dismantling and post-processing processes (e.g. decontamination) in plant decommissioning. In contrast to existing approaches, the dismantling sequence and depth are selected depending on the capacity utilization of required post-processing processes by also considering individual characteristics of respective dismantling tasks (e.g. decontamination success rate, uncertainties regarding the process times). The results can be used in the dismantling of industrial plants (e.g. nuclear power plants) to reduce dismantling time and costs by avoiding bottlenecks such as capacity constraints.

Keywords: dismantling management, logistics planning and control models, nuclear power plant dismantling, reverse logistics

Procedia PDF Downloads 286
7688 Drying Characteristics of Shrimp by Using the Traditional Method of Oven

Authors: I. A. Simsek, S. N. Dogan, A. S. Kipcak, E. Morodor Derun, N. Tugrul

Abstract:

In this study, the drying characteristics of shrimp are studied by using the traditional drying method of oven. Drying temperatures are selected between 60-80°C. Obtained experimental drying results are applied to eleven mathematical models of Alibas, Aghbashlo et al., Henderson and Pabis, Jena and Das, Lewis, Logaritmic, Midilli and Kucuk, Page, Parabolic, Wang and Singh and Weibull. The best model was selected as parabolic based on the highest coefficient of determination (R²) (0.999990 at 80°C) and the lowest χ² (0.000002 at 80°C), and the lowest root mean square error (RMSE) (0.000976 at 80°C) values are compared to other models. The effective moisture diffusivity (Deff) values were calculated using the Fick’s second law’s cylindrical coordinate approximation and are found between 6.61×10⁻⁸ and 6.66×10⁻⁷ m²/s. The activation energy (Ea) was calculated using modified form of Arrhenius equation and is found as 18.315 kW/kg.

Keywords: activation energy, drying, effective moisture diffusivity, modelling, oven, shrimp

Procedia PDF Downloads 169
7687 Modelling the Art Historical Canon: The Use of Dynamic Computer Models in Deconstructing the Canon

Authors: Laura M. F. Bertens

Abstract:

There is a long tradition of visually representing the art historical canon, in schematic overviews and diagrams. This is indicative of the desire for scientific, ‘objective’ knowledge of the kind (seemingly) produced in the natural sciences. These diagrams will, however, always retain an element of subjectivity and the modelling methods colour our perception of the represented information. In recent decades visualisations of art historical data, such as hand-drawn diagrams in textbooks, have been extended to include digital, computational tools. These tools significantly increase modelling strength and functionality. As such, they might be used to deconstruct and amend the very problem caused by traditional visualisations of the canon. In this paper, the use of digital tools for modelling the art historical canon is studied, in order to draw attention to the artificial nature of the static models that art historians are presented with in textbooks and lectures, as well as to explore the potential of digital, dynamic tools in creating new models. To study the way diagrams of the canon mediate the represented information, two modelling methods have been used on two case studies of existing diagrams. The tree diagram Stammbaum der neudeutschen Kunst (1823) by Ferdinand Olivier has been translated to a social network using the program Visone, and the famous flow chart Cubism and Abstract Art (1936) by Alfred Barr has been translated to an ontological model using Protégé Ontology Editor. The implications of the modelling decisions have been analysed in an art historical context. The aim of this project has been twofold. On the one hand the translation process makes explicit the design choices in the original diagrams, which reflect hidden assumptions about the Western canon. Ways of organizing data (for instance ordering art according to artist) have come to feel natural and neutral and implicit biases and the historically uneven distribution of power have resulted in underrepresentation of groups of artists. Over the last decades, scholars from fields such as Feminist Studies, Postcolonial Studies and Gender Studies have considered this problem and tried to remedy it. The translation presented here adds to this deconstruction by defamiliarizing the traditional models and analysing the process of reconstructing new models, step by step, taking into account theoretical critiques of the canon, such as the feminist perspective discussed by Griselda Pollock, amongst others. On the other hand, the project has served as a pilot study for the use of digital modelling tools in creating dynamic visualisations of the canon for education and museum purposes. Dynamic computer models introduce functionalities that allow new ways of ordering and visualising the artworks in the canon. As such, they could form a powerful tool in the training of new art historians, introducing a broader and more diverse view on the traditional canon. Although modelling will always imply a simplification and therefore a distortion of reality, new modelling techniques can help us get a better sense of the limitations of earlier models and can provide new perspectives on already established knowledge.

Keywords: canon, ontological modelling, Protege Ontology Editor, social network modelling, Visone

Procedia PDF Downloads 111
7686 Quality of Service of Transportation Networks: A Hybrid Measurement of Travel Time and Reliability

Authors: Chin-Chia Jane

Abstract:

In a transportation network, travel time refers to the transmission time from source node to destination node, whereas reliability refers to the probability of a successful connection from source node to destination node. With an increasing emphasis on quality of service (QoS), both performance indexes are significant in the design and analysis of transportation systems. In this work, we extend the well-known flow network model for transportation networks so that travel time and reliability are integrated into the QoS measurement simultaneously. In the extended model, in addition to the general arc capacities, each intermediate node has a time weight which is the travel time for per unit of commodity going through the node. Meanwhile, arcs and nodes are treated as binary random variables that switch between operation and failure with associated probabilities. For pre-specified travel time limitation and demand requirement, the QoS of a transportation network is the probability that source can successfully transport the demand requirement to destination while the total transmission time is under the travel time limitation. This work is pioneering, since existing literatures that evaluate travel time reliability via a single optimization path, the proposed QoS focuses the performance of the whole network system. To compute the QoS of transportation networks, we first transfer the extended network model into an equivalent min-cost max-flow network model. In the transferred network, each arc has a new travel time weight which takes value 0. Each intermediate node is replaced by two nodes u and v, and an arc directed from u to v. The newly generated nodes u and v are perfect nodes. The new direct arc has three weights: travel time, capacity, and operation probability. Then the universal set of state vectors is recursively decomposed into disjoint subsets of reliable, unreliable, and stochastic vectors until no stochastic vector is left. The decomposition is made possible by applying existing efficient min-cost max-flow algorithm. Because the reliable subsets are disjoint, QoS can be obtained directly by summing the probabilities of these reliable subsets. Computational experiments are conducted on a benchmark network which has 11 nodes and 21 arcs. Five travel time limitations and five demand requirements are set to compute the QoS value. To make a comparison, we test the exhaustive complete enumeration method. Computational results reveal the proposed algorithm is much more efficient than the complete enumeration method. In this work, a transportation network is analyzed by an extended flow network model where each arc has a fixed capacity, each intermediate node has a time weight, and both arcs and nodes are independent binary random variables. The quality of service of the transportation network is an integration of customer demands, travel time, and the probability of connection. We present a decomposition algorithm to compute the QoS efficiently. Computational experiments conducted on a prototype network show that the proposed algorithm is superior to existing complete enumeration methods.

Keywords: quality of service, reliability, transportation network, travel time

Procedia PDF Downloads 199
7685 Energy Use and Econometric Models of Soybean Production in Mazandaran Province of Iran

Authors: Majid AghaAlikhani, Mostafa Hojati, Saeid Satari-Yuzbashkandi

Abstract:

This paper studies energy use patterns and relationship between energy input and yield for soybean (Glycine max (L.) Merrill) in Mazandaran province of Iran. In this study, data were collected by administering a questionnaire in face-to-face interviews. Results revealed that the highest share of energy consumption belongs to chemical fertilizers (29.29%) followed by diesel (23.42%) and electricity (22.80%). Our investigations showed that a total energy input of 23404.1 MJ.ha-1 was consumed for soybean production. The energy productivity, specific energy, and net energy values were estimated as 0.12 kg MJ-1, 8.03 MJ kg-1, and 49412.71 MJ.ha-1, respectively. The ratio of energy outputs to energy inputs was 3.11. Obtained results indicated that direct, indirect, renewable and non-renewable energies were (56.83%), (43.17%), (15.78%) and (84.22%), respectively. Three econometric models were also developed to estimate the impact of energy inputs on yield. The results of econometric models revealed that impact of chemical, fertilizer, and water on yield were significant at 1% probability level. Also, direct and non-renewable energies were found to be rather high. Cost analysis revealed that total cost of soybean production per ha was around 518.43$. Accordingly, the benefit-cost ratio was estimated as 2.58. The energy use efficiency in soybean production was found as 3.11. This reveals that the inputs used in soybean production are used efficiently. However, due to higher rate of nitrogen fertilizer consumption, sustainable agriculture should be extended and extension staff could be proposed substitution of chemical fertilizer by biological fertilizer or green manure.

Keywords: Cobbe Douglas function, economical analysis, energy efficiency, energy use patterns, soybean

Procedia PDF Downloads 310
7684 Factors Affecting Customer Loyalty in the Independent Surveyor Service Industry in Indonesia

Authors: Sufrin Hannan, Budi Suharjo, Rita Nurmalina, Kirbrandoko

Abstract:

The challenge for independent surveyor service companies now is growing with increasing uncertainty in business. Protection from the government for domestic independent surveyor industry from competitor attack, such as entering the global surveyors to Indonesia also no longer exists. Therefore, building customer loyalty becomes very important to create a long-term relationship between an independent surveyor with its customers. This study aims to develop a model that can be used to build customer loyalty by looking at various factors that determine customer loyalty, especially on independent surveyors for coal inspection in Indonesia. The development of this model uses the relationship marketing approach. Testing of the hypothesis is done by testing the variables that determine customer loyalty, either directly or indirectly, which amounted to 10 variables. The data were collected from 200 questionnaires filled by independent surveyor company decision makers from 51 exporting companies and coal trading companies in Indonesia and analyzed using Structural Equation Model (SEM). The results show that customer loyalty of independent surveyors is influenced by customer satisfaction, trust, switching-barrier, and relationship-bond. Research on customer satisfaction shows that customer satisfaction is influenced by the perceived quality and perceived value, while perceived quality is influenced by reliability, assurance, responsiveness, and empathy.

Keywords: relationship marketing, customer loyalty, customer satisfaction, switching barriers, relationship bonds

Procedia PDF Downloads 153
7683 Teaching Tools for Web Processing Services

Authors: Rashid Javed, Hardy Lehmkuehler, Franz Josef-Behr

Abstract:

Web Processing Services (WPS) have up growing concern in geoinformation research. However, teaching about them is difficult because of the generally complex circumstances of their use. They limit the possibilities for hands- on- exercises on Web Processing Services. To support understanding however a Training Tools Collection was brought on the way at University of Applied Sciences Stuttgart (HFT). It is limited to the scope of Geostatistical Interpolation of sample point data where different algorithms can be used like IDW, Nearest Neighbor etc. The Tools Collection aims to support understanding of the scope, definition and deployment of Web Processing Services. For example it is necessary to characterize the input of Interpolation by the data set, the parameters for the algorithm and the interpolation results (here a grid of interpolated values is assumed). This paper reports on first experiences using a pilot installation. This was intended to find suitable software interfaces for later full implementations and conclude on potential user interface characteristics. Experiences were made with Deegree software, one of several Services Suites (Collections). Being strictly programmed in Java, Deegree offers several OGC compliant Service Implementations that also promise to be of benefit for the project. The mentioned parameters for a WPS were formalized following the paradigm that any meaningful component will be defined in terms of suitable standards. E.g. the data output can be defined as a GML file. But, the choice of meaningful information pieces and user interactions is not free but partially determined by the selected WPS Processing Suite.

Keywords: deegree, interpolation, IDW, web processing service (WPS)

Procedia PDF Downloads 338
7682 Annual Water Level Simulation Using Support Vector Machine

Authors: Maryam Khalilzadeh Poshtegal, Seyed Ahmad Mirbagheri, Mojtaba Noury

Abstract:

In this paper, by application of the input yearly data of rainfall, temperature and flow to the Urmia Lake, the simulation of water level fluctuation were applied by means of three models. According to the climate change investigation the fluctuation of lakes water level are of high interest. This study investigate data-driven models, support vector machines (SVM), SVM method which is a new regression procedure in water resources are applied to the yearly level data of Lake Urmia that is the biggest and the hyper saline lake in Iran. The evaluated lake levels are found to be in good correlation with the observed values. The results of SVM simulation show better accuracy and implementation. The mean square errors, mean absolute relative errors and determination coefficient statistics are used as comparison criteria.

Keywords: simulation, water level fluctuation, urmia lake, support vector machine

Procedia PDF Downloads 344
7681 Living Lab as a Service: Developing Context Induced, Co-creational Innovation Routines as a Process Tool for Nature Based Solutions

Authors: Immanuel Darkwa

Abstract:

Climate change and environmental degradation are existential threats requiring urgent transnational action. The SDGs, as well as regional initiatives the like European Green Deal, as ambitious as they are, put an emphasis on innovatively tackling threats posed by climate change regionally. While co-creational approaches are being propagated, there is no reference blueprint for how potential solutions, particularly nature-based solutions, may be developed and implemented within urban-settings. Using a single case study in Zagreb, Croatia, this paper proposes a workshop-tool for a Living Lab as a Service model for sustainable Nature-Based-Thinking, Nature–Centred-Design and Nature based solutions. The approach is based on a co-creational methodology developed through literature synthesis, expert interviews, focus group discussions, surveys and synthesized through rigorous research analysis and participatory observation. The ensuing tool involves workshop-processes, tested with through-the-process identified stakeholders with distinctive roles and functions. The resulting framework proposes a Nature-Based-Centred-Thinking process tool involving ‘green’ routines supported by a focal unit and a collaborative network, and that allows for the development of nature-based solutions.

Keywords: living labs, nature-based solutions, nature- based design, innovation processes, innovation routines and tools

Procedia PDF Downloads 52
7680 Willingness to Use Mobile Telephone Technology to Improve Pregnancy Outcomes among Women in Lagos, Nigeria

Authors: O. Onigbogi, M. Onigbogi Jr., O. Ojo

Abstract:

Background: The advances in mobile telephone technology has led to a rise in its use globally which has improved service delivery, empowered businesses and changed the way people access information. The practice of many health professionals has also been affected by the information and communications technology (ICT) revolution because they have better access to information. This study was conducted to assess the willingness to participate in mobile technology interventions to improve pregnancy outcomes in Lagos, Nigeria. Materials and Methods: A total of 238 respondents completed self-administered questionnaires. SPSS version 18 data editor was used to analyze data. Univariate odds ratios and 95% confidence intervals (95 % CI) were used to evaluate the correlates of Willingness to Use (WTU) mobile phones to receive health messages during pregnancy. Results: A total of 107 women (45% of the respondents) reported that they will be willing to receive health-related information on their phones during pregnancy. Greater willingness was associated with higher education (OR = 1.25, 95% CI: 1.13–1.53), involvement with community volunteer organizations (OR = 1.25, 95% CI: 1.05–1.52), monetary incentives (OR = 1.37, 95% CI: 1.14–1.45) and nulliparity (OR = 1.39, 95% CI: 1.02–1.42). Decreased willingness was associated with concerns about wrong interpretation of information (OR = 0.42, 95% CI: 0.21–0.54), poor mobile telephone service by providers (OR = 0.61, 95% CI: 0.52–0.78), increase in number of messages (OR = 0.78, 95% CI: 0.53–0.76). Conclusion: The level of WTU recorded indicates that much work still needs to be done before this novel approach could be used adopted in delivering health-related information. Incentives for would-be subjects should also be a part of the planning to encourage greater participation.

Keywords: mobile, outcomes, pregnancy, technology, telephone

Procedia PDF Downloads 204
7679 A Comparative Analysis of Asymmetric Encryption Schemes on Android Messaging Service

Authors: Mabrouka Algherinai, Fatma Karkouri

Abstract:

Today, Short Message Service (SMS) is an important means of communication. SMS is not only used in informal environment for communication and transaction, but it is also used in formal environments such as institutions, organizations, companies, and business world as a tool for communication and transactions. Therefore, there is a need to secure the information that is being transmitted through this medium to ensure security of information both in transit and at rest. But, encryption has been identified as a means to provide security to SMS messages in transit and at rest. Several past researches have proposed and developed several encryption algorithms for SMS and Information Security. This research aims at comparing the performance of common Asymmetric encryption algorithms on SMS security. The research employs the use of three algorithms, namely RSA, McEliece, and RABIN. Several experiments were performed on SMS of various sizes on android mobile device. The experimental results show that each of the three techniques has different key generation, encryption, and decryption times. The efficiency of an algorithm is determined by the time that it takes for encryption, decryption, and key generation. The best algorithm can be chosen based on the least time required for encryption. The obtained results show the least time when McEliece size 4096 is used. RABIN size 4096 gives most time for encryption and so it is the least effective algorithm when considering encryption. Also, the research shows that McEliece size 2048 has the least time for key generation, and hence, it is the best algorithm as relating to key generation. The result of the algorithms also shows that RSA size 1024 is the most preferable algorithm in terms of decryption as it gives the least time for decryption.

Keywords: SMS, RSA, McEliece, RABIN

Procedia PDF Downloads 145
7678 Social Studies Teachers’ Sustained, Collaborative Professional Development Centered Round Innovative Curriculum Materials

Authors: Cory Callahan

Abstract:

Here the author synthesizes findings and implications from two research studies that comprise a continuing line of inquiry into the potential of an innovative professional development program to help in-service teachers understand and implement a complex model of social studies instruction. The paper specifically explores the question: To what degree can a collaborative professional development program centered around innovative curriculum materials help social studies teachers understand and implement a powerful social studies approach? Findings suggest the teachers increasingly incorporated substantive thinking (i.e., second-order historical domain knowledge) into their respective practice and they facilitated students’ use of historical photographs as evidence to begin to answer a compelling question. The teachers also began to effectively support students’ abilities to make claims about the past. Implications include the foregrounding of high-quality questions during planning and the need for explicit guidance in the form of structures and procedures (i.e., scaffolds) to help teachers systematically review students’ work products. The work shared here may contribute to scholarship that posits explanations for why teacher-support is routinely ineffectual and suggests ways to provide substantive collaborative support for in-service social studies teachers.

Keywords: educative curriculum, social studies, professional development, lesson study

Procedia PDF Downloads 44
7677 A Critical Discourse Analysis of Jamaican and Trinidadian News Articles about D/Deafness

Authors: Melissa Angus Baboun

Abstract:

Utilizing a Critical Discourse Analysis (CDA) methodology and a theoretical framework based on disability studies, how Jamaican and Trinidadian newspapers discussed issues relating to the Deaf community were examined. The term deaf was inputted into the search engine tool of the online website for the Jamaica Observer and the Trinidad & Tobago Guardian. All 27 articles that contained the term deaf in its content and were written between August 1, 2017 and November 15, 2017 were chosen for the study. The data analysis was divided into three steps: (1) listing and analysis instances of metaphorical deafness (e.g. fall on deaf ears), (2) categorization of the content of the articles into the models of disability discourse (the medical, socio-cultural, and superscrip models of disability narratives), and (3) the analysis of any additional data found. A total of 42% of the articles pulled for this study did not deal with the Deaf community in any capacity, but rather instances of the use of idiomatic expressions that use deafness as a metaphor for a non-physical, undesirable trait. The most common idiomatic expression found was fall on deaf ears. Regarding the models of disability discourse, eight articles were found to follow the socio-cultural model, two were found to follow the medical model, and two were found to follow the superscrip model. The additional data found in these articles include two instances of the term deaf and mute, an overwhelming use of lower case d for the term deaf, and the misuse of the term translator (to mean interpreter).

Keywords: deafness, disability, news coverage, Caribbean newspapers

Procedia PDF Downloads 216
7676 A Verification Intellectual Property for Multi-Flow Rate Control on Any Single Flow Bus Functional Model

Authors: Pawamana Ramachandra, Jitesh Gupta, Saranga P. Pogula

Abstract:

In verification of high volume and complex packet processing IPs, finer control of flow management aspects (for example, rate, bits/sec etc.) per flow class (or a virtual channel or a software thread) is needed. When any Software/Universal Verification Methodology (UVM) thread arbitration is left to the simulator (e.g., Verilog Compiler Simulator (VCS) or Incisive Enterprise Simulator core simulation engine (NCSIM)), it is hard to predict its pattern of resulting distribution of bandwidth by the simulator thread arbitration. In many cases, the patterns desired in a test scenario may not be accomplished as the simulator might give a different distribution than what was required. This can lead to missing multiple traffic scenarios, specifically deadlock and starvation related. We invented a component (namely Flow Manager Verification IP) to be intervening between the application (test case) and the protocol VIP (with UVM sequencer) to control the bandwidth per thread/virtual channel/flow. The Flow Manager has knobs visible to the UVM sequence/test to configure the required distribution of rate per thread/virtual channel/flow. This works seamlessly and produces rate stimuli to further harness the Design Under Test (DUT) with asymmetric inputs compared to the programmed bandwidth/Quality of Service (QoS) distributions in the Design Under Test.

Keywords: flow manager, UVM sequencer, rated traffic generation, quality of service

Procedia PDF Downloads 77
7675 LACGC: Business Sustainability Research Model for Generations Consumption, Creation, and Implementation of Knowledge: Academic and Non-Academic

Authors: Satpreet Singh

Abstract:

This paper introduces the new LACGC model to sustain the academic and non-academic business to future educational and organizational generations. The consumption of knowledge and the creation of new knowledge is a strength and focal interest of all academics and Non-academic organizations. Implementing newly created knowledge sustains the businesses to the next generation with growth without detriment. Existing models like the Scholar-practitioner model and Organization knowledge creation models focus specifically on academic or non-academic, not both. LACGC model can be used for both Academic and Non-academic at the domestic or international level. Researchers and scholars play a substantial role in finding literature and practice gaps in academic and non-academic disciplines. LACGC model has unrestricted the number of recurrences because the Consumption, Creation, and implementation of new ideas, disciplines, systems, and knowledge is a never-ending process and must continue from one generation to the next.

Keywords: academics, consumption, creation, generations, non-academics, research, sustainability

Procedia PDF Downloads 172
7674 ParkedGuard: An Efficient and Accurate Parked Domain Detection System Using Graphical Locality Analysis and Coarse-To-Fine Strategy

Authors: Chia-Min Lai, Wan-Ching Lin, Hahn-Ming Lee, Ching-Hao Mao

Abstract:

As world wild internet has non-stop developments, making profit by lending registered domain names emerges as a new business in recent years. Unfortunately, the larger the market scale of domain lending service becomes, the riskier that there exist malicious behaviors or malwares hiding behind parked domains will be. Also, previous work for differentiating parked domain suffers two main defects: 1) too much data-collecting effort and CPU latency needed for features engineering and 2) ineffectiveness when detecting parked domains containing external links that are usually abused by hackers, e.g., drive-by download attack. Aiming for alleviating above defects without sacrificing practical usability, this paper proposes ParkedGuard as an efficient and accurate parked domain detector. Several scripting behavioral features were analyzed, while those with special statistical significance are adopted in ParkedGuard to make feature engineering much more cost-efficient. On the other hand, finding memberships between external links and parked domains was modeled as a graph mining problem, and a coarse-to-fine strategy was elaborately designed by leverage the graphical locality such that ParkedGuard outperforms the state-of-the-art in terms of both recall and precision rates.

Keywords: coarse-to-fine strategy, domain parking service, graphical locality analysis, parked domain

Procedia PDF Downloads 392
7673 Making a Difference in a Crisis: How the 24-Hour Surgical Ambulatory Assessment Unit Transformed Emergency Care during COVID-19

Authors: Bindhiya Thomas, Rehana Hafeez

Abstract:

Background: The Surgical Ambulatory Unit (SAU) also known as the Same Day Emergency Care (SDEC) is an established part of many hospitals providing same day emergency care service to surgical patients who would have otherwise required admission through the A&E. Prior to Covid, the SAU was functioning as a 12-hour service, but during the Covid crisis this service was transformed to a 24 hour functioning Surgical Ambulatory Assessment unit (SAAU). We studied the effects that this change brought about in-patient care in our hospital. Objective: The objective of the study was to assess the impact of a 24-hour Surgical Ambulatory Assessment unit on patient care during the time of Covid, in particular its role in freeing A&E capacity and delivering effective patient care. Methods: We collected two sets of data retrospectively. The first set was collected over a 6-week period when the SAU was functioning at the Princess Royal University Hospital. On March 23rd, 2020, the SAU was transformed into a 24-hour SAAU. Following this transformation, a second set of patient data was collected over a period of 6 weeks. A comparison was made between data collected from when the hospital had a 12-hour Surgical Ambulatory unit and later when it was transformed into a 24-hour facility. Its effects on the change in the number of patients breaching the four hour waiting period and the number of emergency surgical admissions. Results: The 24-hour Surgical Ambulatory Assessment unit brought significant reductions in the number of patients breaching the waiting period of 4 hours in A&E from 44% during the period of the 12-hour Surgical Ambulatory care facility to 0% from when the 24-hour Surgical Ambulatory Assessment Unit was established. A 28% reduction was also seen in the number of surgical patients' admissions from A&E. Conclusions: The 24-hour SAAU was found to have a profound positive impact on emergency care of surgical patients. Especially during the Covid crisis, it played a crucial role in providing not only effective and accessible patient care but also in reducing the A&E workload and admissions. It thus proved to be a strategic tool that helped to deal with the immense workload in emergency care during the Covid crisis and helped free much needed headspace at a time of uncertainty for the A&E to better configure their services. If sustained, the 24-hour SAAU could be relied on to augment the NHS emergency services in the future, especially in the event of another crisis.

Keywords: Princess Royal University Hospital, surgical ambulatory assessment unit, surgical ambulatory unit, same day emergency care

Procedia PDF Downloads 144
7672 Scheduling Jobs with Stochastic Processing Times or Due Dates on a Server to Minimize the Number of Tardy Jobs

Authors: H. M. Soroush

Abstract:

The problem of scheduling products and services for on-time deliveries is of paramount importance in today’s competitive environments. It arises in many manufacturing and service organizations where it is desirable to complete jobs (products or services) with different weights (penalties) on or before their due dates. In such environments, schedules should frequently decide whether to schedule a job based on its processing time, due-date, and the penalty for tardy delivery to improve the system performance. For example, it is common to measure the weighted number of late jobs or the percentage of on-time shipments to evaluate the performance of a semiconductor production facility or an automobile assembly line. In this paper, we address the problem of scheduling a set of jobs on a server where processing times or due-dates of jobs are random variables and fixed weights (penalties) are imposed on the jobs’ late deliveries. The goal is to find the schedule that minimizes the expected weighted number of tardy jobs. The problem is NP-hard to solve; however, we explore three scenarios of the problem wherein: (i) both processing times and due-dates are stochastic; (ii) processing times are stochastic and due-dates are deterministic; and (iii) processing times are deterministic and due-dates are stochastic. We prove that special cases of these scenarios are solvable optimally in polynomial time, and introduce efficient heuristic methods for the general cases. Our computational results show that the heuristics perform well in yielding either optimal or near optimal sequences. The results also demonstrate that the stochasticity of processing times or due-dates can affect scheduling decisions. Moreover, the proposed problem is general in the sense that its special cases reduce to some new and some classical stochastic single machine models.

Keywords: number of late jobs, scheduling, single server, stochastic

Procedia PDF Downloads 480
7671 Soap Film Enneper Minimal Surface Model

Authors: Yee Hooi Min, Mohdnasir Abdul Hadi

Abstract:

Tensioned membrane structure in the form of Enneper minimal surface can be considered as a sustainable development for the green environment and technology, it also can be used to support the effectiveness used of energy and the structure. Soap film in the form of Enneper minimal surface model has been studied. The combination of shape and internal forces for the purpose of stiffness and strength is an important feature of membrane surface. For this purpose, form-finding using soap film model has been carried out for Enneper minimal surface models with variables u=v=0.6 and u=v=1.0. Enneper soap film models with variables u=v=0.6 and u=v=1.0 provides an alternative choice for structural engineers to consider the tensioned membrane structure in the form of Enneper minimal surface applied in the building industry. It is expected to become an alternative building material to be considered by the designer.

Keywords: Enneper, minimal surface, soap film, tensioned membrane structure

Procedia PDF Downloads 523
7670 University Students’ Perception on Public Transit in Dhaka City

Authors: Mosabbir Pasha, Ijaj Mahmud Chowdhury, M. A. Afrahim Bhuiyann

Abstract:

With the increasing population and intensive land use, huge traffic demand is generating worldwide both in developing and developed countries. As a developing country, Bangladesh is also facing the same problem in recent years by producing huge numbers of daily trips. As a matter of fact, extensive traffic demand is increasing day by day. Also, transport system in Dhaka is heterogeneous, reflecting the heterogeneity in the socio-economic and land use patterns. As a matter of fact, trips produced here are for different purposes such as work, business, educational etc. Due to the significant concentration of educational institutions a large share of the trips are generated by educational purpose. And one of the major percentages of educational trips is produced by university going students and most of them are travelled by car, bus, train, taxi, rickshaw etc. The aim of the study was to find out the university students’ perception on public transit ridership. A survey was conducted among 330 students from eight different universities. It was found out that 26% of the trips produced by university going students are travelled by public bus service and only 5% are by train. Percentage of car share is 16% and 12% of the trips are travelled by private taxi. From the study, it has been found that more than 42 percent student’s family resides outside of Dhaka, eventually they prefer bus instead of other options. Again those who chose to walk most of the time, of them, over 40 percent students’ family reside outside of Dhaka and of them over 85 percent students have a tendency to live in a mess. They generally choose a neighboring location to their respective university so that they can reach their destination by walk. On the other hand, those who travel by car 80 percent of their family reside inside Dhaka. The study also revealed that the most important reason that restricts students not to use public transit is poor service. Negative attitudes such as discomfort, uneasiness in using public transit also reduces the usage of public transit. The poor waiting area is another major cause of not using public transit. Insufficient security also plays a significant role in not using public transit. On the contrary, the fare is not a problem for students those who use public transit as a mode of transportation. Students also think stations are not far away from their home or institution and they do not need to wait long for the buses or trains. It was also found accessibility to public transit is moderate.

Keywords: traffic demand, fare, poor service, public transit ridership

Procedia PDF Downloads 246
7669 Interpersonal Competence Related to the Practice Learning of Occupational Therapy Students in Hong Kong

Authors: Lik Hang Gary Wong

Abstract:

Background: Practice learning is crucial for preparing the healthcare profession to meet the real challenge upon graduation. Students are required to demonstrate their competence in managing interpersonal challenges, such as teamwork with other professionals and communicating well with the service users, during the placement. Such competence precedes clinical practice, and it may eventually affect students' actual performance in a clinical context. Unfortunately, there were limited studies investigating how such competence affects students' performance in practice learning. Objectives: The aim of this study is to investigate how self-rated interpersonal competence affects students' actual performance during clinical placement. Methods: 40 occupational therapy students from Hong Kong were recruited in this study. Prior to the clinical placement (level two or above), they completed an online survey that included the Interpersonal Communication Competence Scale (ICCS) measuring self-perceived competence in interpersonal communication. Near the end of their placement, the clinical educator rated students’ performance with the Student Practice Evaluation Form - Revised edition (SPEF-R). The SPEF-R measures the eight core competency domains required for an entry-level occupational therapist. This study adopted the cross-sectional observational design. Pearson correlation and multiple regression are conducted to examine the relationship between students' interpersonal communication competence and their actual performance in clinical placement. Results: The ICCS total scores were significantly correlated with all the SPEF-R domains, with correlation coefficient r ranging from 0.39 to 0.51. The strongest association was found with the co-worker communication domain (r = 0.51, p < 0.01), followed by the information gathering domain (r = 0.50, p < 0.01). Regarding the ICCS total scores as the independent variable and the rating in various SPEF-R domains as the dependent variables in the multiple regression analyses, the interpersonal competence measures were identified as a significant predictor of the co-worker communication (R² = 0.33, β = 0.014, SE = 0.006, p = 0.026), information gathering (R² = 0.27, β = 0.018, SE = 0.007, p = 0.011), and service provision (R² = 0.17, β = 0.017, SE = 0.007, p = 0.020). Moreover, some specific communication skills appeared to be especially important to clinical practice. For example, immediacy, which means whether the students were readily approachable on all social occasions, correlated with all the SPEF-R domains, with r-values ranging from 0.45 to 0.33. Other sub-skills, such as empathy, interaction management, and supportiveness, were also found to be significantly correlated to most of the SPEF-R domains. Meanwhile, the ICCS scores correlated differently with the co-worker communication domain (r = 0.51, p < 0.01) and the communication with the service user domain (r = 0.39, p < 0.05). It suggested that different communication skill sets would be required for different interpersonal contexts within the workplace. Conclusion: Students' self-perceived interpersonal communication competence could predict their actual performance during clinical placement. Moreover, some specific communication skills were more important to the co-worker communication but not to the daily interaction with the service users. There were implications on how to better prepare the students to meet the future challenge upon graduation.

Keywords: interpersonal competence, clinical education, healthcare professional education, occupational therapy, occupational therapy students

Procedia PDF Downloads 49
7668 Patients’ Trust in Health Care Systems

Authors: Dilara Usta, Fatos Korkmaz

Abstract:

Background: Individuals who utilise health services maintain relationships with health professionals, insurers and institutions. The nature of these relationships requires service receivers to have trust in the service providers because maintaining health services without reciprocal trust is very difficult. Therefore, individual evaluations of trust within the scope of health services have become increasingly important. Objective: To investigate patients’ trust in the health-care system and their relevant socio-demographical characteristics. Methods: This research was conducted using a descriptive design which included 493 literate patients aged 18-65 years who were hospitalised for a minimum of two days at public university and training&research hospitals in Ankara, Turkey. Patients’ trust in health-care professionals, insurers, and institutions were investigated. Data were collected using a demographic questionnaire and the Multidimensional Trust in Health-Care Systems Scale between September 2015 and April 2016. Results: The participants’ mean age was 47.7±13.1; 70% had a moderate income and 69% had a prior hospitalisation and 63.5% of the patients were satisfied with the health-care services. The mean Multidimensional Trust in Health-Care Systems Scale score for the sample was 61.5±8.3; the provider subscale had a mean of 38.1±5, the insurers subscale had a mean of 12.9±3.7, and institutions subscale had a mean of 10.6±1.9. Conclusion: Patients’ level of trust in the health-care system was above average and the trust level of the patients with higher educational and socio-economic levels was lower compared to the other patients. Health-care professionals should raise awareness about the significance of trust in the health-care system.

Keywords: delivery of health care, health care system, nursing, patients, trust

Procedia PDF Downloads 347
7667 Modeling Biomass and Biodiversity across Environmental and Management Gradients in Temperate Grasslands with Deep Learning and Sentinel-1 and -2

Authors: Javier Muro, Anja Linstadter, Florian Manner, Lisa Schwarz, Stephan Wollauer, Paul Magdon, Gohar Ghazaryan, Olena Dubovyk

Abstract:

Monitoring the trade-off between biomass production and biodiversity in grasslands is critical to evaluate the effects of management practices across environmental gradients. New generations of remote sensing sensors and machine learning approaches can model grasslands’ characteristics with varying accuracies. However, studies often fail to cover a sufficiently broad range of environmental conditions, and evidence suggests that prediction models might be case specific. In this study, biomass production and biodiversity indices (species richness and Fishers’ α) are modeled in 150 grassland plots for three sites across Germany. These sites represent a North-South gradient and are characterized by distinct soil types, topographic properties, climatic conditions, and management intensities. Predictors used are derived from Sentinel-1 & 2 and a set of topoedaphic variables. The transferability of the models is tested by training and validating at different sites. The performance of feed-forward deep neural networks (DNN) is compared to a random forest algorithm. While biomass predictions across gradients and sites were acceptable (r2 0.5), predictions of biodiversity indices were poor (r2 0.14). DNN showed higher generalization capacity than random forest when predicting biomass across gradients and sites (relative root mean squared error of 0.5 for DNN vs. 0.85 for random forest). DNN also achieved high performance when using the Sentinel-2 surface reflectance data rather than different combinations of spectral indices, Sentinel-1 data, or topoedaphic variables, simplifying dimensionality. This study demonstrates the necessity of training biomass and biodiversity models using a broad range of environmental conditions and ensuring spatial independence to have realistic and transferable models where plot level information can be upscaled to landscape scale.

Keywords: ecosystem services, grassland management, machine learning, remote sensing

Procedia PDF Downloads 199
7666 Micro Celebrities in Social Media Instagram and Their Personal Influence in Business Perspective

Authors: Yoga Maulana Putra, Herry Hudrasyah

Abstract:

The Internet has now become an important part of human life; it can be accessed through a computer or even a smartphone almost anywhere and anytime. The Internet has created many social media such as Facebook, Twitter, and Instagram. Instagram has been acquired by Facebook in 2012. Since then, Instagram is growing fast. And now, Instagram is transforming from photo-sharing social media into business tools. As the result, some new behavior has been discovered. Some of Instagram user is becoming popular. These people also being called minor celebrity and they are also being used as marketing tools by many companies to influencing or promoting their product or service. This minor celebrity is existing because of their behavior in using Instagram. The company is using the personal influence of the minor celebrity to promoting and influencing their product or service, and the minor celebrity gets paid as much as their rate card. And their rate card based on their followers and insight. This research is using a qualitative method. An interview is being done to 6 minor celebrities from many different categories such as photographer, travel blogger, lifestyle, food blogger, fashion, and healthcare. Theory of reasoned behavior is being used as the grounded theory to discover the reason for their behavior and personal influence to describe their way to influencing people. The result of the interview is most of the minor celebrities is influenced by their friend’s circle in the process of using Instagram. They also had a different way to use their personal influence to affect their followers when the company employs them.

Keywords: humanities and social sciences, Instagram, minor celebrity, social media

Procedia PDF Downloads 146
7665 Statistical Comparison of Ensemble Based Storm Surge Forecasting Models

Authors: Amin Salighehdar, Ziwen Ye, Mingzhe Liu, Ionut Florescu, Alan F. Blumberg

Abstract:

Storm surge is an abnormal water level caused by a storm. Accurate prediction of a storm surge is a challenging problem. Researchers developed various ensemble modeling techniques to combine several individual forecasts to produce an overall presumably better forecast. There exist some simple ensemble modeling techniques in literature. For instance, Model Output Statistics (MOS), and running mean-bias removal are widely used techniques in storm surge prediction domain. However, these methods have some drawbacks. For instance, MOS is based on multiple linear regression and it needs a long period of training data. To overcome the shortcomings of these simple methods, researchers propose some advanced methods. For instance, ENSURF (Ensemble SURge Forecast) is a multi-model application for sea level forecast. This application creates a better forecast of sea level using a combination of several instances of the Bayesian Model Averaging (BMA). An ensemble dressing method is based on identifying best member forecast and using it for prediction. Our contribution in this paper can be summarized as follows. First, we investigate whether the ensemble models perform better than any single forecast. Therefore, we need to identify the single best forecast. We present a methodology based on a simple Bayesian selection method to select the best single forecast. Second, we present several new and simple ways to construct ensemble models. We use correlation and standard deviation as weights in combining different forecast models. Third, we use these ensembles and compare with several existing models in literature to forecast storm surge level. We then investigate whether developing a complex ensemble model is indeed needed. To achieve this goal, we use a simple average (one of the simplest and widely used ensemble model) as benchmark. Predicting the peak level of Surge during a storm as well as the precise time at which this peak level takes place is crucial, thus we develop a statistical platform to compare the performance of various ensemble methods. This statistical analysis is based on root mean square error of the ensemble forecast during the testing period and on the magnitude and timing of the forecasted peak surge compared to the actual time and peak. In this work, we analyze four hurricanes: hurricanes Irene and Lee in 2011, hurricane Sandy in 2012, and hurricane Joaquin in 2015. Since hurricane Irene developed at the end of August 2011 and hurricane Lee started just after Irene at the beginning of September 2011, in this study we consider them as a single contiguous hurricane event. The data set used for this study is generated by the New York Harbor Observing and Prediction System (NYHOPS). We find that even the simplest possible way of creating an ensemble produces results superior to any single forecast. We also show that the ensemble models we propose generally have better performance compared to the simple average ensemble technique.

Keywords: Bayesian learning, ensemble model, statistical analysis, storm surge prediction

Procedia PDF Downloads 292