Search results for: information technology
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 16143

Search results for: information technology

12243 Improving the Bioprocess Phenotype of Chinese Hamster Ovary Cells Using CRISPR/Cas9 and Sponge Decoy Mediated MiRNA Knockdowns

Authors: Kevin Kellner, Nga Lao, Orla Coleman, Paula Meleady, Niall Barron

Abstract:

Chinese Hamster Ovary (CHO) cells are the prominent cell line used in biopharmaceutical production. To improve yields and find beneficial bioprocess phenotypes genetic engineering plays an essential role in recent research. The miR-23 cluster, specifically miR-24 and miR-27, was first identified as differentially expressed during hypothermic conditions suggesting a role in proliferation and productivity in CHO cells. In this study, we used sponge decoy technology to stably deplete the miRNA expression of the cluster. Furthermore, we implemented the CRISPR/Cas9 system to knockdown miRNA expression. Sponge constructs were designed for an imperfect binding of the miRNA target, protecting from RISC mediated cleavage. GuideRNAs for the CRISPR/Cas9 system were designed to target the seed region of the miRNA. The expression of mature miRNA and precursor were confirmed using RT-qPCR. For both approaches stable expressing mixed populations were generated and characterised in batch cultures. It was shown, that CRISPR/Cas9 can be implemented in CHO cells with achieving high knockdown efficacy of every single member of the cluster. Targeting of one miRNA member showed that its genomic paralog is successfully targeted as well. The stable depletion of miR-24 using CRISPR/Cas9 showed increased growth and specific productivity in a CHO-K1 mAb expressing cell line. This phenotype was further characterized using quantitative label-free LC-MS/MS showing 186 proteins differently expressed with 19 involved in proliferation and 26 involved in protein folding/translation. Targeting miR-27 in the same cell line showed increased viability in late stages of the culture compared to the control. To evaluate the phenotype in an industry relevant cell line; the miR-23 cluster, miR-24 and miR-27 were stably depleted in a Fc fusion CHO-S cell line which showed increased batch titers up to 1.5-fold. In this work, we highlighted that the stable depletion of the miR-23 cluster and its members can improve the bioprocess phenotype concerning growth and productivity in two different cell lines. Furthermore, we showed that using CRISPR/Cas9 is comparable to the traditional sponge decoy technology.

Keywords: Chinese Hamster ovary cells, CRISPR/Cas9, microRNAs, sponge decoy technology

Procedia PDF Downloads 180
12242 Increasing the Efficiency of the Biomass Gasification Technology with Using the Organic Rankin Cycle

Authors: Jaroslav Frantík, Jan Najser

Abstract:

This article deals with increasing the energy efficiency of a plant in terms of optimizing the process. The European Union is striving to achieve the climate-energy package in the area increasing of energy efficiency. The goal of energy efficiency is to reduce primary energy consumption by 20% within the EU until 2020. The objective of saving energy consumption in the Czech Republic was set at 47.84 PJ (13.29 TWh). For reducing electricity consumption, it is possible to choose: a) mandatory increasing of energy efficiency, b) alternative scheme, c) combination of both actions. The Czech Republic has chosen for reducing electricity consumption using-alternative scheme. The presentation is focused on the proposal of a technological unit dealing with the gasification process of processing of biomass with an increase of power in the output. The synthesis gas after gasification of biomass is used as fuel in a cogeneration process of reciprocating internal combustion engine with the classic production of heat and electricity. Subsequently, there is an explanation of the ORC system dealing with the conversion of waste heat to electricity with the using closed cycle of the steam process with organic medium. The arising electricity is distributed to the power grid as a further energy source, or it is used for needs of the partial coverage of the technological unit. Furthermore, there is a presented schematic description of the technology with the identification of energy flows starting from the biomass treatment by drying, through its conversion to gaseous fuel, producing of electricity and utilize of thermal energy with minimizing losses. It has been found that using of ORC system increased the efficiency of the produced electricity by 7.5%.

Keywords: biomass, efficiency, gasification, ORC system

Procedia PDF Downloads 202
12241 Meat Products Demand in Oyo West Local Government: An Application of Almost Ideal Demand System (LA/AIDS)

Authors: B. A. Adeniyi, S. A. Daud, O. Amao

Abstract:

The study investigates consumer demand for meat products in Oyo West Local Government using linear approximate almost ideal demand system (LA/AIDS). Questions that were addressed by the study include: first, what is the type and quantity of meat products available to the household and their demand pattern? Second is the investigation of the factors that affect meat products demand pattern and proportion of income that is spent on them. For the above purpose cross-sectional data were collected from 156 households of the study area and analyzed to reveal the functional relationship between meat products consumption and some socio-economic variables of the household. Results indicated that per capita meat consumption increased as household income and education increased but decreased with age. It was also found that male tend to consume more meat products than their female counterparts and that increase in household size will first increased per caput meat consumption but later decreased it. Price also tends to greatly influence the demand pattern of meat products. The results of elasticity computed from the results of regression analysis revealed that own price elasticity for all meat products were negative which indicated that they were normal products while cross and expenditure elasticity were positive which further confirmed that meat products were normal and substitute products. This study therefore concludes that the relevance of these variables imposed a great challenge to the policy makers and the government, in the sense that more cost effective methods of meat production technology have to be devised in other to make consumption of meat products more affordable.

Keywords: meat products, consumption, animal production, technology

Procedia PDF Downloads 232
12240 Machine Learning Model to Predict TB Bacteria-Resistant Drugs from TB Isolates

Authors: Rosa Tsegaye Aga, Xuan Jiang, Pavel Vazquez Faci, Siqing Liu, Simon Rayner, Endalkachew Alemu, Markos Abebe

Abstract:

Tuberculosis (TB) is a major cause of disease globally. In most cases, TB is treatable and curable, but only with the proper treatment. There is a time when drug-resistant TB occurs when bacteria become resistant to the drugs that are used to treat TB. Current strategies to identify drug-resistant TB bacteria are laboratory-based, and it takes a longer time to identify the drug-resistant bacteria and treat the patient accordingly. But machine learning (ML) and data science approaches can offer new approaches to the problem. In this study, we propose to develop an ML-based model to predict the antibiotic resistance phenotypes of TB isolates in minutes and give the right treatment to the patient immediately. The study has been using the whole genome sequence (WGS) of TB isolates as training data that have been extracted from the NCBI repository and contain different countries’ samples to build the ML models. The reason that different countries’ samples have been included is to generalize the large group of TB isolates from different regions in the world. This supports the model to train different behaviors of the TB bacteria and makes the model robust. The model training has been considering three pieces of information that have been extracted from the WGS data to train the model. These are all variants that have been found within the candidate genes (F1), predetermined resistance-associated variants (F2), and only resistance-associated gene information for the particular drug. Two major datasets have been constructed using these three information. F1 and F2 information have been considered as two independent datasets, and the third information is used as a class to label the two datasets. Five machine learning algorithms have been considered to train the model. These are Support Vector Machine (SVM), Random forest (RF), Logistic regression (LR), Gradient Boosting, and Ada boost algorithms. The models have been trained on the datasets F1, F2, and F1F2 that is the F1 and the F2 dataset merged. Additionally, an ensemble approach has been used to train the model. The ensemble approach has been considered to run F1 and F2 datasets on gradient boosting algorithm and use the output as one dataset that is called F1F2 ensemble dataset and train a model using this dataset on the five algorithms. As the experiment shows, the ensemble approach model that has been trained on the Gradient Boosting algorithm outperformed the rest of the models. In conclusion, this study suggests the ensemble approach, that is, the RF + Gradient boosting model, to predict the antibiotic resistance phenotypes of TB isolates by outperforming the rest of the models.

Keywords: machine learning, MTB, WGS, drug resistant TB

Procedia PDF Downloads 32
12239 A Vision-Based Early Warning System to Prevent Elephant-Train Collisions

Authors: Shanaka Gunasekara, Maleen Jayasuriya, Nalin Harischandra, Lilantha Samaranayake, Gamini Dissanayake

Abstract:

One serious facet of the worsening Human-Elephant conflict (HEC) in nations such as Sri Lanka involves elephant-train collisions. Endangered Asian elephants are maimed or killed during such accidents, which also often result in orphaned or disabled elephants, contributing to the phenomenon of lone elephants. These lone elephants are found to be more likely to attack villages and showcase aggressive behaviour, which further exacerbates the overall HEC. Furthermore, Railway Services incur significant financial losses and disruptions to services annually due to such accidents. Most elephant-train collisions occur due to a lack of adequate reaction time. This is due to the significant stopping distance requirements of trains, as the full braking force needs to be avoided to minimise the risk of derailment. Thus, poor driver visibility at sharp turns, nighttime operation, and poor weather conditions are often contributing factors to this problem. Initial investigations also indicate that most collisions occur in localised “hotspots” where elephant pathways/corridors intersect with railway tracks that border grazing land and watering holes. Taking these factors into consideration, this work proposes the leveraging of recent developments in Convolutional Neural Network (CNN) technology to detect elephants using an RGB/infrared capable camera around known hotspots along the railway track. The CNN was trained using a curated dataset of elephants collected on field visits to elephant sanctuaries and wildlife parks in Sri Lanka. With this vision-based detection system at its core, a prototype unit of an early warning system was designed and tested. This weatherised and waterproofed unit consists of a Reolink security camera which provides a wide field of view and range, an Nvidia Jetson Xavier computing unit, a rechargeable battery, and a solar panel for self-sufficient functioning. The prototype unit was designed to be a low-cost, low-power and small footprint device that can be mounted on infrastructures such as poles or trees. If an elephant is detected, an early warning message is communicated to the train driver using the GSM network. A mobile app for this purpose was also designed to ensure that the warning is clearly communicated. A centralized control station manages and communicates all information through the train station network to ensure coordination among important stakeholders. Initial results indicate that detection accuracy is sufficient under varying lighting situations, provided comprehensive training datasets that represent a wide range of challenging conditions are available. The overall hardware prototype was shown to be robust and reliable. We envision a network of such units may help contribute to reducing the problem of elephant-train collisions and has the potential to act as an important surveillance mechanism in dealing with the broader issue of human-elephant conflicts.

Keywords: computer vision, deep learning, human-elephant conflict, wildlife early warning technology

Procedia PDF Downloads 211
12238 An Experimental Study of Low Concentration CO₂ Capture from Regenerative Thermal Oxidation Tail Gas in Rotating Packed Bed

Authors: Dang HuynhMinhTam, Kuang-Cong Lu, Yi-Hung Chen, Zhung-Yu Lin, Cheng-Siang Cheng

Abstract:

Carbon capture, utilization, and storage (CCUS) technology become a predominant technique to mitigate carbon dioxide and achieve net-zero emissions goals. This research targets to continuously capture the low concentration CO₂ from the tail gas of the regenerative thermal oxidizer (RTO) in the high technology industry. A rotating packed bed (RPB) reactor is investigated to capture the efficiency of CO₂ using a mixture of NaOH/Na₂CO₃ solutions to simulate the real absorbed solution. On a lab scale, semi-batch experiments of continuous gas flow and circulating absorbent solution are conducted to find the optimal parameters and are then examined in a continuous operation. In the semi-batch tests, the carbon capture efficiency and pH variation in the conditions of a low concentration CO₂ (about 1.13 vol%), the NaOH concentration of 1 wt% or 2 wt% mixed with 14 wt% Na₂CO₃, the rotating speed (600, 900, 1200 rpm), the gas-liquid ratio (100, 200, and 400), and the temperature of absorbent solution of 40 ºC are studied. The CO₂ capture efficiency significantly increases with higher rotating speed and smaller gas-liquid ratio, respectively, while the difference between the NaOH concentration of 1 wt% and 2 wt% is relatively small. The maximum capture efficiency is close to 80% in the conditions of the NaOH concentration of 1 wt%, the G/L ratio of 100, and the rotating speed of 1200 rpm within the first 5 minutes. Furthermore, the continuous operation based on similar conditions also demonstrates the steady efficiency of the carbon capture of around 80%.

Keywords: carbon dioxide capture, regenerative thermal oxidizer, rotating packed bed, sodium hydroxide

Procedia PDF Downloads 40
12237 Fusion Neutron Generator Dosimetry and Applications for Medical, Security, and Industry

Authors: Kaouther Bergaui, Nafaa Reguigui, Charles Gary

Abstract:

Characterization and the applications of deuterium-deuterium (DD) neutron generator developed by Adelphie technology and acquired by the National Centre of Nuclear Science and Technology (NCNST) were presented in this work. We study the performance of the neutron generator in terms of neutron yield, production efficiency, and the ionic current as a function of the acceleration voltage at various RF powers. We provide the design and optimization of the PGNAA chamber and thus give insight into the capabilities of the planned PGNAA facility. Additional non-destructive techniques were studied employing the DD neutron generator, such as PGNAA and neutron radiography: The PGNAA is used for determining the concentration of 10B in Si and SiO2 matrices by using a germanium detector HPGe and the results obtained are compared with PGNAA system using a Sodium Iodide detector (NaI (Tl)); Neutron radiography facility was tested and simulated, using a camera device CCD and simulated by the Monte Carlo code; and the explosive detection system (EDS) also simulated using the Monte Carlo code. The study allows us to show that the new models of DD neutron generators are feasible and that superior-quality neutron beams could be produced and used for various applications. The feasibility of Boron neutron capture therapy (BNCT) for cancer treatment using a neutron generator was assessed by optimizing Beam Shaping Assembly (BSA) on a phantom using Monte-Carlo (MCNP6) simulations.

Keywords: neutron generator deuterium-deuterium, Monte Carlo method, radiation, neutron flux, neutron activation analysis, born, neutron radiography, explosive detection, BNCT

Procedia PDF Downloads 174
12236 Machine Learning-Based Techniques for Detecting and Mitigating Cyber-attacks on Automatic Generation Control in Smart Grids

Authors: Sami M. Alshareef

Abstract:

The rapid growth of smart grid technology has brought significant advancements to the power industry. However, with the increasing interconnectivity and reliance on information and communication technologies, smart grids have become vulnerable to cyber-attacks, posing significant threats to the reliable operation of power systems. Among the critical components of smart grids, the Automatic Generation Control (AGC) system plays a vital role in maintaining the balance between generation and load demand. Therefore, protecting the AGC system from cyber threats is of paramount importance to maintain grid stability and prevent disruptions. Traditional security measures often fall short in addressing sophisticated and evolving cyber threats, necessitating the exploration of innovative approaches. Machine learning, with its ability to analyze vast amounts of data and learn patterns, has emerged as a promising solution to enhance AGC system security. Therefore, this research proposal aims to address the challenges associated with detecting and mitigating cyber-attacks on AGC in smart grids by leveraging machine learning techniques on automatic generation control of two-area power systems. By utilizing historical data, the proposed system will learn the normal behavior patterns of AGC and identify deviations caused by cyber-attacks. Once an attack is detected, appropriate mitigation strategies will be employed to safeguard the AGC system. The outcomes of this research will provide power system operators and administrators with valuable insights into the vulnerabilities of AGC systems in smart grids and offer practical solutions to enhance their cyber resilience.

Keywords: machine learning, cyber-attacks, automatic generation control, smart grid

Procedia PDF Downloads 71
12235 Accelerating Personalization Using Digital Tools to Drive Circular Fashion

Authors: Shamini Dhana, G. Subrahmanya VRK Rao

Abstract:

The fashion industry is advancing towards a mindset of zero waste, personalization, creativity, and circularity. The trend of upcycling clothing and materials into personalized fashion is being demanded by the next generation. There is a need for a digital tool to accelerate the process towards mass customization. Dhana’s D/Sphere fashion technology platform uses digital tools to accelerate upcycling. In essence, advanced fashion garments can be designed and developed via reuse, repurposing, recreating activities, and using existing fabric and circulating materials. The D/Sphere platform has the following objectives: to provide (1) An opportunity to develop modern fashion using existing, finished materials and clothing without chemicals or water consumption; (2) The potential for an everyday customer and designer to use the medium of fashion for creative expression; (3) A solution to address the global textile waste generated by pre- and post-consumer fashion; (4) A solution to reduce carbon emissions, water, and energy consumption with the participation of all stakeholders; (5) An opportunity for brands, manufacturers, retailers to work towards zero-waste designs and as an alternative revenue stream. Other benefits of this alternative approach include sustainability metrics, trend prediction, facilitation of disassembly and remanufacture deep learning, and hyperheuristics for high accuracy. A design tool for mass personalization and customization utilizing existing circulating materials and deadstock, targeted to fashion stakeholders will lower environmental costs, increase revenues through up to date upcycled apparel, produce less textile waste during the cut-sew-stitch process, and provide a real design solution for the end customer to be part of circular fashion. The broader impact of this technology will result in a different mindset to circular fashion, increase the value of the product through multiple life cycles, find alternatives towards zero waste, and reduce the textile waste that ends up in landfills. This technology platform will be of interest to brands and companies that have the responsibility to reduce their environmental impact and contribution to climate change as it pertains to the fashion and apparel industry. Today, over 70% of the $3 trillion fashion and apparel industry ends up in landfills. To this extent, the industry needs such alternative techniques to both address global textile waste as well as provide an opportunity to include all stakeholders and drive circular fashion with new personalized products. This type of modern systems thinking is currently being explored around the world by the private sector, organizations, research institutions, and governments. This technological innovation using digital tools has the potential to revolutionize the way we look at communication, capabilities, and collaborative opportunities amongst stakeholders in the development of new personalized and customized products, as well as its positive impacts on society, our environment, and global climate change.

Keywords: circular fashion, deep learning, digital technology platform, personalization

Procedia PDF Downloads 44
12234 Females’ Usage Patterns of Information and Communication Technologies (ICTs) in the Vhembe District, South Africa

Authors: Fulufhelo Oscar Maphiri-Makananise

Abstract:

The main purpose of this paper is to explore and provide substantiated evidence based on the usage patterns of Information and Communication Technologies (ICTs) by females in the Vhembe District in Limpopo-Province, South Africa. The study presents a broader picture and understanding about the usage of ICTs from female’s perspective. The significance of this study stems from the need to discover the role, relevance and usage patterns of ICTs such as smartphones, computers, laptops, and iPods, internet and social networking sites among females following the trends of new media technologies in the society. The main objective of the study was to investigate the usability and accessibility of ICTs to empower the Vhembe District females in South Africa. The study used quantitative research method together with elements of qualitative research to determine the major ideas, perceptions and usage patterns of ICTs by females in the District. Data collection involved structured and self-administered questionnaire with both closed-ended and open-ended questions. Two groups of respondents participated in this study. Media Studies female students (n=50) at the University of Venda provided their ideas and perceptions about the usefulness and usage patterns of ICTs such as smartphones, internet and computers at the university level, while the second group were (n=50) Makhado comprehensive school learners who also provided their perceptions and ideas about the use of ICTs at the high school level. Also, the study provides a more balanced, accurate and rational results on the pertinent issues that concern the use of ICTs by females in the Vhembe District. The researcher also believes that the findings of the study are useful as a guideline and model for ICT intervention that work as an empowerment to women in South Africa. The study showed that the main purpose of using ICTs by females was to search information for writing assignments, conducting research, dating, exchanging ideas and networking with friends and relatives that are also members of social networking sites and maintaining existing friends in real life. The study further revealed that most females were using ICTs for social purposes and accessing the internet than entertaining themselves. The finding also indicated a high number of females that used ICTs for e-learning (62%) and social purposes (85%). Moreover, the study centred on providing strong insightful information on the females’ usage patterns and their perceptions of ICTs in the Vhembe district of Limpopo province.

Keywords: female users, information and communication technologies, internet, usage patterns

Procedia PDF Downloads 201
12233 A Model of Teacher Leadership in History Instruction

Authors: Poramatdha Chutimant

Abstract:

The objective of the research was to propose a model of teacher leadership in history instruction for utilization. Everett M. Rogers’ Diffusion of Innovations Theory is applied as theoretical framework. Qualitative method is to be used in the study, and the interview protocol used as an instrument to collect primary data from best practices who awarded by Office of National Education Commission (ONEC). Open-end questions will be used in interview protocol in order to gather the various data. Then, information according to international context of history instruction is the secondary data used to support in the summarizing process (Content Analysis). Dendrogram is a key to interpret and synthesize the primary data. Thus, secondary data comes as the supportive issue in explanation and elaboration. In-depth interview is to be used to collected information from seven experts in educational field. The focal point is to validate a draft model in term of future utilization finally.

Keywords: history study, nationalism, patriotism, responsible citizenship, teacher leadership

Procedia PDF Downloads 265
12232 Ontological Modeling Approach for Statistical Databases Publication in Linked Open Data

Authors: Bourama Mane, Ibrahima Fall, Mamadou Samba Camara, Alassane Bah

Abstract:

At the level of the National Statistical Institutes, there is a large volume of data which is generally in a format which conditions the method of publication of the information they contain. Each household or business data collection project includes a dissemination platform for its implementation. Thus, these dissemination methods previously used, do not promote rapid access to information and especially does not offer the option of being able to link data for in-depth processing. In this paper, we present an approach to modeling these data to publish them in a format intended for the Semantic Web. Our objective is to be able to publish all this data in a single platform and offer the option to link with other external data sources. An application of the approach will be made on data from major national surveys such as the one on employment, poverty, child labor and the general census of the population of Senegal.

Keywords: Semantic Web, linked open data, database, statistic

Procedia PDF Downloads 164
12231 Documentation Project on Decorated Wooden Coffins From Luxor, in the Cairo Museum

Authors: Hassan Mohmed, Mohamed Ismail, Aiman Rezk

Abstract:

Introduction: This project aims to document and preserve decorated wooden coffins which were discovered in Luxor by Egyptian mission at Luxor, (SR Numbers:2514,2519,2520,2521,5469).These decorated wooden coffins dates back to Egyptian New Kingdom period and has been transferred to the Cairo Museum, to be displayed at the museum. These decorated wooden coffins discovered in the cache-tomb of Bab el-gasus at Deir el-Bahari, Luxor. This site has been dictated for the burials of priests of Amun through 18th Dynasty the coffins owners held these titles, which are as follows: "the embalmer of the beautiful-house (the place of embalming)" and "the servant in the place of truth". Methodology: Methodology: The project objectives making such decorated wooden coffins more visible to visitors through the use of 3D reconstructed coffins and high resolution photos which describe the history of using the wooden coffins during the Ancient Egyptian history Especially, The Cairo Museum is going to exhibit decorated wooden coffins in New kingdom. The project goals is to document decorated wooden coffins and arrange an exhibition, where such decorated wooden coffins going to be displayed next to the Ramses 2nd coffin, This research focuses on the text analyses and the technology. Paleographic information found on these objects. Conclusion: The project shows the importance of using coffins in Ancient Egypt, and connecting their usage through Ancient Egyptian periods; the coffins had a unique Symbolized in ancient Egypt and connect the public with their kings. The Egyptian put coffins in their tombs that they hope to save their bodies’ afterlife. This research will be beneficial and useful for the heritage and ancient civilizations, Indeed this study will open a destination in order to know how to identify these collections and how to exhibit them commensurate with the natural of the ancient Egyptian history and heritage.

Keywords: archaeology, decorated wooden coffins, 3D digital tools for heritage management, museums

Procedia PDF Downloads 59
12230 Modeling the Human Harbor: An Equity Project in New York City, New York USA

Authors: Lauren B. Birney

Abstract:

The envisioned long-term outcome of this three-year research, and implementation plan is for 1) teachers and students to design and build their own computational models of real-world environmental-human health phenomena occurring within the context of the “Human Harbor” and 2) project researchers to evaluate the degree to which these integrated Computer Science (CS) education experiences in New York City (NYC) public school classrooms (PreK-12) impact students’ computational-technical skill development, job readiness, career motivations, and measurable abilities to understand, articulate, and solve the underlying phenomena at the center of their models. This effort builds on the partnership’s successes over the past eight years in developing a benchmark Model of restoration-based Science, Technology, Engineering, and Math (STEM) education for urban public schools and achieving relatively broad-based implementation in the nation’s largest public school system. The Billion Oyster Project Curriculum and Community Enterprise for Restoration Science (BOP-CCERS STEM + Computing) curriculum, teacher professional developments, and community engagement programs have reached more than 200 educators and 11,000 students at 124 schools, with 84 waterfront locations and Out of School of Time (OST) programs. The BOP-CCERS Partnership is poised to develop a more refined focus on integrating computer science across the STEM domains; teaching industry-aligned computational methods and tools; and explicitly preparing students from the city’s most under-resourced and underrepresented communities for upwardly mobile careers in NYC’s ever-expanding “digital economy,” in which jobs require computational thinking and an increasing percentage require discreet computer science technical skills. Project Objectives include the following: 1. Computational Thinking (CT) Integration: Integrate computational thinking core practices across existing middle/high school BOP-CCERS STEM curriculum as a means of scaffolding toward long term computer science and computational modeling outcomes. 2. Data Science and Data Analytics: Enabling Researchers to perform interviews with Teachers, students, community members, partners, stakeholders, and Science, Technology, Engineering, and Mathematics (STEM) industry Professionals. Collaborative analysis and data collection were also performed. As a centerpiece, the BOP-CCERS partnership will expand to include a dedicated computer science education partner. New York City Department of Education (NYCDOE), Computer Science for All (CS4ALL) NYC will serve as the dedicated Computer Science (CS) lead, advising the consortium on integration and curriculum development, working in tandem. The BOP-CCERS Model™ also validates that with appropriate application of technical infrastructure, intensive teacher professional developments, and curricular scaffolding, socially connected science learning can be mainstreamed in the nation’s largest urban public school system. This is evidenced and substantiated in the initial phases of BOP-CCERS™. The BOP-CCERS™ student curriculum and teacher professional development have been implemented in approximately 24% of NYC public middle schools, reaching more than 250 educators and 11,000 students directly. BOP-CCERS™ is a fully scalable and transferable educational model, adaptable to all American school districts. In all settings of the proposed Phase IV initiative, the primary beneficiary group will be underrepresented NYC public school students who live in high-poverty neighborhoods and are traditionally underrepresented in the STEM fields, including African Americans, Latinos, English language learners, and children from economically disadvantaged households. In particular, BOP-CCERS Phase IV will explicitly prepare underrepresented students for skilled positions within New York City’s expanding digital economy, computer science, computational information systems, and innovative technology sectors.

Keywords: computer science, data science, equity, diversity and inclusion, STEM education

Procedia PDF Downloads 42
12229 Information in Public Domain: How Far It Measures Government's Accountability

Authors: Sandip Mitra

Abstract:

Studies on Governance and Accountability has often stressed the need to release Data in public domain to increase transparency ,which otherwise act as an evidence of performance. However, inefficient handling, lack of capacity and the dynamics of transfers (especially fund transfers) are important issues which need appropriate attention. E-Governance alone can not serve as a measure of transparency as long as a comprehensive planning is instituted. Studies on Governance and public exposure has often triggered public opinion in favour or against any government. The root of the problem (especially in local governments) lies in the management of the governance. The participation of the people in the local government functioning, the networks within and outside the locality, synergy with various layers of Government are crucial in understanding the activities of any government. Unfortunately, data on such issues are not released in the public domain .If they are at all released , the extraction of information is often hindered for complicated designs. A Study has been undertaken with a few local Governments in India. The data has been analysed to substantiate the views.

Keywords: accountability, e-governance, transparency, local government

Procedia PDF Downloads 418
12228 Correlation Analysis between Sensory Processing Sensitivity (SPS), Meares-Irlen Syndrome (MIS) and Dyslexia

Authors: Kaaryn M. Cater

Abstract:

Students with sensory processing sensitivity (SPS), Meares-Irlen Syndrome (MIS) and dyslexia can become overwhelmed and struggle to thrive in traditional tertiary learning environments. An estimated 50% of tertiary students who disclose learning related issues are dyslexic. This study explores the relationship between SPS, MIS and dyslexia. Baseline measures will be analysed to establish any correlation between these three minority methods of information processing. SPS is an innate sensitivity trait found in 15-20% of the population and has been identified in over 100 species of animals. Humans with SPS are referred to as Highly Sensitive People (HSP) and the measure of HSP is a 27 point self-test known as the Highly Sensitive Person Scale (HSPS). A 2016 study conducted by the author established base-line data for HSP students in a tertiary institution in New Zealand. The results of the study showed that all participating HSP students believed the knowledge of SPS to be life-changing and useful in managing life and study, in addition, they believed that all tutors and in-coming students should be given information on SPS. MIS is a visual processing and perception disorder that is found in approximately 10% of the population and has a variety of symptoms including visual fatigue, headaches and nausea. One way to ease some of these symptoms is through the use of colored lenses or overlays. Dyslexia is a complex phonological based information processing variation present in approximately 10% of the population. An estimated 50% of dyslexics are thought to have MIS. The study exploring possible correlations between these minority forms of information processing is due to begin in February 2017. An invitation will be extended to all first year students enrolled in degree programmes across all faculties and schools within the institution. An estimated 900 students will be eligible to participate in the study. Participants will be asked to complete a battery of on-line questionnaires including the Highly Sensitive Person Scale, the International Dyslexia Association adult self-assessment and the adapted Irlen indicator. All three scales have been used extensively in literature and have been validated among many populations. All participants whose score on any (or some) of the three questionnaires suggest a minority method of information processing will receive an invitation to meet with a learning advisor, and given access to counselling services if they choose. Meeting with a learning advisor is not mandatory, and some participants may choose not to receive help. Data will be collected using the Question Pro platform and base-line data will be analysed using correlation and regression analysis to identify relationships and predictors between SPS, MIS and dyslexia. This study forms part of a larger three year longitudinal study and participants will be required to complete questionnaires at annual intervals in subsequent years of the study until completion of (or withdrawal from) their degree. At these data collection points, participants will be questioned on any additional support received relating to their minority method(s) of information processing. Data from this study will be available by April 2017.

Keywords: dyslexia, highly sensitive person (HSP), Meares-Irlen Syndrome (MIS), minority forms of information processing, sensory processing sensitivity (SPS)

Procedia PDF Downloads 218
12227 Explaining Irregularity in Music by Entropy and Information Content

Authors: Lorena Mihelac, Janez Povh

Abstract:

In 2017, we conducted a research study using data consisting of 160 musical excerpts from different musical styles, to analyze the impact of entropy of the harmony on the acceptability of music. In measuring the entropy of harmony, we were interested in unigrams (individual chords in the harmonic progression) and bigrams (the connection of two adjacent chords). In this study, it has been found that 53 musical excerpts out from 160 were evaluated by participants as very complex, although the entropy of the harmonic progression (unigrams and bigrams) was calculated as low. We have explained this by particularities of chord progression, which impact the listener's feeling of complexity and acceptability. We have evaluated the same data twice with new participants in 2018 and with the same participants for the third time in 2019. These three evaluations have shown that the same 53 musical excerpts, found to be difficult and complex in the study conducted in 2017, are exhibiting a high feeling of complexity again. It was proposed that the content of these musical excerpts, defined as “irregular,” is not meeting the listener's expectancy and the basic perceptual principles, creating a higher feeling of difficulty and complexity. As the “irregularities” in these 53 musical excerpts seem to be perceived by the participants without being aware of it, affecting the pleasantness and the feeling of complexity, they have been defined as “subliminal irregularities” and the 53 musical excerpts as “irregular.” In our recent study (2019) of the same data (used in previous research works), we have proposed a new measure of the complexity of harmony, “regularity,” based on the irregularities in the harmonic progression and other plausible particularities in the musical structure found in previous studies. We have in this study also proposed a list of 10 different particularities for which we were assuming that they are impacting the participant’s perception of complexity in harmony. These ten particularities have been tested in this paper, by extending the analysis in our 53 irregular musical excerpts from harmony to melody. In the examining of melody, we have used the computational model “Information Dynamics of Music” (IDyOM) and two information-theoretic measures: entropy - the uncertainty of the prediction before the next event is heard, and information content - the unexpectedness of an event in a sequence. In order to describe the features of melody in these musical examples, we have used four different viewpoints: pitch, interval, duration, scale degree. The results have shown that the texture of melody (e.g., multiple voices, homorhythmic structure) and structure of melody (e.g., huge interval leaps, syncopated rhythm, implied harmony in compound melodies) in these musical excerpts are impacting the participant’s perception of complexity. High information content values were found in compound melodies in which implied harmonies seem to have suggested additional harmonies, affecting the participant’s perception of the chord progression in harmony by creating a sense of an ambiguous musical structure.

Keywords: entropy and information content, harmony, subliminal (ir)regularity, IDyOM

Procedia PDF Downloads 116
12226 Digitalize or Die-Responsible Innovations in Healthcare and Welfare Sectors

Authors: T. Iakovleva

Abstract:

Present paper suggests a theoretical model that describes the process of the development of responsible innovations on the firm level in health and welfare sectors. There is a need to develop new firm strategies in these sectors. This paper suggests to look on the concept of responsible innovation that was originally developed on the social level and to apply this new concept to the new area of firm strategy. The rapid global diffusion of information and communication technologies has greatly improved access to knowledge. At the same time, communication is cheap, information is a commodity, and global trade increases technological diffusion. As a result, firms and users, including those outside of industrialized nations, get early exposure to the latest technologies and information. General-purpose technologies such as mobile phones and 3D printers enable individuals to solve local needs and customize products. The combined effect of these changes is having a profound impact on the innovation landscape. Meanwhile, the healthcare sector is facing unprecedented challenges, which are magnified by budgetary constraints, an aging population and the desire to provide care for all. On the other hand, patients themselves are changing. They are savvier about their diseases, they expect their relation with the healthcare professionals to be open and interactive, but above all they want to be part of the decision process. All of this is a reflection of what is already happening in other industries where customers have access to large amount of information and became educated buyers. This article addresses the question of how ICT research and innovation may contribute to developing solutions to grand societal challenges in a responsible way. A broad definition of the concept of responsibility in the context of innovation is adopted in this paper. Responsibility is thus seen as a collective, uncertain and future-oriented activity. This opens the questions of how responsibilities are perceived and distributed and how innovation and science can be governed and stewarded towards socially desirable and acceptable ends. This article addresses a central question confronting politicians, business leaders, and regional planners.

Keywords: responsible innovation, ICT, healthcare, welfare sector

Procedia PDF Downloads 183
12225 Formation of Convergence Culture in the Framework of Conventional Media and New Media

Authors: Berkay Buluş, Aytekin İşman, Kübra Yüzüncüyıl

Abstract:

Developments in media and communication technologies have changed the way we use media. The importance of convergence culture has been increasing day by day within the framework of these developments. With new media, it is possible to say that social networks are the most powerful platforms that are integrated to this digitalization process. Although social networks seem like the place that people can socialize, they can also be utilized as places of production. On the other hand, audience has become users within the framework of transformation from national to global broadcasting. User generated contents make conventional media and new media collide. In this study, these communication platforms will be examined not as platforms that replace one another but mediums that unify each other. In the light of this information, information that is produced by users regarding new media platforms and all new media use practices are called convergence culture. In other words, convergence culture means intersections of conventional and new media. In this study, examples of convergence culture will be analyzed in detail.

Keywords: new media, convergence culture, convergence, use of new media, user generated content

Procedia PDF Downloads 250
12224 Analysis of the Effect of Increased Self-Awareness on the Amount of Food Thrown Away

Authors: Agnieszka Dubiel, Artur Grabowski, Tomasz Przerywacz, Mateusz Roganowicz, Patrycja Zioty

Abstract:

Food waste is one of the most significant challenges humanity is facing nowadays. Every year, reports from global organizations show the scale of the phenomenon, although society's awareness is still insufficient. One-third of the food produced in the world is wasted at various points in the food supply chain. Wastes are present from the delivery through the food preparation and distribution to the end of the sale and consumption. The first step in understanding and resisting the phenomenon is a thorough analysis of the everyday behaviors of humanity. This concept is understood as finding the correlation between the type of food and the reason for throwing it out and wasting it. Those actions were identified as a critical step in the start of work to develop technology to prevent food waste. In this paper, the problem mentioned above was analyzed by focusing on the inhabitants of Central Europe, especially Poland, aged 20-30. This paper provides an insight into collecting data through dedicated software and an organized database. The proposed database contains information on the amount, type, and reasons for wasting food in households. A literature review supported the work to answer research questions, compare the situation in Poland with the problem analyzed in other countries, and find research gaps. The proposed article examines the cause of food waste and its quantity in detail. This review complements previous reviews by emphasizing social and economic innovation in Poland's food waste management. The paper recommends a course of action for future research on food waste management and prevention related to the handling and disposal of food, emphasizing households, i.e., the last link in the supply chain.

Keywords: food waste, food waste reduction, consumer food waste, human-food interaction

Procedia PDF Downloads 103
12223 Apply Activity-Based Costing Management System by Key Success Paths to Promote the Competitive Advantages and Operation Performance

Authors: Mei-Fang Wu, Shu-Li Wang, Feng-Tsung Cheng

Abstract:

Highly developed technology and highly competitive global market highlight the important role of competitive advantages and operation performances in sustainable company operation. Activity-Based Costing (ABC) provides accurate operation cost and operation performance information. Rich literature provide relevant research with cases study on Activity-Based Costing application, and yet, there is no research studying on cause relationship between key success factors of applying Activity-Based Costing and its specific outcomes, such as profitability or share market. These relationships provide the ways to handle the key success factors to achieve the specific outcomes for ensuring to promote the competitive advantages and operation performances. The main purposes of this research are exploring the key success paths by Key Success Paths approach which will lead the ways to apply Activity-Base Costing. The Key Success Paths is the innovative method which is exploring the cause relationships and explaining what are the effects of key success factors to specific outcomes of Activity-Based Costing implementation. The cause relationships between key success factors and successful specific outcomes are Key Success Paths (KSPs). KSPs are the guidelines to lead the cost management strategies to achieve the goals of competitive advantages and operation performances. The research findings indicate that good management system design may impact the good outcomes of Activity-Based Costing application and achieve to outstanding competitive advantage, operating performance and profitability as well by KSPs exploration.

Keywords: activity-based costing, key success factors, key success paths approach, key success paths, key failure paths

Procedia PDF Downloads 374
12222 A Comparative Analysis of ARIMA and Threshold Autoregressive Models on Exchange Rate

Authors: Diteboho Xaba, Kolentino Mpeta, Tlotliso Qejoe

Abstract:

This paper assesses the in-sample forecasting of the South African exchange rates comparing a linear ARIMA model and a SETAR model. The study uses a monthly adjusted data of South African exchange rates with 420 observations. Akaike information criterion (AIC) and the Schwarz information criteria (SIC) are used for model selection. Mean absolute error (MAE), root mean squared error (RMSE) and mean absolute percentage error (MAPE) are error metrics used to evaluate forecast capability of the models. The Diebold –Mariano (DM) test is employed in the study to check forecast accuracy in order to distinguish the forecasting performance between the two models (ARIMA and SETAR). The results indicate that both models perform well when modelling and forecasting the exchange rates, but SETAR seemed to outperform ARIMA.

Keywords: ARIMA, error metrices, model selection, SETAR

Procedia PDF Downloads 227
12221 Internet Versus Muslim Communities Challenges, Problems and Solutions

Authors: Bashir Muhammad

Abstract:

The present research contains the definition of the internet, the inter-relationship between and globalization as well as the divergent views of scholars on internet net-work. Additionally, both the positive and the negative impacts of the internet on Muslim communities were elucidated. As an example, it is part of the positive effect that the internet constitutes a vital source of vast information and data acquisition in various academic sciences in general and Islamic Studies in particular. The most recent and current facts and scientific discoveries by specialists of various ramifications could be fund as fast as possible. Many other exciting points were also cited. And on the negative side of the internet, among many other points, it releases uncontrolled promiscuous pictures and sometimes misguiding information about Islam, which could gradually and easily destroy the sound moral up bring of our young Muslim generation and pollute their positive thinking and reasoning. Another problem is that, Muslims in most cases pertaining to internet services are passive consumers, having no power to control it and manipulate it for their welfare and well being. Due to that, they have to pay the price for that, directly or indirectly.

Keywords: internet, muslim, challenges, communities

Procedia PDF Downloads 105
12220 WhatsApp Application and Challenges of Radio Broadcasting in Northern Nigeria: Special Interest on FRCN Kaduna

Authors: Aliyu Damri

Abstract:

This study analyzed the emergence of WhatsApp and how employees at the Federal Radio Corporation of Nigeria, Kaduna defined the concept base on their vast broadcasting experiences for over five decades and application of the phenomenon to the radio station. It also analyzed the nature, patterns, dimensions, features, challenges as well as the effects of WhatsApp as a social networking site with specific interest on the radio outlet. Also, the study identified how the radio organization responded to the challenges in an attempt to adapt to the new pattern of broadcasting characterized by many technological transformations. The study further explained in details such skills journalists need to function optimally using WhatsApp as well as the impacts of the WhatsApp on radio broadcasting. It used a combination of published materials, focus group discussion, in depth interviews and participant observation on the activities of the radio stations to address the research questions. The data generated provided insight to better understand the challenges posed to FRCN Kaduna as a result of WhatsApp application and how FRCN Kaduna responded to the challenges. It also provided information on the skills journalists need to function optimally in using WhatsApp application in the radio station. The interview and focus group discussion’s transcripts and the published materials were analyzed along thematic pattern related to the research questions in the study. The dominant response relied heavily on change in the radio station’s organizational and technical integration of newsrooms, the use of a multiskilled workforce, application of a flexible and user-friendly technology in all aspects of production, expansion of the station’s services in to new media such as internet and mobile phones as well as sharing of ideas across different units in the radio outfit.

Keywords: broadcasting, challenge, northern Nigeria, radio, WhatsApp application

Procedia PDF Downloads 115
12219 Modelling and Control of Binary Distillation Column

Authors: Narava Manose

Abstract:

Distillation is a very old separation technology for separating liquid mixtures that can be traced back to the chemists in Alexandria in the first century A. D. Today distillation is the most important industrial separation technology. By the eleventh century, distillation was being used in Italy to produce alcoholic beverages. At that time, distillation was probably a batch process based on the use of just a single stage, the boiler. The word distillation is derived from the Latin word destillare, which means dripping or trickling down. By at least the sixteenth century, it was known that the extent of separation could be improved by providing multiple vapor-liquid contacts (stages) in a so called Rectifactorium. The term rectification is derived from the Latin words rectefacere, meaning to improve. Modern distillation derives its ability to produce almost pure products from the use of multi-stage contacting. Throughout the twentieth century, multistage distillation was by far the most widely used industrial method for separating liquid mixtures of chemical components.The basic principle behind this technique relies on the different boiling temperatures for the various components of the mixture, allowing the separation between the vapor from the most volatile component and the liquid of other(s) component(s). •Developed a simple non-linear model of a binary distillation column using Skogestad equations in Simulink. •We have computed the steady-state operating point around which to base our analysis and controller design. However, the model contains two integrators because the condenser and reboiler levels are not controlled. One particular way of stabilizing the column is the LV-configuration where we use D to control M_D, and B to control M_B; such a model is given in cola_lv.m where we have used two P-controllers with gains equal to 10.

Keywords: modelling, distillation column, control, binary distillation

Procedia PDF Downloads 260
12218 Aligning Informatics Study Programs with Occupational and Qualifications Standards

Authors: Patrizia Poscic, Sanja Candrlic, Danijela Jaksic

Abstract:

The University of Rijeka, Department of Informatics participated in the Stand4Info project, co-financed by the European Union, with the main idea of an alignment of study programs with occupational and qualifications standards in the field of Informatics. A brief overview of our research methodology, goals and deliverables is shown. Our main research and project objectives were: a) development of occupational standards, qualification standards and study programs based on the Croatian Qualifications Framework (CROQF), b) higher education quality improvement in the field of information and communication sciences, c) increasing the employability of students of information and communication technology (ICT) and science, and d) continuously improving competencies of teachers in accordance with the principles of CROQF. CROQF is a reform instrument in the Republic of Croatia for regulating the system of qualifications at all levels through qualifications standards based on learning outcomes and following the needs of the labor market, individuals and society. The central elements of CROQF are learning outcomes - competences acquired by the individual through the learning process and proved afterward. The place of each acquired qualification is set by the level of the learning outcomes belonging to that qualification. The placement of qualifications at respective levels allows the comparison and linking of different qualifications, as well as linking of Croatian qualifications' levels to the levels of the European Qualifications Framework and the levels of the Qualifications framework of the European Higher Education Area. This research has made 3 proposals of occupational standards for undergraduate study level (System Analyst, Developer, ICT Operations Manager), and 2 for graduate (master) level (System Architect, Business Architect). For each occupational standard employers have provided a list of key tasks and associated competencies necessary to perform them. A set of competencies required for each particular job in the workplace was defined and each set of competencies as described in more details by its individual competencies. Based on sets of competencies from occupational standards, sets of learning outcomes were defined and competencies from the occupational standard were linked with learning outcomes. For each learning outcome, as well as for the set of learning outcomes, it was necessary to specify verification method, material, and human resources. The task of the project was to suggest revision and improvement of the existing study programs. It was necessary to analyze existing programs and determine how they meet and fulfill defined learning outcomes. This way, one could see: a) which learning outcomes from the qualifications standards are covered by existing courses, b) which learning outcomes have yet to be covered, c) are they covered by mandatory or elective courses, and d) are some courses unnecessary or redundant. Overall, the main research results are: a) completed proposals of qualification and occupational standards in the field of ICT, b) revised curricula of undergraduate and master study programs in ICT, c) sustainable partnership and association stakeholders network, d) knowledge network - informing the public and stakeholders (teachers, students, and employers) about the importance of CROQF establishment, and e) teachers educated in innovative methods of teaching.

Keywords: study program, qualification standard, occupational standard, higher education, informatics and computer science

Procedia PDF Downloads 130
12217 Approach for Updating a Digital Factory Model by Photogrammetry

Authors: R. Hellmuth, F. Wehner

Abstract:

Factory planning has the task of designing products, plants, processes, organization, areas, and the construction of a factory. The requirements for factory planning and the building of a factory have changed in recent years. Regular restructuring is becoming more important in order to maintain the competitiveness of a factory. Restrictions in new areas, shorter life cycles of product and production technology as well as a VUCA world (Volatility, Uncertainty, Complexity & Ambiguity) lead to more frequent restructuring measures within a factory. A digital factory model is the planning basis for rebuilding measures and becomes an indispensable tool. Short-term rescheduling can no longer be handled by on-site inspections and manual measurements. The tight time schedules require up-to-date planning models. Due to the high adaptation rate of factories described above, a methodology for rescheduling factories on the basis of a modern digital factory twin is conceived and designed for practical application in factory restructuring projects. The focus is on rebuild processes. The aim is to keep the planning basis (digital factory model) for conversions within a factory up to date. This requires the application of a methodology that reduces the deficits of existing approaches. The aim is to show how a digital factory model can be kept up to date during ongoing factory operation. A method based on photogrammetry technology is presented. The focus is on developing a simple and cost-effective solution to track the many changes that occur in a factory building during operation. The method is preceded by a hardware and software comparison to identify the most economical and fastest variant. 

Keywords: digital factory model, photogrammetry, factory planning, restructuring

Procedia PDF Downloads 100
12216 The Social Psychology of Illegal Game Room Addiction in the Historic Chinatown District of Honolulu, Hawaii: Illegal Compulsive Gambling, Chinese-Polynesian Organized Crime Syndicates, Police Corruption, and Loan Sharking Rings

Authors: Gordon James Knowles

Abstract:

Historically the Chinatown district in Sandwich Islands has been plagued with the traditional vice crimes of illegal drugs, gambling, and prostitution since the early 1800s. However, a new form of psychologically addictive arcade style table gambling machines has become the dominant form of illegal revenue made in Honolulu, Hawaii. This study attempts to document the drive, desire, or will to play and wager with arcade style video gaming and understand the role of illegal game rooms in facilitating pathological gambling addiction. Indicators of police corruption by Chinese organized crime syndicates related to protection rackets, bribery, and pay-offs were revealed. Information fusion from a police science and sociological intelligence perspective indicates insurgent warfare is being waged on the streets of Honolulu by the People’s Republic of China. This state-sponsored communist terrorism in the Hawaiian Islands used “contactless” irregular warfare entailing: (1) the deployment of psychologically addictive gambling machines, (2) the distribution of the physically addictive fentanyl drug as a lethal chemical weapon, and (3) psychological warfare by circulating pro-China anti-American propaganda newspapers targeted at the small island populace.

Keywords: Chinese and Polynesian organized crime, china daily newspaper, electronic arcade style table games, gaming technology addiction, illegal compulsive gambling, and police intelligence

Procedia PDF Downloads 56
12215 Investigating the Impact of Individual Risk-Willingness and Group-Interaction Effects on Business Model Innovation Decisions

Authors: Sarah Müller-Sägebrecht

Abstract:

Today’s volatile environment challenges executives to make the right strategic decisions to gain sustainable success. Entrepreneurship scholars postulate mainly positive effects of environmental changes on entrepreneurship behavior, such as developing new business opportunities, promoting ingenuity, and the satisfaction of resource voids. A strategic solution approach to overcome threatening environmental changes and catch new business opportunities is business model innovation (BMI). Although this research stream has gained further importance in the last decade, BMI research is still insufficient. Especially BMI barriers, such as inefficient strategic decision-making processes, need to be identified. Strategic decisions strongly impact organizational future and are, therefore, usually made in groups. Although groups draw on a more extensive information base than single individuals, group-interaction effects can influence the decision-making process - in a favorable but also unfavorable way. Decisions are characterized by uncertainty and risk, whereby their intensity is perceived individually differently. Individual risk-willingness influences which option humans choose. The special nature of strategic decisions, such as in BMI processes, is that these decisions are not made individually but in groups due to their high organizational scope. These groups consist of different personalities whose individual risk-willingness can vary considerably. It is known from group decision theory that these individuals influence each other, observable in different group-interaction effects. The following research questions arise: i) Which impact has the individual risk-willingness on BMI decisions? And ii) how do group interaction effects impact BMI decisions? After conducting 26 in-depth interviews with executives from the manufacturing industry, the applied Gioia methodology reveals the following results: i) Risk-averse decision-makers have an increased need to be guided by facts. The more information available to them, the lower they perceive uncertainty and the more willing they are to pursue a specific decision option. However, the results also show that social interaction does not change the individual risk-willingness in the decision-making process. ii) Generally, it could be observed that during BMI decisions, group interaction is primarily beneficial to increase the group’s information base for making good decisions, less than for social interaction. Further, decision-makers mainly focus on information available to all decision-makers in the team but less on personal knowledge. This work contributes to strategic decision-making literature twofold. First, it gives insights into how group-interaction effects influence an organization’s strategic BMI decision-making. Second, it enriches risk-management research by highlighting how individual risk-willingness impacts organizational strategic decision-making. To date, it was known in BMI research that risk aversion would be an internal BMI barrier. However, with this study, it becomes clear that it is not risk aversion that inhibits BMI. Instead, the lack of information prevents risk-averse decision-makers from choosing a riskier option. Simultaneously, results show that risk-averse decision-makers are not easily carried away by the higher risk-willingness of their team members. Instead, they use social interaction to gather missing information. Therefore, executives need to provide sufficient information to all decision-makers to catch promising business opportunities.

Keywords: business model innovation, decision-making, group biases, group decisions, group-interaction effects, risk-willingness

Procedia PDF Downloads 82
12214 Investigating the Determinants and Growth of Financial Technology Depth of Penetration among the Heterogeneous Africa Economies

Authors: Tochukwu Timothy Okoli, Devi Datt Tewari

Abstract:

The high rate of Fintech adoption has not transmitted to greater financial inclusion and development in Africa. This problem is attributed to poor Fintech diversification and usefulness in the continent. This concept is referred to as the Fintech depth of penetration in this study. The study, therefore, assessed its determinants and growth process in a panel of three emergings, twenty-four frontiers and five fragile African economies disaggregated with dummies over the period 2004-2018 to allow for heterogeneity between groups. The System Generalized Method of Moments (GMM) technique reveals that the average depth of Mobile banking and automated teller machine (ATM) is a dynamic heterogeneity process. Moreover, users' previous experiences/compatibility, trial-ability/income, and financial development were the major factors that raise its usefulness, whereas perceived risk, financial openness, and inflation rate significantly limit its usefulness. The growth rate of Mobile banking, ATM, and Internet banking in 2018 is, on average 41.82, 0.4, and 20.8 per cent respectively greater than its average rates in 2004. These greater averages after the 2009 financial crisis suggest that countries resort to Fintech as a risk-mitigating tool. This study, therefore, recommends greater Fintech diversification through improved literacy, institutional development, financial liberalization, and continuous innovation.

Keywords: depth of fintech, emerging Africa, financial technology, internet banking, mobile banking

Procedia PDF Downloads 114