Search results for: inclusive business models
7076 Computational Fluid Dynamicsfd Simulations of Air Pollutant Dispersion: Validation of Fire Dynamic Simulator Against the Cute Experiments of the Cost ES1006 Action
Authors: Virginie Hergault, Siham Chebbah, Bertrand Frere
Abstract:
Following in-house objectives, Central laboratory of Paris police Prefecture conducted a general review on models and Computational Fluid Dynamics (CFD) codes used to simulate pollutant dispersion in the atmosphere. Starting from that review and considering main features of Large Eddy Simulation, Central Laboratory Of Paris Police Prefecture (LCPP) postulates that the Fire Dynamics Simulator (FDS) model, from National Institute of Standards and Technology (NIST), should be well suited for air pollutant dispersion modeling. This paper focuses on the implementation and the evaluation of FDS in the frame of the European COST ES1006 Action. This action aimed at quantifying the performance of modeling approaches. In this paper, the CUTE dataset carried out in the city of Hamburg, and its mock-up has been used. We have performed a comparison of FDS results with wind tunnel measurements from CUTE trials on the one hand, and, on the other, with the models results involved in the COST Action. The most time-consuming part of creating input data for simulations is the transfer of obstacle geometry information to the format required by SDS. Thus, we have developed Python codes to convert automatically building and topographic data to the FDS input file. In order to evaluate the predictions of FDS with observations, statistical performance measures have been used. These metrics include the fractional bias (FB), the normalized mean square error (NMSE) and the fraction of predictions within a factor of two of observations (FAC2). As well as the CFD models tested in the COST Action, FDS results demonstrate a good agreement with measured concentrations. Furthermore, the metrics assessment indicate that FB and NMSE meet the tolerance acceptable.Keywords: numerical simulations, atmospheric dispersion, cost ES1006 action, CFD model, cute experiments, wind tunnel data, numerical results
Procedia PDF Downloads 1337075 Africa as Endemically a War Continent: Explaining the Changing Pattern of Armed Conflicts in Africa
Authors: Kenneth Azaigba
Abstract:
The history of post-colonial African States has been dubbed a history of endemic warfare in existing literature. Indeed, Africa political environment is characterized by a multiplicity of threats to peace and security. Africa's leading drivers of conflict include abundant (especially mineral) resources, personal rule and attendant political authoritarianism, manipulation of identity politics across ethnicity, marginalization of communities, as well as electoral mal-practices resulting in contested legitimacy and resultant violence. However, the character of armed conflicts in Africa is changing. This paper attempts to reconstruct the trajectory of armed conflicts in Africa and explain the changing pattern of armed conflict. The paper contends that large scale political violence in Africa is on the decline rendering the endemic thesis an inappropriate paradigm in explaining political conflicts in Africa. The paper also posits that though small scale conflicts are springing up and exhibiting trans-border dimensions, these patterns of armed conflicts are not peculiar to Africa but emerging waves of global conflicts. The paper explains that the shift in the scale of warfare in Africa is a function of a multiplicity of post-cold war global contradictions. Inclusive governance, social justice and economic security are articulated as workable panaceas for mitigating warfare in Africa.Keywords: Africa, conflicts, pattern, war
Procedia PDF Downloads 3877074 Empowerment Model: A Strategy for Supporting Creative Economy through Traditional Weaving in Anajiaka Village
Authors: Sita Yuliastuti Amijaya, Wiyatiningsih Wiyatiningsih, Paulus Bawole
Abstract:
Weaving skills were not originally a way to earn money for the traditional people on Sumba Island. Weaving is a leisure activity carried out between farming and caring for families. It is quite understandable if the weavers are women. At this time, weaving crafts become a unique potential inherent in an area, so that the weaver women also have the potential to drive economic activity in regional tourism sector. This study aims to measure the sustainability of traditional weaving business activities in Anajiaka Village, Umbu Ratu Nggay Barat, Central Sumba Regency, which is able to support the creative economy. The analysis was performed using qualitative descriptive methods by comparing the criteria of smart living and smart economy in the study of smart city. This study found that business sustainability will be better maintained if it is bound in a joint commitment, for example by forming a group of craftsmen. Other challenges besides the commitment of the group members are aspects of local government support and related agencies, in the form of guidance, funding, and promotion. In addition, fabric order targets, maintaining family and community balance, are recognized as obstacles for craftsmen. The modern marketing model is not yet mastered by the craftsmen group, so it needs assistance for future development.Keywords: agriculture, craftsmen, creativepreneur, smart economy, smart living
Procedia PDF Downloads 1677073 Masquerade and “What Comes Behind Six Is More Than Seven”: Thoughts on Art History and Visual Culture Research Methods
Authors: Osa D Egonwa
Abstract:
In the 21st century, the disciplinary boundaries of past centuries that we often create through mainstream art historical classification, techniques and sources may have been eroded by visual culture, which seems to provide a more inclusive umbrella for the new ways artists go about the creative process and its resultant commodities. Over the past four decades, artists in Africa have resorted to new materials, techniques and themes which have affected our ways of research on these artists and their art. Frontline artists such as El Anatsui, Yinka Shonibare, Erasmus Onyishi are demonstrating that any material is just suitable for artistic expression. Most of times, these materials come with their own techniques/effects and visual syntax: a combination of materials compounds techniques, formal aesthetic indexes, halo effects, and iconography. This tends to challenge the categories and we lean on to view, think and talk about them. This renders our main stream art historical research methods inadequate, thus suggesting new discursive concepts, terms and theories. This paper proposed the Africanist eclectic methods derived from the dual framework of Masquerade Theory and What Comes Behind Six is More Than Seven. This paper shares thoughts/research on art historical methods, terminological re-alignments on classification/source data, presentational format and interpretation arising from the emergent trends in our subject. The outcome provides useful tools to mediate new thoughts and experiences in recent African art and visual culture.Keywords: art historical methods, classifications, concepts, re-alignment
Procedia PDF Downloads 1107072 Generalized Correlation Coefficient in Genome-Wide Association Analysis of Cognitive Ability in Twins
Authors: Afsaneh Mohammadnejad, Marianne Nygaard, Jan Baumbach, Shuxia Li, Weilong Li, Jesper Lund, Jacob v. B. Hjelmborg, Lene Christensen, Qihua Tan
Abstract:
Cognitive impairment in the elderly is a key issue affecting the quality of life. Despite a strong genetic background in cognition, only a limited number of single nucleotide polymorphisms (SNPs) have been found. These explain a small proportion of the genetic component of cognitive function, thus leaving a large proportion unaccounted for. We hypothesize that one reason for this missing heritability is the misspecified modeling in data analysis concerning phenotype distribution as well as the relationship between SNP dosage and the phenotype of interest. In an attempt to overcome these issues, we introduced a model-free method based on the generalized correlation coefficient (GCC) in a genome-wide association study (GWAS) of cognitive function in twin samples and compared its performance with two popular linear regression models. The GCC-based GWAS identified two genome-wide significant (P-value < 5e-8) SNPs; rs2904650 near ZDHHC2 on chromosome 8 and rs111256489 near CD6 on chromosome 11. The kinship model also detected two genome-wide significant SNPs, rs112169253 on chromosome 4 and rs17417920 on chromosome 7, whereas no genome-wide significant SNPs were found by the linear mixed model (LME). Compared to the linear models, more meaningful biological pathways like GABA receptor activation, ion channel transport, neuroactive ligand-receptor interaction, and the renin-angiotensin system were found to be enriched by SNPs from GCC. The GCC model outperformed the linear regression models by identifying more genome-wide significant genetic variants and more meaningful biological pathways related to cognitive function. Moreover, GCC-based GWAS was robust in handling genetically related twin samples, which is an important feature in handling genetic confounding in association studies.Keywords: cognition, generalized correlation coefficient, GWAS, twins
Procedia PDF Downloads 1247071 Training AI to Be Empathetic and Determining the Psychotype of a Person During a Conversation with a Chatbot
Authors: Aliya Grig, Konstantin Sokolov, Igor Shatalin
Abstract:
The report describes the methodology for collecting data and building an ML model for determining the personality psychotype using profiling and personality traits methods based on several short messages of a user communicating on an arbitrary topic with a chitchat bot. In the course of the experiments, the minimum amount of text was revealed to confidently determine aspects of personality. Model accuracy - 85%. Users' language of communication is English. AI for a personalized communication with a user based on his mood, personality, and current emotional state. Features investigated during the research: personalized communication; providing empathy; adaptation to a user; predictive analytics. In the report, we describe the processes that captures both structured and unstructured data pertaining to a user in large quantities and diverse forms. This data is then effectively processed through ML tools to construct a knowledge graph and draw inferences regarding users of text messages in a comprehensive manner. Specifically, the system analyzes users' behavioral patterns and predicts future scenarios based on this analysis. As a result of the experiments, we provide for further research on training AI models to be empathetic, creating personalized communication for a userKeywords: AI, empathetic, chatbot, AI models
Procedia PDF Downloads 927070 Machine Learning in Patent Law: How Genetic Breeding Algorithms Challenge Modern Patent Law Regimes
Authors: Stefan Papastefanou
Abstract:
Artificial intelligence (AI) is an interdisciplinary field of computer science with the aim of creating intelligent machine behavior. Early approaches to AI have been configured to operate in very constrained environments where the behavior of the AI system was previously determined by formal rules. Knowledge was presented as a set of rules that allowed the AI system to determine the results for specific problems; as a structure of if-else rules that could be traversed to find a solution to a particular problem or question. However, such rule-based systems typically have not been able to generalize beyond the knowledge provided. All over the world and especially in IT-heavy industries such as the United States, the European Union, Singapore, and China, machine learning has developed to be an immense asset, and its applications are becoming more and more significant. It has to be examined how such products of machine learning models can and should be protected by IP law and for the purpose of this paper patent law specifically, since it is the IP law regime closest to technical inventions and computing methods in technical applications. Genetic breeding models are currently less popular than recursive neural network method and deep learning, but this approach can be more easily described by referring to the evolution of natural organisms, and with increasing computational power; the genetic breeding method as a subset of the evolutionary algorithms models is expected to be regaining popularity. The research method focuses on patentability (according to the world’s most significant patent law regimes such as China, Singapore, the European Union, and the United States) of AI inventions and machine learning. Questions of the technical nature of the problem to be solved, the inventive step as such, and the question of the state of the art and the associated obviousness of the solution arise in the current patenting processes. Most importantly, and the key focus of this paper is the problem of patenting inventions that themselves are developed through machine learning. The inventor of a patent application must be a natural person or a group of persons according to the current legal situation in most patent law regimes. In order to be considered an 'inventor', a person must actually have developed part of the inventive concept. The mere application of machine learning or an AI algorithm to a particular problem should not be construed as the algorithm that contributes to a part of the inventive concept. However, when machine learning or the AI algorithm has contributed to a part of the inventive concept, there is currently a lack of clarity regarding the ownership of artificially created inventions. Since not only all European patent law regimes but also the Chinese and Singaporean patent law approaches include identical terms, this paper ultimately offers a comparative analysis of the most relevant patent law regimes.Keywords: algorithms, inventor, genetic breeding models, machine learning, patentability
Procedia PDF Downloads 1087069 Improved Soil and Snow Treatment with the Rapid Update Cycle Land-Surface Model for Regional and Global Weather Predictions
Authors: Tatiana G. Smirnova, Stan G. Benjamin
Abstract:
Rapid Update Cycle (RUC) land surface model (LSM) was a land-surface component in several generations of operational weather prediction models at the National Center for Environment Prediction (NCEP) at the National Oceanic and Atmospheric Administration (NOAA). It was designed for short-range weather predictions with an emphasis on severe weather and originally was intentionally simple to avoid uncertainties from poorly known parameters. Nevertheless, the RUC LSM, when coupled with the hourly-assimilating atmospheric model, can produce a realistic evolution of time-varying soil moisture and temperature, as well as the evolution of snow cover on the ground surface. This result is possible only if the soil/vegetation/snow component of the coupled weather prediction model has sufficient skill to avoid long-term drift. RUC LSM was first implemented in the operational NCEP Rapid Update Cycle (RUC) weather model in 1998 and later in the Weather Research Forecasting Model (WRF)-based Rapid Refresh (RAP) and High-resolution Rapid Refresh (HRRR). Being available to the international WRF community, it was implemented in operational weather models in Austria, New Zealand, and Switzerland. Based on the feedback from the US weather service offices and the international WRF community and also based on our own validation, RUC LSM has matured over the years. Also, a sea-ice module was added to RUC LSM for surface predictions over the Arctic sea-ice. Other modifications include refinements to the snow model and a more accurate specification of albedo, roughness length, and other surface properties. At present, RUC LSM is being tested in the regional application of the Unified Forecast System (UFS). The next generation UFS-based regional Rapid Refresh FV3 Standalone (RRFS) model will replace operational RAP and HRRR at NCEP. Over time, RUC LSM participated in several international model intercomparison projects to verify its skill using observed atmospheric forcing. The ESM-SnowMIP was the last of these experiments focused on the verification of snow models for open and forested regions. The simulations were performed for ten sites located in different climatic zones of the world forced with observed atmospheric conditions. While most of the 26 participating models have more sophisticated snow parameterizations than in RUC, RUC LSM got a high ranking in simulations of both snow water equivalent and surface temperature. However, ESM-SnowMIP experiment also revealed some issues in the RUC snow model, which will be addressed in this paper. One of them is the treatment of grid cells partially covered with snow. RUC snow module computes energy and moisture budgets of snow-covered and snow-free areas separately by aggregating the solutions at the end of each time step. Such treatment elevates the importance of computing in the model snow cover fraction. Improvements to the original simplistic threshold-based approach have been implemented and tested both offline and in the coupled weather model. The detailed description of changes to the snow cover fraction and other modifications to RUC soil and snow parameterizations will be described in this paper.Keywords: land-surface models, weather prediction, hydrology, boundary-layer processes
Procedia PDF Downloads 887068 Nature of Polaronic Hopping Conduction Mechanism in Polycrystalline and Nanocrystalline Gd0.5Sr0.5MnO3 Compounds
Authors: Soma Chatterjee, I. Das
Abstract:
In the present study, we have investigated the structural, electrical and magneto-transport properties of polycrystalline and nanocrystalline Gd0.5Sr0.5MnO3 compounds. The variation of transport properties is modified by tuning the grain size of the material. In the high-temperature semiconducting region, temperature-dependent resistivity data can be well explained by the non-adiabatic small polaron hopping (SPH) mechanism. In addition, the resistivity data for all compounds in the low-temperature paramagnetic region can also be well explained by the variable range hopping (VRH) model. The parameters obtained from SPH and VRH mechanisms are found to be reasonable. In the case of nanocrystalline compounds, there is an overlapping temperature range where both SPH and VRH models are valid simultaneously, and a new conduction mechanism - variable range hopping of small polaron s(VR-SPH) is satisfactorily valid for the whole temperature range of these compounds. However, for the polycrystalline compound, the overlapping temperature region between VRH and SPH models does not exist and the VR-SPH mechanism is not valid here. Thus, polarons play a leading role in selecting different conduction mechanisms in different temperature ranges.Keywords: electrical resistivity, manganite, small polaron hopping, variable range hopping, variable range of small polaron hopping
Procedia PDF Downloads 897067 XAI Implemented Prognostic Framework: Condition Monitoring and Alert System Based on RUL and Sensory Data
Authors: Faruk Ozdemir, Roy Kalawsky, Peter Hubbard
Abstract:
Accurate estimation of RUL provides a basis for effective predictive maintenance, reducing unexpected downtime for industrial equipment. However, while models such as the Random Forest have effective predictive capabilities, they are the so-called ‘black box’ models, where interpretability is at a threshold to make critical diagnostic decisions involved in industries related to aviation. The purpose of this work is to present a prognostic framework that embeds Explainable Artificial Intelligence (XAI) techniques in order to provide essential transparency in Machine Learning methods' decision-making mechanisms based on sensor data, with the objective of procuring actionable insights for the aviation industry. Sensor readings have been gathered from critical equipment such as turbofan jet engine and landing gear, and the prediction of the RUL is done by a Random Forest model. It involves steps such as data gathering, feature engineering, model training, and evaluation. These critical components’ datasets are independently trained and evaluated by the models. While suitable predictions are served, their performance metrics are reasonably good; such complex models, however obscure reasoning for the predictions made by them and may even undermine the confidence of the decision-maker or the maintenance teams. This is followed by global explanations using SHAP and local explanations using LIME in the second phase to bridge the gap in reliability within industrial contexts. These tools analyze model decisions, highlighting feature importance and explaining how each input variable affects the output. This dual approach offers a general comprehension of the overall model behavior and detailed insight into specific predictions. The proposed framework, in its third component, incorporates the techniques of causal analysis in the form of Granger causality tests in order to move beyond correlation toward causation. This will not only allow the model to predict failures but also present reasons, from the key sensor features linked to possible failure mechanisms to relevant personnel. The causality between sensor behaviors and equipment failures creates much value for maintenance teams due to better root cause identification and effective preventive measures. This step contributes to the system being more explainable. Surrogate Several simple models, including Decision Trees and Linear Models, can be used in yet another stage to approximately represent the complex Random Forest model. These simpler models act as backups, replicating important jobs of the original model's behavior. If the feature explanations obtained from the surrogate model are cross-validated with the primary model, the insights derived would be more reliable and provide an intuitive sense of how the input variables affect the predictions. We then create an iterative explainable feedback loop, where the knowledge learned from the explainability methods feeds back into the training of the models. This feeds into a cycle of continuous improvement both in model accuracy and interpretability over time. By systematically integrating new findings, the model is expected to adapt to changed conditions and further develop its prognosis capability. These components are then presented to the decision-makers through the development of a fully transparent condition monitoring and alert system. The system provides a holistic tool for maintenance operations by leveraging RUL predictions, feature importance scores, persistent sensor threshold values, and autonomous alert mechanisms. Since the system will provide explanations for the predictions given, along with active alerts, the maintenance personnel can make informed decisions on their end regarding correct interventions to extend the life of the critical machinery.Keywords: predictive maintenance, explainable artificial intelligence, prognostic, RUL, machine learning, turbofan engines, C-MAPSS dataset
Procedia PDF Downloads 67066 Predictive Value of ¹⁸F-Fdg Accumulation in Visceral Fat Activity to Detect Colorectal Cancer Metastases
Authors: Amil Suleimanov, Aigul Saduakassova, Denis Vinnikov
Abstract:
Objective: To assess functional visceral fat (VAT) activity evaluated by ¹⁸F-fluorodeoxyglucose (¹⁸F-FDG) positron emission tomography/computed tomography (PET/CT) as a predictor of metastases in colorectal cancer (CRC). Materials and methods: We assessed 60 patients with histologically confirmed CRC who underwent 18F-FDG PET/CT after a surgical treatment and courses of chemotherapy. Age, histology, stage, and tumor grade were recorded. Functional VAT activity was measured by maximum standardized uptake value (SUVmax) using ¹⁸F-FDG PET/CT and tested as a predictor of later metastases in eight abdominal locations (RE – Epigastric Region, RLH – Left Hypochondriac Region, RRL – Right Lumbar Region, RU – Umbilical Region, RLL – Left Lumbar Region, RRI – Right Inguinal Region, RP – Hypogastric (Pubic) Region, RLI – Left Inguinal Region) and pelvic cavity (P) in the adjusted regression models. We also report the best areas under the curve (AUC) for SUVmax with the corresponding sensitivity (Se) and specificity (Sp). Results: In both adjusted for age regression models and ROC analysis, 18F-FDG accumulation in RLH (cutoff SUVmax 0.74; Se 75%; Sp 61%; AUC 0.668; p = 0.049), RU (cutoff SUVmax 0.78; Se 69%; Sp 61%; AUC 0.679; p = 0.035), RRL (cutoff SUVmax 1.05; Se 69%; Sp 77%; AUC 0.682; p = 0.032) and RRI (cutoff SUVmax 0.85; Se 63%; Sp 61%; AUC 0.672; p = 0.043) could predict later metastases in CRC patients, as opposed to age, sex, primary tumor location, tumor grade and histology. Conclusions: VAT SUVmax is significantly associated with later metastases in CRC patients and can be used as their predictor.Keywords: ¹⁸F-FDG, PET/CT, colorectal cancer, predictive value
Procedia PDF Downloads 1177065 Combination of Artificial Neural Network Model and Geographic Information System for Prediction Water Quality
Authors: Sirilak Areerachakul
Abstract:
Water quality has initiated serious management efforts in many countries. Artificial Neural Network (ANN) models are developed as forecasting tools in predicting water quality trend based on historical data. This study endeavors to automatically classify water quality. The water quality classes are evaluated using 6 factor indices. These factors are pH value (pH), Dissolved Oxygen (DO), Biochemical Oxygen Demand (BOD), Nitrate Nitrogen (NO3N), Ammonia Nitrogen (NH3N) and Total Coliform (T-Coliform). The methodology involves applying data mining techniques using multilayer perceptron (MLP) neural network models. The data consisted of 11 sites of Saen Saep canal in Bangkok, Thailand. The data is obtained from the Department of Drainage and Sewerage Bangkok Metropolitan Administration during 2007-2011. The results of multilayer perceptron neural network exhibit a high accuracy multilayer perception rate at 94.23% in classifying the water quality of Saen Saep canal in Bangkok. Subsequently, this encouraging result could be combined with GIS data improves the classification accuracy significantly.Keywords: artificial neural network, geographic information system, water quality, computer science
Procedia PDF Downloads 3437064 Exploring the Applications of Neural Networks in the Adaptive Learning Environment
Authors: Baladitya Swaika, Rahul Khatry
Abstract:
Computer Adaptive Tests (CATs) is one of the most efficient ways for testing the cognitive abilities of students. CATs are based on Item Response Theory (IRT) which is based on item selection and ability estimation using statistical methods of maximum information selection/selection from posterior and maximum-likelihood (ML)/maximum a posteriori (MAP) estimators respectively. This study aims at combining both classical and Bayesian approaches to IRT to create a dataset which is then fed to a neural network which automates the process of ability estimation and then comparing it to traditional CAT models designed using IRT. This study uses python as the base coding language, pymc for statistical modelling of the IRT and scikit-learn for neural network implementations. On creation of the model and on comparison, it is found that the Neural Network based model performs 7-10% worse than the IRT model for score estimations. Although performing poorly, compared to the IRT model, the neural network model can be beneficially used in back-ends for reducing time complexity as the IRT model would have to re-calculate the ability every-time it gets a request whereas the prediction from a neural network could be done in a single step for an existing trained Regressor. This study also proposes a new kind of framework whereby the neural network model could be used to incorporate feature sets, other than the normal IRT feature set and use a neural network’s capacity of learning unknown functions to give rise to better CAT models. Categorical features like test type, etc. could be learnt and incorporated in IRT functions with the help of techniques like logistic regression and can be used to learn functions and expressed as models which may not be trivial to be expressed via equations. This kind of a framework, when implemented would be highly advantageous in psychometrics and cognitive assessments. This study gives a brief overview as to how neural networks can be used in adaptive testing, not only by reducing time-complexity but also by being able to incorporate newer and better datasets which would eventually lead to higher quality testing.Keywords: computer adaptive tests, item response theory, machine learning, neural networks
Procedia PDF Downloads 1757063 Effects of Nano-Coating on the Mechanical Behavior of Nanoporous Metals
Authors: Yunus Onur Yildiz, Mesut Kirca
Abstract:
In this study, mechanical properties of a nanoporous metal coated with a different metallic material are studied through a new atomistic modelling technique and molecular dynamics (MD) simulations. This new atomistic modelling technique is based on the Voronoi tessellation method for the purpose of geometric representation of the ligaments. With the proposed technique, atomistic models of nanoporous metals which have randomly oriented ligaments with non-uniform mass distribution along the ligament axis can be generated by enabling researchers to control both ligament length and diameter. Furthermore, by the utilization of this technique, atomistic models of coated nanoporous materials can be numerically obtained for further mechanical or thermal characterization. In general, this study consists of two stages. At the first stage, we use algorithms developed for generating atomic coordinates of the coated nanoporous material. In this regard, coordinates of randomly distributed points are determined in a controlled way to be employed in the establishment of the Voronoi tessellation, which results in randomly oriented and intersected line segments. Then, line segment representation of the Voronoi tessellation is transformed to atomic structure by a special process. This special process includes generation of non-uniform volumetric core region in which atoms can be generated based on a specific crystal structure. As an extension, this technique can be used for coating of nanoporous structures by creating another volumetric region encapsulating the core region in which atoms for the coating material are generated. The ultimate goal of the study at this stage is to generate atomic coordinates that can be employed in the MD simulations of randomly organized coated nanoporous structures. At the second stage of the study, mechanical behavior of the coated nanoporous models is investigated by examining deformation mechanisms through MD simulations. In this way, the effect of coating on the mechanical behavior of the selected material couple is investigated.Keywords: atomistic modelling, molecular dynamic, nanoporous metals, voronoi tessellation
Procedia PDF Downloads 2777062 Technology in Commercial Law Enforcement: Tanzania, Canada, and Singapore Comparatively
Authors: Katarina Revocati Mteule
Abstract:
The background of this research arises from global demands for fair business opportunities. As one of responses to these demands, nations embarked on reforms in commercial laws. In 1990s Tanzania resorted to economic transformation through liberalization to attract more investments included reform in commercial laws enforcement. This research scrutinizes the effectiveness of reforms in Tanzania in comparison with Canada and Singapore and the role of technology. The methodology to be used is doctrinal legal research mixed with international comparative legal research. It involves comparative analysis of library, online, and internet resources as well as Case Laws and Statutory Laws. Tanzania, Canada and Singapore are sampled comparators basing on their distinct level of economic development. The criteria of analysis includes the nature of reforms, type of technology, technological infrastructure and human resource technical competence in each country. As the world progresses towards reforms in commercial laws, improvements in law, policy, and regulatory frameworks are paramount. Specifically, commercial laws are essential in contract enforcement and dispute resolution and how it copes with modern technologies is a concern. Harnessing the best technology is necessary to cope with the modernity in world businesses. In line with this, Tanzania is improving its business environment, including law enforcement mechanisms that are supportive to investments. Reforms such as specialized commercial law enforcement coupled with alternative dispute resolutions such as arbitration, mediation, and reconciliation are emphasized. Court technology as one of the reform tools given high priority. This research evaluates the progress and the effectiveness of the reforms in Commercial Laws towards friendly business environment in Tanzania in comparison with Canada and Singapore. The experience of Tanzania is compared with Canada and Singapore to see what to improve for each country to enhance quick and fair enforcement of commercial law. The research proposes necessary global standards of procedures and in national laws to offer a business-friendly environment and the use of appropriate technology. Solutions are proposed in tackling the challenges of delays in enforcing Commercial Laws such as case management, funding, legal and procedural hindrances, laxity among staff, and abuse of Court process among litigants, all in line with modern technology. It is the finding of the research that proper use of technology has managed to reduce case backlogs and time taken to resolve a commercial dispute, to increase court integrity by minimizing human contacts in commercial law enforcement which may lead to solicitation of favors and saving of parties’ time due to online service. Among the three countries, each one is facing a distinct challenge due to the level of poverty and remoteness from online service. How solutions are found in one country is a lesson to another. To conclude, this paper is suggesting solutions for improving the commercial law enforcement mechanisms in line with modern technology. The call for technological transformation is essential for the enforcement of commercial laws.Keywords: commercial law, enforcement, technology
Procedia PDF Downloads 587061 Social Entrepreneurship and Inclusive Growth
Authors: Sudheer Gupta
Abstract:
Approximately 4 billion citizens of the world live on the equivalent of less than $8 a day. This segment constitutes a $5 trillion global market that remains under-served. Multinational corporations have historically tended to focus their innovation efforts on the upper segments of the economic pyramid. The academic literature has also been dominated by theories and frameworks of innovation that are valid when applied to the developed markets and consumer segments, but fail to adequately account for the challenges and realities of new product and service creation for the poor. Theories of entrepreneurship developed in the context of developed markets similarly ignore the challenges and realities of operating in developing economies that can be characterized by missing institutions, missing markets, information and infrastructural challenges, and resource constraints. Social entrepreneurs working in such contexts develop solutions differently. In this talk, we summarize lessons learnt from a long-term research project that involves data collection from a broad range of social entrepreneurs in developing countries working towards solutions to alleviate poverty, and grounded theory-building efforts. We aim to develop a better understanding of consumers, producers, and other stakeholder involvement, thus laying the foundation to build a robust theory of innovation and entrepreneurship for the poor.Keywords: poverty alleviation, social enterprise, social innovation, development
Procedia PDF Downloads 3997060 Emulation of a Wind Turbine Using Induction Motor Driven by Field Oriented Control
Authors: L. Benaaouinate, M. Khafallah, A. Martinez, A. Mesbahi, T. Bouragba
Abstract:
This paper concerns with the modeling, simulation, and emulation of a wind turbine emulator for standalone wind energy conversion systems. By using emulation system, we aim to reproduce the dynamic behavior of the wind turbine torque on the generator shaft: it provides the testing facilities to optimize generator control strategies in a controlled environment, without reliance on natural resources. The aerodynamic, mechanical, electrical models have been detailed as well as the control of pitch angle using Fuzzy Logic for horizontal axis wind turbines. The wind turbine emulator consists mainly of an induction motor with AC power drive with torque control. The control of the induction motor and the mathematical models of the wind turbine are designed with MATLAB/Simulink environment. The simulation results confirm the effectiveness of the induction motor control system and the functionality of the wind turbine emulator for providing all necessary parameters of the wind turbine system such as wind speed, output torque, power coefficient and tip speed ratio. The findings are of direct practical relevance.Keywords: electrical generator, induction motor drive, modeling, pitch angle control, real time control, renewable energy, wind turbine, wind turbine emulator
Procedia PDF Downloads 2347059 Green Hydrogen: Exploring Economic Viability and Alluring Business Scenarios
Authors: S. Sakthivel
Abstract:
Currently, the global economy is based on the hydrocarbon economy, which is referencing the global hydrocarbon industry. Problems of using these fossil fuels (like oil, NG, coal) are emitting greenhouse gases (GHGs) and price fluctuation, supply/distribution, etc. These challenges can be overcome by using clean energy as hydrogen. The hydrogen economy is the use of hydrogen as a low carbon fuel, particularly for hydrogen vehicles, alternative industrial feedstock, power generation, and energy storage, etc. Engineering consulting firms have a significant role in this ambition and green hydrogen value chain (i.e., integration of renewables, production, storage, and distribution to end-users). Typically, the cost of green hydrogen is a function of the price of electricity needed, the cost of the electrolyser, and the operating cost to run the system. This article focuses on economic viability and explores the alluring business scenarios globally. Break-even analysis was carried out for green hydrogen production and in order to evaluate and compare the impact of the electricity price on the production costs of green hydrogen and relate it to fossil fuel-based brown/grey/blue hydrogen costs. It indicates that the cost of green hydrogen production will fall drastically due to the declining costs of renewable electricity prices and along with the improvement and scaling up of electrolyser manufacturing. For instance, in a scenario where electricity prices are below US$ 40/MWh, green hydrogen cost is expected to reach cost competitiveness.Keywords: green hydrogen, cost analysis, break-even analysis, renewables, electrolyzer
Procedia PDF Downloads 1437058 Duality of Leagility and Governance: A New Normal Demand Network Management Paradigm under Pandemic
Authors: Jacky Hau
Abstract:
The prevalence of emerging technologies disrupts various industries as well as consumer behavior. Data collection has been in the fingertip and inherited through enabled Internet-of-things (IOT) devices. Big data analytics (BDA) becomes possible and allows real-time demand network management (DNM) through leagile supply chain. To enhance further on its resilience and predictability, governance is going to be examined to promote supply chain transparency and trust in an efficient manner. Leagility combines lean thinking and agile techniques in supply chain management. It aims at reducing costs and waste, as well as maintaining responsiveness to any volatile consumer demand by means of adjusting the decoupling point where the product flow changes from push to pull. Leagility would only be successful when collaborative planning, forecasting, and replenishment (CPFR) process or alike is in place throughout the supply chain business entities. Governance and procurement of the supply chain, however, is crucial and challenging for the execution of CPFR as every entity has to walk-the-talk generously for the sake of overall benefits of supply chain performance, not to mention the complexity of exercising the polices at both of within across various supply chain business entities on account of organizational behavior and mutual trust. Empirical survey results showed that the effective timespan on demand forecasting had been drastically shortening in the magnitude of months to weeks planning horizon, thus agility shall come first and preferably following by lean approach in a timely manner.Keywords: governance, leagility, procure-to-pay, source-to-contract
Procedia PDF Downloads 1117057 Upward Millennium: Enterprise Resource Planning (ERP) Development and Implementation in Pakistani Organizations
Authors: Sara Aziz, Madiha Arooj, Hira Rizwani, Wasim Irshad
Abstract:
Enterprise Resource Planning (ER) as component of Information Resource System has turned up as one of the most demanding software in market for the new millennium. ERP system automates the core activities of any organization such as finance, manufacturing and supply chain management, human resource etc. to generate an access to the information in real time environment. Despite this fact many of the organizations globally particularly in developing country Pakistan are unaware and avoid adopting it. The development and implementation of ERP system is a complex and challenging process. This research was aimed to explore the benefits and coping strategies (with reference to end user reaction) of organizations those have implemented ERP. The problems addressed in this study focused the challenges and key success factors regarding implementing ERP Pakistani Organizations. Secondly, it has explored the stumbling blocks and business integration of those organizations that are not implementing ERP. The public and corporate sector organizations in Pakistan were selected to collect the data. The research finding shows that the organizational culture, openness towards adoption and learning, deployment and development, top management commitment and change systems, business processes and compatibility and user acceptance and reaction are contributing factors for successful implementation and development of ERP system. This research is thus an addition to enhance knowledge and understanding of implementation of ERP system in Pakistan.Keywords: ERP system, user acceptance and involvement, change management, organizational culture
Procedia PDF Downloads 2817056 Mathematical Modeling of Thin Layer Drying Behavior of Bhimkol (Musa balbisiana) Pulp
Authors: Ritesh Watharkar, Sourabh Chakraborty, Brijesh Srivastava
Abstract:
Reduction of water from the fruits and vegetables using different drying techniques is widely employed to prolong the shelf life of these food commodities. Heat transfer occurs inside the sample by conduction and mass transfer takes place by diffusion in accordance with temperature and moisture concentration gradient respectively during drying. This study was undertaken to study and model the thin layer drying behavior of Bhimkol pulp. The drying was conducted in a tray drier at 500c temperature with 5, 10 and 15 % concentrations of added maltodextrin. The drying experiments were performed at 5mm thickness of the thin layer and the constant air velocity of 0.5 m/s.Drying data were fitted to different thin layer drying models found in the literature. Comparison of fitted models was based on highest R2(0.9917), lowest RMSE (0.03201), and lowest SSE (0.01537) revealed Middle equation as the best-fitted model for thin layer drying with 10% concentration of maltodextrin. The effective diffusivity was estimated based on the solution of Fick’s law of diffusion which is found in the range of 3.0396 x10-09 to 5.0661 x 10-09. There was a reduction in drying time with the addition of maltodextrin as compare to the raw pulp.Keywords: Bhimkol, diffusivity, maltodextrine, Midilli model
Procedia PDF Downloads 2117055 Support Vector Regression Combined with Different Optimization Algorithms to Predict Global Solar Radiation on Horizontal Surfaces in Algeria
Authors: Laidi Maamar, Achwak Madani, Abdellah El Ahdj Abdellah
Abstract:
The aim of this work is to use Support Vector regression (SVR) combined with dragonfly, firefly, Bee Colony and particle swarm Optimization algorithm to predict global solar radiation on horizontal surfaces in some cities in Algeria. Combining these optimization algorithms with SVR aims principally to enhance accuracy by fine-tuning the parameters, speeding up the convergence of the SVR model, and exploring a larger search space efficiently; these parameters are the regularization parameter (C), kernel parameters, and epsilon parameter. By doing so, the aim is to improve the generalization and predictive accuracy of the SVR model. Overall, the aim is to leverage the strengths of both SVR and optimization algorithms to create a more powerful and effective regression model for various cities and under different climate conditions. Results demonstrate close agreement between predicted and measured data in terms of different metrics. In summary, SVM has proven to be a valuable tool in modeling global solar radiation, offering accurate predictions and demonstrating versatility when combined with other algorithms or used in hybrid forecasting models.Keywords: support vector regression (SVR), optimization algorithms, global solar radiation prediction, hybrid forecasting models
Procedia PDF Downloads 357054 A Conv-Long Short-term Memory Deep Learning Model for Traffic Flow Prediction
Authors: Ali Reza Sattarzadeh, Ronny J. Kutadinata, Pubudu N. Pathirana, Van Thanh Huynh
Abstract:
Traffic congestion has become a severe worldwide problem, affecting everyday life, fuel consumption, time, and air pollution. The primary causes of these issues are inadequate transportation infrastructure, poor traffic signal management, and rising population. Traffic flow forecasting is one of the essential and effective methods in urban congestion and traffic management, which has attracted the attention of researchers. With the development of technology, undeniable progress has been achieved in existing methods. However, there is a possibility of improvement in the extraction of temporal and spatial features to determine the importance of traffic flow sequences and extraction features. In the proposed model, we implement the convolutional neural network (CNN) and long short-term memory (LSTM) deep learning models for mining nonlinear correlations and their effectiveness in increasing the accuracy of traffic flow prediction in the real dataset. According to the experiments, the results indicate that implementing Conv-LSTM networks increases the productivity and accuracy of deep learning models for traffic flow prediction.Keywords: deep learning algorithms, intelligent transportation systems, spatiotemporal features, traffic flow prediction
Procedia PDF Downloads 1717053 Comprehensive Study of Data Science
Authors: Asifa Amara, Prachi Singh, Kanishka, Debargho Pathak, Akshat Kumar, Jayakumar Eravelly
Abstract:
Today's generation is totally dependent on technology that uses data as its fuel. The present study is all about innovations and developments in data science and gives an idea about how efficiently to use the data provided. This study will help to understand the core concepts of data science. The concept of artificial intelligence was introduced by Alan Turing in which the main principle was to create an artificial system that can run independently of human-given programs and can function with the help of analyzing data to understand the requirements of the users. Data science comprises business understanding, analyzing data, ethical concerns, understanding programming languages, various fields and sources of data, skills, etc. The usage of data science has evolved over the years. In this review article, we have covered a part of data science, i.e., machine learning. Machine learning uses data science for its work. Machines learn through their experience, which helps them to do any work more efficiently. This article includes a comparative study image between human understanding and machine understanding, advantages, applications, and real-time examples of machine learning. Data science is an important game changer in the life of human beings. Since the advent of data science, we have found its benefits and how it leads to a better understanding of people, and how it cherishes individual needs. It has improved business strategies, services provided by them, forecasting, the ability to attend sustainable developments, etc. This study also focuses on a better understanding of data science which will help us to create a better world.Keywords: data science, machine learning, data analytics, artificial intelligence
Procedia PDF Downloads 827052 Hybrid Approach for Software Defect Prediction Using Machine Learning with Optimization Technique
Authors: C. Manjula, Lilly Florence
Abstract:
Software technology is developing rapidly which leads to the growth of various industries. Now-a-days, software-based applications have been adopted widely for business purposes. For any software industry, development of reliable software is becoming a challenging task because a faulty software module may be harmful for the growth of industry and business. Hence there is a need to develop techniques which can be used for early prediction of software defects. Due to complexities in manual prediction, automated software defect prediction techniques have been introduced. These techniques are based on the pattern learning from the previous software versions and finding the defects in the current version. These techniques have attracted researchers due to their significant impact on industrial growth by identifying the bugs in software. Based on this, several researches have been carried out but achieving desirable defect prediction performance is still a challenging task. To address this issue, here we present a machine learning based hybrid technique for software defect prediction. First of all, Genetic Algorithm (GA) is presented where an improved fitness function is used for better optimization of features in data sets. Later, these features are processed through Decision Tree (DT) classification model. Finally, an experimental study is presented where results from the proposed GA-DT based hybrid approach is compared with those from the DT classification technique. The results show that the proposed hybrid approach achieves better classification accuracy.Keywords: decision tree, genetic algorithm, machine learning, software defect prediction
Procedia PDF Downloads 3297051 Dynamics of Piaget’s Cognitive Learning Approach and Vygotsky’s Sociocultural Theory in Different Stages of Medical and Allied Health Education
Authors: Ferissa B. Ablola
Abstract:
The two learning theories which were evidently used in medical education include cognitive and sociocultural frameworks. The interplay of different learning theories in education is vital since most of the existing theories have specific focus of development. In addition, a certain theory is best fit with a particular learning outcome and audience profile. The application of learning theories is education is said to be dynamic and becomes more complex with increasing educational level. This systematic review aims to describe the possible shift from integration of cognitive learning theory to employment of socio-cultural approach in medical and health-allied education over the years among students, educators and the learning institution through systematic review following the PRISMA guidelines. In addition, the changes in teaching modality and individual acceptance of the shift of learning framework among cognitive constructivist and social constructivist will also be documented. This present review may serve as baseline information on the connection of two widely used theories in medical education in different year levels. Further, this study emphasizes the significance of the alignment of different learning theories and combination of insights from several educational frameworks, would permit the creation of a teaching/learning design with real theoretical depth. A more inclusive systematic review is necessary to involve more related studies, and exploration of interaction among other learning theories in health and other fields of study is encouraged.Keywords: learning theory, cognitive, sociocultural, medical education
Procedia PDF Downloads 267050 Identifying and Quantifying Factors Affecting Traffic Crash Severity under Heterogeneous Traffic Flow
Authors: Praveen Vayalamkuzhi, Veeraragavan Amirthalingam
Abstract:
Studies on safety on highways are becoming the need of the hour as over 400 lives are lost every day in India due to road crashes. In order to evaluate the factors that lead to different levels of crash severity, it is necessary to investigate the level of safety of highways and their relation to crashes. In the present study, an attempt is made to identify the factors that contribute to road crashes and to quantify their effect on the severity of road crashes. The study was carried out on a four-lane divided rural highway in India. The variables considered in the analysis includes components of horizontal alignment of highway, viz., straight or curve section; time of day, driveway density, presence of median; median opening; gradient; operating speed; and annual average daily traffic. These variables were considered after a preliminary analysis. The major complexities in the study are the heterogeneous traffic and the speed variation between different classes of vehicles along the highway. To quantify the impact of each of these factors, statistical analyses were carried out using Logit model and also negative binomial regression. The output from the statistical models proved that the variables viz., horizontal components of the highway alignment; driveway density; time of day; operating speed as well as annual average daily traffic show significant relation with the severity of crashes viz., fatal as well as injury crashes. Further, the annual average daily traffic has significant effect on the severity compared to other variables. The contribution of highway horizontal components on crash severity is also significant. Logit models can predict crashes better than the negative binomial regression models. The results of the study will help the transport planners to look into these aspects at the planning stage itself in the case of highways operated under heterogeneous traffic flow condition.Keywords: geometric design, heterogeneous traffic, road crash, statistical analysis, level of safety
Procedia PDF Downloads 3027049 The Advancements of Transformer Models in Part-of-Speech Tagging System for Low-Resource Tigrinya Language
Authors: Shamm Kidane, Ibrahim Abdella, Fitsum Gaim, Simon Mulugeta, Sirak Asmerom, Natnael Ambasager, Yoel Ghebrihiwot
Abstract:
The call for natural language processing (NLP) systems for low-resource languages has become more apparent than ever in the past few years, with the arduous challenges still present in preparing such systems. This paper presents an improved dataset version of the Nagaoka Tigrinya Corpus for Parts-of-Speech (POS) classification system in the Tigrinya language. The size of the initial Nagaoka dataset was incremented, totaling the new tagged corpus to 118K tokens, which comprised the 12 basic POS annotations used previously. The additional content was also annotated manually in a stringent manner, followed similar rules to the former dataset and was formatted in CONLL format. The system made use of the novel approach in NLP tasks and use of the monolingually pre-trained TiELECTRA, TiBERT and TiRoBERTa transformer models. The highest achieved score is an impressive weighted F1-score of 94.2%, which surpassed the previous systems by a significant measure. The system will prove useful in the progress of NLP-related tasks for Tigrinya and similarly related low-resource languages with room for cross-referencing higher-resource languages.Keywords: Tigrinya POS corpus, TiBERT, TiRoBERTa, conditional random fields
Procedia PDF Downloads 1037048 Problems in English into Thai Translation Normally Found in Thai University Students
Authors: Anochao Phetcharat
Abstract:
This research aims to study problems of translation basic knowledge, particularly from English into Thai. The researcher used 38 2nd-year non-English speaking students of Suratthani Rajabhat University as samples. The samples were required to translate an A4-sized article from English into Thai assigned as a part of BEN0202 Translation for Business, a requirement subject for Business English Department, which was also taught by the researcher. After completion of the translation, numerous problems were found and the research grouped them into 4 major types. The normally occurred problems in English-Thai translation works are the lack of knowledge in terms of parts of speech, word-by-word translation employment, misspellings as well as the poor knowledge in English language structure. However, this research is currently under the process of data analysis and shall be completed by the beginning of August. The researcher, nevertheless, predicts that all the above-mentioned problems, will support the researcher’s hypothesizes, that are; 1) the lack of knowledge in terms of parts of speech causes the mistranslation problem; 2) employing word-by-word translation technique hugely results in the mistranslation problem; 3) misspellings yields the mistranslation problem; and 4) the poor knowledge in English language structure also brings about translation errors. The research also predicts that, of all the aforementioned problems, the following ones are found the most, respectively: the poor knowledge in English language structure, word-by-word translation employment, the lack of knowledge in terms of parts of speech, and misspellings.Keywords: problem, student, Thai, translation
Procedia PDF Downloads 4367047 Landslide Hazard Assessment Using Physically Based Mathematical Models in Agricultural Terraces at Douro Valley in North of Portugal
Authors: C. Bateira, J. Fernandes, A. Costa
Abstract:
The Douro Demarked Region (DDR) is a production Porto wine region. On the NE of Portugal, the strong incision of the Douro valley developed very steep slopes, organized with agriculture terraces, have experienced an intense and deep transformation in order to implement the mechanization of the work. The old terrace system, based on stone vertical wall support structure, replaced by terraces with earth embankments experienced a huge terrace instability. This terrace instability has important economic and financial consequences on the agriculture enterprises. This paper presents and develops cartographic tools to access the embankment instability and identify the area prone to instability. The priority on this evaluation is related to the use of physically based mathematical models and develop a validation process based on an inventory of the past embankment instability. We used the shallow landslide stability model (SHALSTAB) based on physical parameters such us cohesion (c’), friction angle(ф), hydraulic conductivity, soil depth, soil specific weight (ϱ), slope angle (α) and contributing areas by Multiple Flow Direction Method (MFD). A terraced area can be analysed by this models unless we have very detailed information representative of the terrain morphology. The slope angle and the contributing areas depend on that. We can achieve that propose using digital elevation models (DEM) with great resolution (pixel with 40cm side), resulting from a set of photographs taken by a flight at 100m high with pixel resolution of 12cm. The slope angle results from this DEM. In the other hand, the MFD contributing area models the internal flow and is an important element to define the spatial variation of the soil saturation. That internal flow is based on the DEM. That is supported by the statement that the interflow, although not coincident with the superficial flow, have important similitude with it. Electrical resistivity monitoring values which related with the MFD contributing areas build from a DEM of 1m resolution and revealed a consistent correlation. That analysis, performed on the area, showed a good correlation with R2 of 0,72 and 0,76 at 1,5m and 2m depth, respectively. Considering that, a DEM with 1m resolution was the base to model the real internal flow. Thus, we assumed that the contributing area of 1m resolution modelled by MFD is representative of the internal flow of the area. In order to solve this problem we used a set of generalized DEMs to build the contributing areas used in the SHALSTAB. Those DEMs, with several resolutions (1m and 5m), were built from a set of photographs with 50cm resolution taken by a flight with 5km high. Using this maps combination, we modelled several final maps of terrace instability and performed a validation process with the contingency matrix. The best final instability map resembles the slope map from a DEM of 40cm resolution and a MFD map from a DEM of 1m resolution with a True Positive Rate (TPR) of 0,97, a False Positive Rate of 0,47, Accuracy (ACC) of 0,53, Precision (PVC) of 0,0004 and a TPR/FPR ratio of 2,06.Keywords: agricultural terraces, cartography, landslides, SHALSTAB, vineyards
Procedia PDF Downloads 177