Search results for: Google Cloud Platform
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 2997

Search results for: Google Cloud Platform

1887 English Grammatical Errors of Arabic Sentence Translations Done by Machine Translations

Authors: Muhammad Fathurridho

Abstract:

Grammar as a rule used by every language to be understood by everyone is always related to syntax and morphology. Arabic grammar is different with another languages’ grammars. It has more rules and difficulties. This paper aims to investigate and describe the English grammatical errors of machine translation systems in translating Arabic sentences, including declarative, exclamation, imperative, and interrogative sentences, specifically in year 2018 which can be supported with artificial intelligence’s role. The Arabic sample sentences which are divided into two; verbal and nominal sentence of several Arabic published texts will be examined as the source language samples. The translated sentences done by several popular online machine translation systems, including Google Translate, Microsoft Bing, Babylon, Facebook, Hellotalk, Worldlingo, Yandex Translate, and Tradukka Translate are the material objects of this research. Descriptive method that will be taken to finish this research will show the grammatical errors of English target language, and classify them. The conclusion of this paper has showed that the grammatical errors of machine translation results are varied and generally classified into morphological, syntactical, and semantic errors in all type of Arabic words (Noun, Verb, and Particle), and it will be one of the evaluations for machine translation’s providers to correct them in order to improve their understandable results.

Keywords: Arabic, Arabic-English translation, machine translation, grammatical errors

Procedia PDF Downloads 149
1886 Application of the Carboxylate Platform in the Consolidated Bioconversion of Agricultural Wastes to Biofuel Precursors

Authors: Sesethu G. Njokweni, Marelize Botes, Emile W. H. Van Zyl

Abstract:

An alternative strategy to the production of bioethanol is by examining the degradability of biomass in a natural system such as the rumen of mammals. This anaerobic microbial community has higher cellulolytic activities than microbial communities from other habitats and degrades cellulose to produce volatile fatty acids (VFA), methane and CO₂. VFAs have the potential to serve as intermediate products for electrochemical conversion to hydrocarbon fuels. In vitro mimicking of this process would be more cost-effective than bioethanol production as it does not require chemical pre-treatment of biomass, a sterile environment or added enzymes. The strategies of the carboxylate platform and the co-cultures of a bovine ruminal microbiota from cannulated cows were combined in order to investigate and optimize the bioconversion of agricultural biomass (apple and grape pomace, citrus pulp, sugarcane bagasse and triticale straw) to high value VFAs as intermediates for biofuel production in a consolidated bioprocess. Optimisation of reactor conditions was investigated using five different ruminal inoculum concentrations; 5,10,15,20 and 25% with fixed pH at 6.8 and temperature at 39 ˚C. The ANKOM 200/220 fiber analyser was used to analyse in vitro neutral detergent fiber (NDF) disappearance of the feedstuffs. Fresh and cryo-frozen (5% DMSO and 50% glycerol for 3 months) rumen cultures were tested for the retainment of fermentation capacity and durability in 72 h fermentations in 125 ml serum vials using a FURO medical solutions 6-valve gas manifold to induce anaerobic conditions. Fermentation of apple pomace, triticale straw, and grape pomace showed no significant difference (P > 0.05) in the effect of 15 and 20 % inoculum concentrations for the total VFA yield. However, high performance liquid chromatographic separation within the two inoculum concentrations showed a significant difference (P < 0.05) in acetic acid yield, with 20% inoculum concentration being the optimum at 4.67 g/l. NDF disappearance of 85% in 96 h and total VFA yield of 11.5 g/l in 72 h (A/P ratio = 2.04) for apple pomace entailed that it was the optimal feedstuff for this process. The NDF disappearance and VFA yield of DMSO (82% NDF disappearance and 10.6 g/l VFA) and glycerol (90% NDF disappearance and 11.6 g/l VFA) stored rumen also showed significantly similar degradability of apple pomace with lack of treatment effect differences compared to a fresh rumen control (P > 0.05). The lack of treatment effects was a positive sign in indicating that there was no difference between the stored samples and the fresh rumen control. Retaining of the fermentation capacity within the preserved cultures suggests that its metabolic characteristics were preserved due to resilience and redundancy of the rumen culture. The amount of degradability and VFA yield within a short span was similar to other carboxylate platforms that have longer run times. This study shows that by virtue of faster rates and high extent of degradability, small scale alternatives to bioethanol such as rumen microbiomes and other natural fermenting microbiomes can be employed to enhance the feasibility of biofuels large-scale implementation.

Keywords: agricultural wastes, carboxylate platform, rumen microbiome, volatile fatty acids

Procedia PDF Downloads 127
1885 Increasing the System Availability of Data Centers by Using Virtualization Technologies

Authors: Chris Ewe, Naoum Jamous, Holger Schrödl

Abstract:

Like most entrepreneurs, data center operators pursue goals such as profit-maximization, improvement of the company’s reputation or basically to exist on the market. Part of those aims is to guarantee a given quality of service. Quality characteristics are specified in a contract called the service level agreement. Central part of this agreement is non-functional properties of an IT service. The system availability is one of the most important properties as it will be shown in this paper. To comply with availability requirements, data center operators can use virtualization technologies. A clear model to assess the effect of virtualization functions on the parts of a data center in relation to the system availability is still missing. This paper aims to introduce a basic model that shows these connections, and consider if the identified effects are positive or negative. Thus, this work also points out possible disadvantages of the technology. In consequence, the paper shows opportunities as well as risks of data center virtualization in relation to system availability.

Keywords: availability, cloud computing IT service, quality of service, service level agreement, virtualization

Procedia PDF Downloads 530
1884 Exploring the Quest for Centralized Identity in Mohsin Hamid's "The Last White Man": Post-Apocalyptic Transformations and Societal Reconfigurations

Authors: Kashifa Khalid, Eesham Fatima

Abstract:

This study aims to analyze the loss of identity and its impact on one’s life in ‘The Last White Man’ by Mohsin Hamid. The theory of Alienation Effect by Bertolt Brecht has been applied to the text as Hamid offers the readers a unique perspective, alluding to significant themes like identity, race, and death. The aspects of defamiliarization align impeccably with the plot, as existence and the corresponding concept of identity seem to have dissolved into utter chaos. This extends from the unexplained transformation to the way the entire world unravels from its general norm into a dystopian mayhem. The characters, starting with the protagonist Anders, have lost their center. One’s own self transforms into the ‘other,’ and the struggle is to get refamiliarized with one’s own self. Alienation and isolation only rise as the construct of race and identity is taken apart brick by brick, ironically at its own pace as many new realities are blown to bits. The inseparable relationship between identity and grief under the ever-looming cloud of ‘death’ is studied in detail. The theoretical framework and thematic aspects harmonize in accordance with the writing style put forth by Hamid, tying all the loose ends together.

Keywords: alienation, chaos, identity, transformation

Procedia PDF Downloads 39
1883 Deep Learning Approach to Trademark Design Code Identification

Authors: Girish J. Showkatramani, Arthi M. Krishna, Sashi Nareddi, Naresh Nula, Aaron Pepe, Glen Brown, Greg Gabel, Chris Doninger

Abstract:

Trademark examination and approval is a complex process that involves analysis and review of the design components of the marks such as the visual representation as well as the textual data associated with marks such as marks' description. Currently, the process of identifying marks with similar visual representation is done manually in United States Patent and Trademark Office (USPTO) and takes a considerable amount of time. Moreover, the accuracy of these searches depends heavily on the experts determining the trademark design codes used to catalog the visual design codes in the mark. In this study, we explore several methods to automate trademark design code classification. Based on recent successes of convolutional neural networks in image classification, we have used several different convolutional neural networks such as Google’s Inception v3, Inception-ResNet-v2, and Xception net. The study also looks into other techniques to augment the results from CNNs such as using Open Source Computer Vision Library (OpenCV) to pre-process the images. This paper reports the results of the various models trained on year of annotated trademark images.

Keywords: trademark design code, convolutional neural networks, trademark image classification, trademark image search, Inception-ResNet-v2

Procedia PDF Downloads 228
1882 The Implantable MEMS Blood Pressure Sensor Model With Wireless Powering And Data Transmission

Authors: Vitaliy Petrov, Natalia Shusharina, Vitaliy Kasymov, Maksim Patrushev, Evgeny Bogdanov

Abstract:

The leading worldwide death reasons are ischemic heart disease and other cardiovascular illnesses. Generally, the common symptom is high blood pressure. Long-time blood pressure control is very important for the prophylaxis, correct diagnosis and timely therapy. Non-invasive methods which are based on Korotkoff sounds are impossible to apply often and for a long time. Implantable devices can combine longtime monitoring with high accuracy of measurements. The main purpose of this work is to create a real-time monitoring system for decreasing the death rate from cardiovascular diseases. These days implantable electronic devices began to play an important role in medicine. Usually implantable devices consist of a transmitter, powering which could be wireless with a special made battery and measurement circuit. Common problems in making implantable devices are short lifetime of the battery, big size and biocompatibility. In these work, blood pressure measure will be the focus because it’s one of the main symptoms of cardiovascular diseases. Our device will consist of three parts: the implantable pressure sensor, external transmitter and automated workstation in a hospital. The Implantable part of pressure sensors could be based on piezoresistive or capacitive technologies. Both sensors have some advantages and some limitations. The Developed circuit is based on a small capacitive sensor which is made of the technology of microelectromechanical systems (MEMS). The Capacitive sensor can provide high sensitivity, low power consumption and minimum hysteresis compared to the piezoresistive sensor. For this device, it was selected the oscillator-based circuit where frequency depends from the capacitance of sensor hence from capacitance one can calculate pressure. The external device (transmitter) used for wireless charging and signal transmission. Some implant devices for these applications are passive, the external device sends radio wave signal on internal LC circuit device. The external device gets reflected the signal from the implant and from a change of frequency is possible to calculate changing of capacitance and then blood pressure. However, this method has some disadvantages, such as the patient position dependence and static using. Developed implantable device doesn’t have these disadvantages and sends blood pressure data to the external part in real-time. The external device continuously sends information about blood pressure to hospital cloud service for analysis by a physician. Doctor’s automated workstation at the hospital also acts as a dashboard, which displays actual medical data of patients (which require attention) and stores it in cloud service. Usually, critical heart conditions occur few hours before heart attack but the device is able to send an alarm signal to the hospital for an early action of medical service. The system was tested with wireless charging and data transmission. These results can be used for ASIC design for MEMS pressure sensor.

Keywords: MEMS sensor, RF power, wireless data, oscillator-based circuit

Procedia PDF Downloads 587
1881 Modelling of Reactive Methodologies in Auto-Scaling Time-Sensitive Services With a MAPE-K Architecture

Authors: Óscar Muñoz Garrigós, José Manuel Bernabeu Aubán

Abstract:

Time-sensitive services are the base of the cloud services industry. Keeping low service saturation is essential for controlling response time. All auto-scalable services make use of reactive auto-scaling. However, reactive auto-scaling has few in-depth studies. This presentation shows a model for reactive auto-scaling methodologies with a MAPE-k architecture. Queuing theory can compute different properties of static services but lacks some parameters related to the transition between models. Our model uses queuing theory parameters to relate the transition between models. It associates MAPE-k related times, the sampling frequency, the cooldown period, the number of requests that an instance can handle per unit of time, the number of incoming requests at a time instant, and a function that describes the acceleration in the service's ability to handle more requests. This model is later used as a solution to horizontally auto-scale time-sensitive services composed of microservices, reevaluating the model’s parameters periodically to allocate resources. The solution requires limiting the acceleration of the growth in the number of incoming requests to keep a constrained response time. Business benefits determine such limits. The solution can add a dynamic number of instances and remains valid under different system sizes. The study includes performance recommendations to improve results according to the incoming load shape and business benefits. The exposed methodology is tested in a simulation. The simulator contains a load generator and a service composed of two microservices, where the frontend microservice depends on a backend microservice with a 1:1 request relation ratio. A common request takes 2.3 seconds to be computed by the service and is discarded if it takes more than 7 seconds. Both microservices contain a load balancer that assigns requests to the less loaded instance and preemptively discards requests if they are not finished in time to prevent resource saturation. When load decreases, instances with lower load are kept in the backlog where no more requests are assigned. If the load grows and an instance in the backlog is required, it returns to the running state, but if it finishes the computation of all requests and is no longer required, it is permanently deallocated. A few load patterns are required to represent the worst-case scenario for reactive systems: the following scenarios test response times, resource consumption and business costs. The first scenario is a burst-load scenario. All methodologies will discard requests if the rapidness of the burst is high enough. This scenario focuses on the number of discarded requests and the variance of the response time. The second scenario contains sudden load drops followed by bursts to observe how the methodology behaves when releasing resources that are lately required. The third scenario contains diverse growth accelerations in the number of incoming requests to observe how approaches that add a different number of instances can handle the load with less business cost. The exposed methodology is compared against a multiple threshold CPU methodology allocating/deallocating 10 or 20 instances, outperforming the competitor in all studied metrics.

Keywords: reactive auto-scaling, auto-scaling, microservices, cloud computing

Procedia PDF Downloads 90
1880 Quantitative Seismic Interpretation in the LP3D Concession, Central of the Sirte Basin, Libya

Authors: Tawfig Alghbaili

Abstract:

LP3D Field is located near the center of the Sirt Basin in the Marada Trough approximately 215 km south Marsa Al Braga City. The Marada Trough is bounded on the west by a major fault, which forms the edge of the Beda Platform, while on the east, a bounding fault marks the edge of the Zelten Platform. The main reservoir in the LP3D Field is Upper Paleocene Beda Formation. The Beda Formation is mainly limestone interbedded with shale. The reservoir average thickness is 117.5 feet. To develop a better understanding of the characterization and distribution of the Beda reservoir, quantitative seismic data interpretation has been done, and also, well logs data were analyzed. Six reflectors corresponding to the tops of the Beda, Hagfa Shale, Gir, Kheir Shale, Khalifa Shale, and Zelten Formations were picked and mapped. Special work was done on fault interpretation part because of the complexities of the faults at the structure area. Different attribute analyses were done to build up more understanding of structures lateral extension and to view a clear image of the fault blocks. Time to depth conversion was computed using velocity modeling generated from check shot and sonic data. The simplified stratigraphic cross-section was drawn through the wells A1, A2, A3, and A4-LP3D. The distribution and the thickness variations of the Beda reservoir along the study area had been demonstrating. Petrophysical analysis of wireline logging also was done and Cross plots of some petrophysical parameters are generated to evaluate the lithology of reservoir interval. Structure and Stratigraphic Framework was designed and run to generate different model like faults, facies, and petrophysical models and calculate the reservoir volumetric. This study concluded that the depth structure map of the Beda formation shows the main structure in the area of study, which is north to south faulted anticline. Based on the Beda reservoir models, volumetric for the base case has been calculated and it has STOIIP of 41MMSTB and Recoverable oil of 10MMSTB. Seismic attributes confirm the structure trend and build a better understanding of the fault system in the area.

Keywords: LP3D Field, Beda Formation, reservoir models, Seismic attributes

Procedia PDF Downloads 209
1879 Political Views and Information and Communication Technology (ICT) in Tertiary Institutions in Achieving the Millennium Development Goals (MDGS)

Authors: Perpetual Nwakaego Ibe

Abstract:

The Millennium Development Goals (MDGs), were an integrated project formed to eradicate many unnatural situations the citizens of the third world country may found themselves in. The MDGs, to be a sustainable project for the future depends 100% on the actions of governments, multilateral institutions and civil society. This paper first looks at the political views on the MDGs and relates it to the current electoral situations around the country by underlining the drastic changes over the few months. The second part of the paper presents ICT in tertiary institutions as one of the solutions in terms of the success of the MDGs. ICT is vital in all phases of educational process and development of the cloud connectivity is an added advantage of Information and Communication Technology (ICT) for sharing a common data bank for research purposes among UNICEF, RED CROSS, NPS, INEC, NMIC, and WHO. Finally, the paper concludes with areas that needs twigging and recommendations for the tertiary institutions committed to delivering an ambitious set of goals. A combination of observation, and document materials for data gathering was employed as the methodology for carrying out this research.

Keywords: MDG, ICT, data bank, database

Procedia PDF Downloads 197
1878 Sound Source Localisation and Augmented Reality for On-Site Inspection of Prefabricated Building Components

Authors: Jacques Cuenca, Claudio Colangeli, Agnieszka Mroz, Karl Janssens, Gunther Riexinger, Antonio D'Antuono, Giuseppe Pandarese, Milena Martarelli, Gian Marco Revel, Carlos Barcena Martin

Abstract:

This study presents an on-site acoustic inspection methodology for quality and performance evaluation of building components. The work focuses on global and detailed sound source localisation, by successively performing acoustic beamforming and sound intensity measurements. A portable experimental setup is developed, consisting of an omnidirectional broadband acoustic source and a microphone array and sound intensity probe. Three main acoustic indicators are of interest, namely the sound pressure distribution on the surface of components such as walls, windows and junctions, the three-dimensional sound intensity field in the vicinity of junctions, and the sound transmission loss of partitions. The measurement data is post-processed and converted into a three-dimensional numerical model of the acoustic indicators with the help of the simultaneously acquired geolocation information. The three-dimensional acoustic indicators are then integrated into an augmented reality platform superimposing them onto a real-time visualisation of the spatial environment. The methodology thus enables a measurement-supported inspection process of buildings and the correction of errors during construction and refurbishment. Two experimental validation cases are shown. The first consists of a laboratory measurement on a full-scale mockup of a room, featuring a prefabricated panel. The latter is installed with controlled defects such as lack of insulation and joint sealing material. It is demonstrated that the combined acoustic and augmented reality tool is capable of identifying acoustic leakages from the building defects and assist in correcting them. The second validation case is performed on a prefabricated room at a near-completion stage in the factory. With the help of the measurements and visualisation tools, the homogeneity of the partition installation is evaluated and leakages from junctions and doors are identified. Furthermore, the integration of acoustic indicators together with thermal and geometrical indicators via the augmented reality platform is shown.

Keywords: acoustic inspection, prefabricated building components, augmented reality, sound source localization

Procedia PDF Downloads 376
1877 The Regulation of Reputational Information in the Sharing Economy

Authors: Emre Bayamlıoğlu

Abstract:

This paper aims to provide an account of the legal and the regulative aspects of the algorithmic reputation systems with a special emphasis on the sharing economy (i.e., Uber, Airbnb, Lyft) business model. The first section starts with an analysis of the legal and commercial nature of the tripartite relationship among the parties, namely, the host platform, individual sharers/service providers and the consumers/users. The section further examines to what extent an algorithmic system of reputational information could serve as an alternative to legal regulation. Shortcomings are explained and analyzed with specific examples from Airbnb Platform which is a pioneering success in the sharing economy. The following section focuses on the issue of governance and control of the reputational information. The section first analyzes the legal consequences of algorithmic filtering systems to detect undesired comments and how a delicate balance could be struck between the competing interests such as freedom of speech, privacy and the integrity of the commercial reputation. The third section deals with the problem of manipulation by users. Indeed many sharing economy businesses employ certain techniques of data mining and natural language processing to verify consistency of the feedback. Software agents referred as "bots" are employed by the users to "produce" fake reputation values. Such automated techniques are deceptive with significant negative effects for undermining the trust upon which the reputational system is built. The third section is devoted to explore the concerns with regard to data mobility, data ownership, and the privacy. Reputational information provided by the consumers in the form of textual comment may be regarded as a writing which is eligible to copyright protection. Algorithmic reputational systems also contain personal data pertaining both the individual entrepreneurs and the consumers. The final section starts with an overview of the notion of reputation as a communitarian and collective form of referential trust and further provides an evaluation of the above legal arguments from the perspective of public interest in the integrity of reputational information. The paper concludes with certain guidelines and design principles for algorithmic reputation systems, to address the above raised legal implications.

Keywords: sharing economy, design principles of algorithmic regulation, reputational systems, personal data protection, privacy

Procedia PDF Downloads 464
1876 Establishing Student Support Strategies for Virtual Learning in Learning Management System Based on Grounded Theory

Authors: Farhad Shafiepour Motlagh, Narges Salehi

Abstract:

Purpose: The purpose of this study was to support student strategies for virtual learning in the learning management system. Methodology: The research method was based on grounded theory. The statistical population included all the articles of the ten years 2022-2010, and the sampling method was purposeful to the extent of theoretical saturation (n=31 ). Data collection was done by referring to the authoritative scientific databases of Emerald, Springer, Elsevier, Google Scholar, Sage Publication, and Science Direct. For data analysis, open coding, axial coding, and selective coding were used. Results: The results showed that causal conditions include cognitive empowerment (comprehension, analysis, composition), emotional empowerment (learning motivation, involvement in the learning system, enthusiasm for learning), psychomotor empowerment (learning to master, internalizing learning skills, creativity in learning). Conclusion: Supporting students requires their empowerment in three dimensions: cognitive, emotional empowerment, and psychomotor empowerment. In such a way that by introducing them to enter the learning management system, the capacities of the system, the toolkit of learning in the system, improve the motivation to learn in them, and in such a case, by learning more in the learning management system, they will reach mastery learning.

Keywords: student support, virtual education, learning management system, electronic

Procedia PDF Downloads 298
1875 A Multi Cordic Architecture on FPGA Platform

Authors: Ahmed Madian, Muaz Aljarhi

Abstract:

Coordinate Rotation Digital Computer (CORDIC) is a unique digital computing unit intended for the computation of mathematical operations and functions. This paper presents a multi-CORDIC processor that integrates different CORDIC architectures on a single FPGA chip and allows the user to select the CORDIC architecture to proceed with based on what he wants to calculate and his/her needs. Synthesis show that radix 2 CORDIC has the lowest clock delay, radix 8 CORDIC has the highest LUT usage and lowest register usage while Hybrid Radix 4 CORDIC had the highest clock delay.

Keywords: multi, CORDIC, FPGA, processor

Procedia PDF Downloads 466
1874 Nutritional Quality Assessment and Safety Evaluation of Food Crops

Authors: Olawole Emmanuel Aina, Liziwe Lizbeth Mugivhisa, Joshua Oluwole Olowoyo, Chikwela Lawrence Obi

Abstract:

In sustained and consistent efforts to improve food security, numerous and different methods are proposed and used in the production of food crops, and farm produce to meet the demands of consumers. However, unregulated and indiscriminate methods of production present another problem that may expose consumers of these food crops to potential health risks. Therefore, it is imperative that a thorough assessment of farm produce is carried out due to the growing trend of health-conscious consumers preference for minimally processed or raw farm produce. This study evaluated the safety and nutritional quality of food crops. The objectives were to compare the nutritional quality of organic and inorganic farm produce in one hand and, on the other, evaluate the safety of farm produce with respect to trace metal and pathogenic contamination. We conducted a broad systematic search of peer-reviewed published literatures from databases and search engines such as science direct, web-of-science, Google scholar, and Scopus. This study concluded that there is no conclusive evidence to support the notion of nutritional superiority of organic food crops over their inorganic counterparts and there are documented reports of pathogenic and metal contaminations of food crops.

Keywords: food crops, fruits and vegetables, pathogens, nutrition, trace metals

Procedia PDF Downloads 79
1873 Double Encrypted Data Communication Using Cryptography and Steganography

Authors: Adine Barett, Jermel Watson, Anteneh Girma, Kacem Thabet

Abstract:

In information security, secure communication of data across networks has always been a problem at the forefront. Transfer of information across networks is susceptible to being exploited by attackers engaging in malicious activity. In this paper, we leverage steganography and cryptography to create a layered security solution to protect the information being transmitted. The first layer of security leverages crypto- graphic techniques to scramble the information so that it cannot be deciphered even if the steganography-based layer is compromised. The second layer of security relies on steganography to disguise the encrypted in- formation so that it cannot be seen. We consider three cryptographic cipher methods in the cryptography layer, namely, Playfair cipher, Blowfish cipher, and Hills cipher. Then, the encrypted message is passed through the least significant bit (LSB) to the steganography algorithm for further encryption. Both encryption approaches are combined efficiently to help secure information in transit over a network. This multi-layered encryption is a solution that will benefit cloud platforms, social media platforms and networks that regularly transfer private information such as banks and insurance companies.

Keywords: cryptography, steganography, layered security, Cipher, encryption

Procedia PDF Downloads 77
1872 Gut-Microbiota-Brain-Axis, Leaky Gut, Leaky Brain: Pathophysiology of Second Brain Aging and Alzheimer’s Disease- A Neuroscientific Riddle

Authors: Bilal Ahmad

Abstract:

Alzheimer’s disease (AD) is one of the most common neurodegenerative illnesses. However, how Gut-microbiota plays a role in the pathogenesis of AD is not well elucidated. The purpose of this literature review is to summarize and understand the current findings that may elucidate the gut microbiota's role in the development of AD. Methods: A literature review of all the relevant papers known to the author was conducted. Relevant articles, abstracts and research papers were collected from well-accepted web sources like PubMed, PMC, and Google Scholar. Results: Recent studies have shown that Gut-microbiota has an important role in the progression of AD via Gut-Microbiota-Brain Axis. The onset of AD supports the ‘Hygiene Hypothesis’, which shows that AD might begin in the Gut, causing dysbiosis, which interferes with the intestinal barrier by releasing pro-inflammatory cytokines and making its way up to the brain via the blood-brain barrier (BBB). Molecular mechanisms lipopolysaccharides and serotonin kynurenine (tryptophan) pathways have a direct association with inflammation, the immune system, neurodegeneration, and AD. Conclusion: The studies helped to analyze the molecular basis of AD, other neurological conditions like depression, autism, and Parkinson's disease and how they are linked to Gut-microbiota. Further, studies to explore the therapeutic effects of probiotics in AD and cognitive enhancement should be warranted to provide significant clinical and practical value.

Keywords: gut-microbiota, Alzheimer’s disease, second brain aging, lipopolysaccharides, short-chain fatty acids

Procedia PDF Downloads 39
1871 Urbanization and House Water Supply in Nigeria

Authors: Oluronke Odunjo

Abstract:

The world is becoming increasingly urbanized and Nigeria is not left out. One of the indicators of human developments is housing and as such, water is needed by households for survival. This Paper assesses sources of water being used by residents in the newly urbanized areas of Ogbomoso, Southwest, Nigeria. Multistage sampling technique was used and Oke-Adunin Community was purposively selected for the study as it has large concentration of staff and students of Ladoke Akintola University of Technology. The area was captured with Google earth and two hundred and twenty two inhabited houses were found. Questionnaire was the instrument for data collection which was administered using total enumeration technique. Data obtained however, were analyzed with descriptive and inferential statistical analyses. Findings revealed that most of the respondents were male, while 36.03% house owners were between the ages of 46 and 55 years. Sources of water used by residents include well (56.94%), water vendors (17.77%), rain (15.29%) and borehole (3.72%). Distance travelled by house owners to sources of water was as high as 5.06 metres, resulting into stress (30.00 %), depression (25.00%) and aggressiveness (18.75%). Result of correlation analysis between the sources of water of respondents and disease prevalence showed that both rain water and water vendor had very strong positive correlation with typhoid, diarrhea and dysentery, while well water only had positive correlation with dysentery. Recommendations were therefore, proffered towards solving the problems associated with water in the area.

Keywords: newly urbanized area, Ogbomoso, sources of water, residents

Procedia PDF Downloads 195
1870 A Framework for Automated Nuclear Waste Classification

Authors: Seonaid Hume, Gordon Dobie, Graeme West

Abstract:

Detecting and localizing radioactive sources is a necessity for safe and secure decommissioning of nuclear facilities. An important aspect for the management of the sort-and-segregation process is establishing the spatial distributions and quantities of the waste radionuclides, their type, corresponding activity, and ultimately classification for disposal. The data received from surveys directly informs decommissioning plans, on-site incident management strategies, the approach needed for a new cell, as well as protecting the workforce and the public. Manual classification of nuclear waste from a nuclear cell is time-consuming, expensive, and requires significant expertise to make the classification judgment call. Also, in-cell decommissioning is still in its relative infancy, and few techniques are well-developed. As with any repetitive and routine tasks, there is the opportunity to improve the task of classifying nuclear waste using autonomous systems. Hence, this paper proposes a new framework for the automatic classification of nuclear waste. This framework consists of five main stages; 3D spatial mapping and object detection, object classification, radiological mapping, source localisation based on gathered evidence and finally, waste classification. The first stage of the framework, 3D visual mapping, involves object detection from point cloud data. A review of related applications in other industries is provided, and recommendations for approaches for waste classification are made. Object detection focusses initially on cylindrical objects since pipework is significant in nuclear cells and indeed any industrial site. The approach can be extended to other commonly occurring primitives such as spheres and cubes. This is in preparation of stage two, characterizing the point cloud data and estimating the dimensions, material, degradation, and mass of the objects detected in order to feature match them to an inventory of possible items found in that nuclear cell. Many items in nuclear cells are one-offs, have limited or poor drawings available, or have been modified since installation, and have complex interiors, which often and inadvertently pose difficulties when accessing certain zones and identifying waste remotely. Hence, this may require expert input to feature match objects. The third stage, radiological mapping, is similar in order to facilitate the characterization of the nuclear cell in terms of radiation fields, including the type of radiation, activity, and location within the nuclear cell. The fourth stage of the framework takes the visual map for stage 1, the object characterization from stage 2, and radiation map from stage 3 and fuses them together, providing a more detailed scene of the nuclear cell by identifying the location of radioactive materials in three dimensions. The last stage involves combining the evidence from the fused data sets to reveal the classification of the waste in Bq/kg, thus enabling better decision making and monitoring for in-cell decommissioning. The presentation of the framework is supported by representative case study data drawn from an application in decommissioning from a UK nuclear facility. This framework utilises recent advancements of the detection and mapping capabilities of complex radiation fields in three dimensions to make the process of classifying nuclear waste faster, more reliable, cost-effective and safer.

Keywords: nuclear decommissioning, radiation detection, object detection, waste classification

Procedia PDF Downloads 198
1869 Intimate Partner Violence Concerns during COVID-19 Pandemic

Authors: Fatemeh Abdollahi, Munn-Sann Lye, Jamshid Yazdani Charati, Mehran Zarghami

Abstract:

Background: In March 2020, the World Health Organization (WHO) declared the outbreak of a new coronavirus disease, COVID-19, as a public health concern and pandemic. This situation is generating psychological consequences such as stress, anxiety, depression, and intimate partner violence (IPV) throughout the population. This is a brief note on the magnitude of this threat and different ways for abused women to minimize the effects of it in their daily life. Methods: A literature review was conducted using the MEDLINE, PSYCHINFO, and SCIENCE DIRECT databases. The keywords used included intimate partner violence, abuse, victims, pandemic, quarantine, coronavirus, and COVID-19. A Google search was also conducted using these words to identify reports published in non-indexed health care and social science journals. The literature search was restricted to English language studies. Results: The prevalence of IPV and its consequences are rising during such a pandemic. Having sufficient support from healthcare workers and acquaintances is critical for women in such circumstances. Conclusion: Community members, healthcare providers, governments, and policymakers should be informed of the increased risk of IPV during such a pandemic. They should provide a supporting structure for abused women. Social networking is also a good approach that could help abusive women during this situation.

Keywords: covid-19, intimate partner violence, pandemic, women

Procedia PDF Downloads 62
1868 Predictive Analytics for Theory Building

Authors: Ho-Won Jung, Donghun Lee, Hyung-Jin Kim

Abstract:

Predictive analytics (data analysis) uses a subset of measurements (the features, predictor, or independent variable) to predict another measurement (the outcome, target, or dependent variable) on a single person or unit. It applies empirical methods in statistics, operations research, and machine learning to predict the future, or otherwise unknown events or outcome on a single or person or unit, based on patterns in data. Most analyses of metabolic syndrome are not predictive analytics but statistical explanatory studies that build a proposed model (theory building) and then validate metabolic syndrome predictors hypothesized (theory testing). A proposed theoretical model forms with causal hypotheses that specify how and why certain empirical phenomena occur. Predictive analytics and explanatory modeling have their own territories in analysis. However, predictive analytics can perform vital roles in explanatory studies, i.e., scientific activities such as theory building, theory testing, and relevance assessment. In the context, this study is to demonstrate how to use our predictive analytics to support theory building (i.e., hypothesis generation). For the purpose, this study utilized a big data predictive analytics platform TM based on a co-occurrence graph. The co-occurrence graph is depicted with nodes (e.g., items in a basket) and arcs (direct connections between two nodes), where items in a basket are fully connected. A cluster is a collection of fully connected items, where the specific group of items has co-occurred in several rows in a data set. Clusters can be ranked using importance metrics, such as node size (number of items), frequency, surprise (observed frequency vs. expected), among others. The size of a graph can be represented by the numbers of nodes and arcs. Since the size of a co-occurrence graph does not depend directly on the number of observations (transactions), huge amounts of transactions can be represented and processed efficiently. For a demonstration, a total of 13,254 metabolic syndrome training data is plugged into the analytics platform to generate rules (potential hypotheses). Each observation includes 31 predictors, for example, associated with sociodemographic, habits, and activities. Some are intentionally included to get predictive analytics insights on variable selection such as cancer examination, house type, and vaccination. The platform automatically generates plausible hypotheses (rules) without statistical modeling. Then the rules are validated with an external testing dataset including 4,090 observations. Results as a kind of inductive reasoning show potential hypotheses extracted as a set of association rules. Most statistical models generate just one estimated equation. On the other hand, a set of rules (many estimated equations from a statistical perspective) in this study may imply heterogeneity in a population (i.e., different subpopulations with unique features are aggregated). Next step of theory development, i.e., theory testing, statistically tests whether a proposed theoretical model is a plausible explanation of a phenomenon interested in. If hypotheses generated are tested statistically with several thousand observations, most of the variables will become significant as the p-values approach zero. Thus, theory validation needs statistical methods utilizing a part of observations such as bootstrap resampling with an appropriate sample size.

Keywords: explanatory modeling, metabolic syndrome, predictive analytics, theory building

Procedia PDF Downloads 273
1867 Sustainable Separation of Nicotine from Its Aqueous Solutions

Authors: Zoran Visak, Joana Lopes, Vesna Najdanovic-Visak

Abstract:

Within this study, the separation of nicotine from its aqueous solutions, using inorganic salt sodium chloride or ionic liquid (molten salt) ECOENG212® as salting-out media, was carried out. Thus, liquid-liquid equilibria of the ternary solutions (nicotine+water+NaCl) and (nicotine+water+ECOENG212®) were determined at ambient pressure, 0.1 MPa, at three temperatures. The related phase diagrams were constructed in two manners: by adding the determined cloud-points and by the chemical analysis of phases in equilibrium (tie-line data). The latter were used to calculate two important separation parameters - partition coefficients of nicotine and separation factors. The impacts of the initial compositions of the mother solutions and of temperature on the liquid-liquid phase separation and partition coefficients were analyzed and discussed. The results obtained clearly showed that both investigated salts are good salting-out media for the efficient and sustainable separation of nicotine from its solutions with water. However, when compared, sodium chloride exhibited much better separation performance than the ionic liquid.

Keywords: nicotine, liquid-liquid separation, inorganic salt, ionic liquid

Procedia PDF Downloads 307
1866 Personalized Tissues and Organs Replacement – a Peek into the Future

Authors: Asaf Toker

Abstract:

Matricelf developed a technology that enables the production of autologous engineered tissue composed of matrix and cells derived from patients Omentum biopsy. The platform showed remarkable pre-clinical results for several medical conditions. The company recently licensed the technology that enabled scientist at Tel Aviv university that 3D printed a human heart from human cells and matrix for the first time in human history. The company plans to conduct its first human clinical trial for Acute Spinal Cord Injury (SCI) early in 2023.

Keywords: tissue engineering, regenerative medicine, spinal Cord Injury, autologous implants, iPSC

Procedia PDF Downloads 116
1865 Statistical Comparison of Machine and Manual Translation: A Corpus-Based Study of Gone with the Wind

Authors: Yanmeng Liu

Abstract:

This article analyzes and compares the linguistic differences between machine translation and manual translation, through a case study of the book Gone with the Wind. As an important carrier of human feeling and thinking, the literature translation poses a huge difficulty for machine translation, and it is supposed to expose distinct translation features apart from manual translation. In order to display linguistic features objectively, tentative uses of computerized and statistical evidence to the systematic investigation of large scale translation corpora by using quantitative methods have been deployed. This study compiles bilingual corpus with four versions of Chinese translations of the book Gone with the Wind, namely, Piao by Chunhai Fan, Piao by Huairen Huang, translations by Google Translation and Baidu Translation. After processing the corpus with the software of Stanford Segmenter, Stanford Postagger, and AntConc, etc., the study analyzes linguistic data and answers the following questions: 1. How does the machine translation differ from manual translation linguistically? 2. Why do these deviances happen? This paper combines translation study with the knowledge of corpus linguistics, and concretes divergent linguistic dimensions in translated text analysis, in order to present linguistic deviances in manual and machine translation. Consequently, this study provides a more accurate and more fine-grained understanding of machine translation products, and it also proposes several suggestions for machine translation development in the future.

Keywords: corpus-based analysis, linguistic deviances, machine translation, statistical evidence

Procedia PDF Downloads 139
1864 Perception of Nursing Students’ Engagement With Emergency Remote Learning During COVID 19 Pandemic

Authors: Jansirani Natarajan, Mickael Antoinne Joseph

Abstract:

The COVID-19 pandemic has interrupted face-to-face education and forced universities into an emergency remote teaching curriculum over a short duration. This abrupt transition in the Spring 2020 semester left both faculty and students without proper preparation for continuing higher education in an online environment. Online learning took place in different formats, including fully synchronous, fully asynchronous, and blended in our university through the e-learning platform MOODLE. Studies have shown that students’ engagement, is a critical factor for optimal online teaching. Very few studies have assessed online engagement with ERT during the COVID-19 pandemic. Purpose: Therefore, this study, sought to understand how the sudden transition to emergency remote teaching impacted nursing students’ engagement with online courses in a Middle Eastern public university. Method: A cross-sectional descriptive research design was adopted in this study. Data were collected through a self-reported online survey using Dixon’s online students’ engagement questionnaire from a sample of 177 nursing students after the ERT learning semester. Results The maximum possible engagement score was 95, and the maximum scores in the domains of skills engagement, emotional engagement, participation engagement, and performance engagement were 30, 25, 30, and 10 respectively. Dixson (2010) noted that a mean item score of ≥3.5 (total score of ≥66.5) represents a highly engaged student. The majority of the participants were females (71.8%) and 84.2% were regular BSN students. Most of them (32.2%) were second-year students and 52% had a CGPA between 2 and 3. Most participants (56.5%) had low engagement scores with ERT learning during the COVID lockdown. Among the four engagement domains, 78% had low engagement scores for the participation domain. There was no significant association found between the engagement and the demographic characteristics of the participants. Conclusion The findings supported the importance of engaging students in all four categories skill, emotional, performance, and participation. Based on the results, training sessions were organized for faculty on various strategies for engaging nursing students in all domains by using the facilities available in the MOODLE (online e-learning platform). It added value as a dashboard of information regarding ERT for the administrators and nurse educators to introduce numerous active learning strategies to improve the quality of teaching and learning of nursing students in the University.

Keywords: engagement, perception, emergency remote learning, COVID-19

Procedia PDF Downloads 57
1863 John Cunningham Virus Interaction with Multiple Sclerosis Disease Progression

Authors: Sina Mahdavi

Abstract:

Background and Objective: Multiple sclerosis (MS) is the most common inflammatory autoimmune disease of the central nervous system (CNS) that affects the myelination process in the CNS. Complex interactions of various "environmental or infectious" factors may act as triggers in autoimmunity and disease progression. The association between viral infections, especially the John Cunningham virus (JCV) and MS is one potential cause that is not well understood. This study aims to summarize the available data on JCV infection in MS disease progression. Materials and Methods: For this study, the keywords "Multiple sclerosis", " John Cunningham virus ", and "central nervous system" in the databases PubMed, Google Scholar, Sid, and MagIran between 2019 and 2022 were searched, and 12 articles were chosen, studied, and analyzed. Results: MS patients are candidates for natalizumab therapy, which inhibits lymphocyte migration and increases the risk of progressive multifocal leukoencephalopathy (PML), a rare lytic infection of glial cells caused by JCV. Oligodendrocytes may be the target of JCV infection in the central nervous system (CNS). Conclusion: There is a high expression of JCV during the natalizumab treatment period for MS patients, suggesting that the virus may play a role in the development of MS by inducing an inflammatory state. Therefore, it is necessary to evaluate anti-JCV antibody serum as an important risk factor for the development of PML before deciding on the treatment course for these patients.

Keywords: multiple sclerosis, John Cunningham virus, central nervous system, autoimmunity

Procedia PDF Downloads 130
1862 Soft Robotic System for Mechanical Stimulation of Scaffolds During Dynamic Cell Culture

Authors: Johanna Perdomo, Riki Lamont, Edmund Pickering, Naomi C. Paxton, Maria A. Woodruff

Abstract:

Background: Tissue Engineering (TE) has combined advanced materials, such as biomaterials, to create affordable scaffolds and dynamic systems to generate stimulation of seeded cells on these scaffolds, improving and maintaining the cellular growth process in a cell culture. However, Few TE skin products have been clinically translated, and more research is required to produce highly biomimetic skin substitutes that mimic the native elasticity of skin in a controlled manner. Therefore, this work will be focused on the fabrication of a novel mechanical system to enhance the TE treatment approaches for the reparation of damaged tissue skin. Aims: To archive this, a soft robotic device will be created to emulate different deformation of skin stress. The design of this soft robot will allow the attachment of scaffolds, which will then be mechanically actuated. This will provide a novel and highly adaptable platform for dynamic cell culture. Methods: Novel, low-cost soft robot is fabricated via 3D printed moulds and silicone. A low cost, electro-mechanical device was constructed to actuate the soft robot through the controlled combination of positive and negative air pressure to control the different state of movements. Mechanical tests were conducted to assess the performance and calibration of each electronic component. Similarly, pressure-displacement test was performed on scaffolds, which were attached to the soft robot, applying various mechanical loading regimes. Lastly, digital image correlation test was performed to obtain strain distributions over the soft robot’s surface. Results: The control system can control and stabilise positive pressure changes for long hours. Similarly, pressure-displacement test demonstrated that scaffolds with 5µm of diameter and wavy geometry can displace at 100%, applying a maximum pressure of 1.5 PSI. Lastly, during the inflation state, the displacement of silicone was measured using DIC method, and this showed a parameter of 4.78 mm and strain of 0.0652. Discussion And Conclusion: The developed soft robot system provides a novel and low-cost platform for the dynamic actuation of tissue scaffolds with a target towards dynamic cell culture.

Keywords: soft robot, tissue engineering, mechanical stimulation, dynamic cell culture, bioreactor

Procedia PDF Downloads 94
1861 A Four-Step Ortho-Rectification Procedure for Geo-Referencing Video Streams from a Low-Cost UAV

Authors: B. O. Olawale, C. R. Chatwin, R. C. D. Young, P. M. Birch, F. O. Faithpraise, A. O. Olukiran

Abstract:

Ortho-rectification is the process of geometrically correcting an aerial image such that the scale is uniform. The ortho-image formed from the process is corrected for lens distortion, topographic relief, and camera tilt. This can be used to measure true distances, because it is an accurate representation of the Earth’s surface. Ortho-rectification and geo-referencing are essential to pin point the exact location of targets in video imagery acquired at the UAV platform. This can only be achieved by comparing such video imagery with an existing digital map. However, it is only when the image is ortho-rectified with the same co-ordinate system as an existing map that such a comparison is possible. The video image sequences from the UAV platform must be geo-registered, that is, each video frame must carry the necessary camera information before performing the ortho-rectification process. Each rectified image frame can then be mosaicked together to form a seamless image map covering the selected area. This can then be used for comparison with an existing map for geo-referencing. In this paper, we present a four-step ortho-rectification procedure for real-time geo-referencing of video data from a low-cost UAV equipped with multi-sensor system. The basic procedures for the real-time ortho-rectification are: (1) Decompilation of video stream into individual frames; (2) Finding of interior camera orientation parameters; (3) Finding the relative exterior orientation parameters for each video frames with respect to each other; (4) Finding the absolute exterior orientation parameters, using self-calibration adjustment with the aid of a mathematical model. Each ortho-rectified video frame is then mosaicked together to produce a 2-D planimetric mapping, which can be compared with a well referenced existing digital map for the purpose of georeferencing and aerial surveillance. A test field located in Abuja, Nigeria was used for testing our method. Fifteen minutes video and telemetry data were collected using the UAV and the data collected were processed using the four-step ortho-rectification procedure. The results demonstrated that the geometric measurement of the control field from ortho-images are more reliable than those from original perspective photographs when used to pin point the exact location of targets on the video imagery acquired by the UAV. The 2-D planimetric accuracy when compared with the 6 control points measured by a GPS receiver is between 3 to 5 meters.

Keywords: geo-referencing, ortho-rectification, video frame, self-calibration

Procedia PDF Downloads 476
1860 Integrating One Health Approach with National Policies to Improve Health Security post-COVID-19 in Vietnam

Authors: Yasser Sanad, Thu Trang Dao

Abstract:

Introduction: Implementing the One Health (OH) approach requires an integrated, interdisciplinary, and cross-sectoral methodology. OH is a key tool for developing and implementing programs and projects and includes developing ambitious policies that consider the common needs and benefits of human, animal, plant, and ecosystem health. OH helps humanity readjust its path to environmentally friendly and impartial sustainability. As co-leader of the Global Health Security Agenda’s Zoonotic Disease Action Package, Vietnam pioneered a strong OH approach to effectively address early waves of the COVID-19 outbreak in-country. Context and Aim: The repeated surges in COVID-19 in Vietnam challenged the capabilities of the national system and disclosed the gaps in multi-sectoral coordination and resilience. To address this, FHI 360 advocated for the standardization of the OH platform by government actors to increase the resiliency of the system during and post COVID-19. Methods: FHI 360 coordinated technical resources to develop and implement evidence-based OH policies, promoting high-level policy dialogue between the Ministries of Health, Agriculture, and the Environment, and policy research to inform developed policies and frameworks. Through discussions, an OH-building Partnership (OHP) was formed, linking climate change, the environment, and human and animal health. Findings: The OHP Framework created a favorable policy environment within and between sectors, as well as between governments and international health security partners. It also promoted strategic dialogue, resource mobilization, policy advocacy, and integration of international systems with National Steering Committees to ensure accountability and emphasize national ownership. Innovative contribution to policy, practice and/or research: OHP was an effective evidence-based research-to-policy platform linking to the National One Health Strategic Plan (2021-2025). Collectively they serve as a national framework for the implementation and monitoring of OH activities. Through the adoption of policies and plans, the risk of zoonotic pathogens, environmental agent spillover, and antimicrobial resistance can be minimized through strengthening multi-sectoral OH collaboration for health security.

Keywords: one health, national policies, health security, COVID-19, Vietnam

Procedia PDF Downloads 98
1859 The Multiple Sclerosis and the Role of Human Herpesvirus 6 in Its Progression

Authors: Sina Mahdavi

Abstract:

Background and Objective: Multiple sclerosis (MS) is an inflammatory autoimmune disease of the CNS that affects the myelination process in the central nervous system (CNS). Complex interactions of various "environmental or infectious" factors may act as triggers in autoimmunity and disease progression. The association between viral infections, especially Human Herpesvirus 6 (HHV-6), and MS is one potential cause that is not well understood. In this study, we aim to summarize the available data on HHV-6 infection in MS disease progression. Materials and Methods: For this study, the keywords "Multiple sclerosis", " Human Herpesvirus 6 ", and "central nervous system" in the databases PubMed and Google Scholar between 2017 and 2022 were searched, and 12 articles were chosen, studied, and analyzed. Results: HHV 6 tends towards TCD 4+ lymphocytes and enters the CNS due to the weakening of the blood-brain barrier due to inflammatory damage. Following the observation that the HHV-6 U24 protein has a seven amino acid sequence with myelin basic protein, which is one of the main components of the myelin sheath, it could cause a molecular mimicry mechanism followed by cross-reactivity. Reactivation of HHV-6 in the CNS can cause the release of proinflammatory cytokines, including TNF-α, leading to immune-mediated demyelination in patients with MS. Conclusion: There is a high expression of endogenous retroviruses during the course of MS, which indicates the relationship between HHV-6 and MS, and that this virus can play a role in the development of MS by creating an inflammatory state. Therefore, measures to modulate the expression of HHV-6 may be effective in reducing inflammatory processes in demyelinated areas of MS patients.

Keywords: multiple sclerosis, human herpesvirus 6, central nervous system, autoimmunity

Procedia PDF Downloads 106
1858 The First Transcriptome Assembly of Marama Bean: An African Orphan Crop

Authors: Ethel E. Phiri, Lionel Hartzenberg, Percy Chimwamuromba, Emmanuel Nepolo, Jens Kossmann, James R. Lloyd

Abstract:

Orphan crops are underresearched and underutilized food plant species that have not been categorized as major food crops, but have the potential to be economically and agronomically significant. They have been documented to have the ability to tolerate extreme environmental conditions. However, limited research has been conducted to uncover their potential as food crop species. The New Partnership for Africa’s Development (NEPAD) has classified Marama bean, Tylosema esculentum, as an orphan crop. The plant is one of the 101 African orphan crops that must have their genomes sequenced, assembled, and annotated in the foreseeable future. Marama bean is a perennial leguminous plant that primarily grows in poor, arid soils in southern Africa. The plants produce large tubers that can weigh as much as 200kg. While the foliage provides fodder, the tuber is carbohydrate rich and is a staple food source for rural communities in Namibia. Also, the edible seeds are protein- and oil-rich. Marama Bean plants respond rapidly to increased temperatures and severe water scarcity without extreme consequences. Advances in molecular biology and biotechnology have made it possible to effectively transfer technologies between model- and major crops to orphan crops. In this research, the aim was to assemble the first transcriptomic analysis of Marama Bean RNA-sequence data. Many model plant species have had their genomes sequenced and their transcriptomes assembled. Therefore the availability of transcriptome data for a non-model crop plant species will allow for gene identification and comparisons between various species. The data has been sequenced using the Ilumina Hiseq 2500 sequencing platform. Data analysis is underway. In essence, this research will eventually evaluate the potential use of Marama Bean as a crop species to improve its value in agronomy. data for a non-model crop plant species will allow for gene identification and comparisons between various species. The data has been sequenced using the Ilumina Hiseq 2500 sequencing platform. Data analysis is underway. In essence, this researc will eventually evaluate the potential use of Marama bean as a crop species to improve its value in agronomy.

Keywords: 101 African orphan crops, RNA-Seq, Tylosema esculentum, underutilised crop plants

Procedia PDF Downloads 356