Search results for: rubber wood processing workers
3820 Effect of Air Temperatures (°C) and Slice Thickness (mm) on Drying Characteristics and Some Quality Properties of Omani Banana
Authors: Atheer Al-Maqbali, Mohammed Al-Rizeiqi, Pankaj Pathare
Abstract:
There is an ever-increased demand for the consumption of banana products in Oman and elsewhere in the region due to the nutritional value and the decent taste of the product. There are approximately 3,751 acres of land designated for banana cultivation in the Sultanate of Oman, which produces approximately 18,447 tons of banana product. The fresh banana product is extremely perishable, resulting in a significant post-harvest economic loss. Since the product has high sensory acceptability, the drying method is a common method for processing fresh banana products. This study aims to use the drying technology in the production of dried bananas to preserve the largest amount of natural color and delicious taste for the consumer. The study also aimed to assess the shelf stability of both water activity (aw) and color (L*, a*, b*) for fresh and finished dried bananas by using a Conventional Air Drying System. Water activity aw, color characteristic L a b, and product’s hardness were analyzed for 3mm, 5mm, and7 mm thickness at different temperaturesoC. All data were analyzed statistically using STATA 13.0, and α ≤ 0.05 was considered for the significance level. The study is useful to banana farmers to improve cultivation, food processors to optimize producer’s output and policy makers in the optimization of banana processing and post-harvest management of the products.Keywords: banana, drying, oman, quality, thickness, hardness, color
Procedia PDF Downloads 923819 Co-Composting of Poultry Manure with Different Organic Amendments
Authors: M. E. Silva, I. Brás
Abstract:
To study the influence of different organic amendments on the quality of poultry manure compost, three pilot composting trials were carried out with different mixes: poultry manure/carcasse meal/ashes/grape pomace (Pile 1), poultry manure/ cellulosic sludge (Pile 2) and poultry manure (Pile 3). For all piles, wood chips were applied as bulking agent. The process was monitored, over time, by evaluating standard physical and chemical parameters, such as, pH, electric conductivity, moisture, organic matter and ash content, total carbon and total nitrogen content, carbon/nitrogen ratio (C/N) and content in mineral elements. Piles 1 and 2 reached a thermophilic phase, however having different trends. Pile 1 reached this phase earlier than Pile 2. For both, the pH showed a slight alkaline character and the electric conductivity was lower than 2 mS/cm. Also, the initial C/N value was 22 and reached values lower than 15 at the end of composting process. The total N content of the Pile 1 increased slightly during composting, in contrast with the others piles. At the end of composting process, the phosphorus content ranged between 54 and 236 mg/kg dry matter, for Pile 2 and 3, respectively. Generally, the Piles 1 and 3 exhibited similar heavy metals content. This study showed that organic amendments can be used as carbon source, given that the final composts presented parameters within the range of those recommended in the 2nd Draft of EU regulation proposal (DG Env.A.2 2001) for compost quality.Keywords: co-composting, compost quality, organic ammendment, poultry manure
Procedia PDF Downloads 3053818 The Relationship between Quality of Work and Employment, Self-Perceived Health and Use of Health Services among the Older Japanese Workforce
Authors: Jacques Wels
Abstract:
Japan has one of the highest average retirement ages within the OCDE and is paving the way to raise the retirement age to 70. However, the Japanese labour market is facing two main issues that can have detrimental effects on health: non-standard employment forms are widespread among the ageing workforce, and poor working conditions can contribute to explain poor health in late career. To assess such a relationship, the study uses data from JSTAR. Using mediation analysis, it particularly looks at the association between job dissatisfaction, employment status, self-perceived health (SPH), and use of health care services. Results show that work quality and employment status are associated with SPH. Contract work has a particularly negative impact and therefore contributes to explain the use of health care services but is not significantly associated with lower job satisfaction levels. SPH is a good predictor of the use of health care services.Keywords: self-reported health, occupational health, employment, older workers, mediation
Procedia PDF Downloads 1203817 Segmentation of the Liver and Spleen From Abdominal CT Images Using Watershed Approach
Authors: Belgherbi Aicha, Hadjidj Ismahen, Bessaid Abdelhafid
Abstract:
The phase of segmentation is an important step in the processing and interpretation of medical images. In this paper, we focus on the segmentation of liver and spleen from the abdomen computed tomography (CT) images. The importance of our study comes from the fact that the segmentation of ROI from CT images is usually a difficult task. This difficulty is the gray’s level of which is similar to the other organ also the ROI are connected to the ribs, heart, kidneys, etc. Our proposed method is based on the anatomical information and mathematical morphology tools used in the image processing field. At first, we try to remove the surrounding and connected organs and tissues by applying morphological filters. This first step makes the extraction of interest regions easier. The second step consists of improving the quality of the image gradient. In this step, we propose a method for improving the image gradient to reduce these deficiencies by applying the spatial filters followed by the morphological filters. Thereafter we proceed to the segmentation of the liver, spleen. To validate the segmentation technique proposed, we have tested it on several images. Our segmentation approach is evaluated by comparing our results with the manual segmentation performed by an expert. The experimental results are described in the last part of this work.The system has been evaluated by computing the sensitivity and specificity between the semi-automatically segmented (liver and spleen) contour and the manually contour traced by radiological experts.Keywords: CT images, liver and spleen segmentation, anisotropic diffusion filter, morphological filters, watershed algorithm
Procedia PDF Downloads 4953816 Visual Template Detection and Compositional Automatic Regular Expression Generation for Business Invoice Extraction
Authors: Anthony Proschka, Deepak Mishra, Merlyn Ramanan, Zurab Baratashvili
Abstract:
Small and medium-sized businesses receive over 160 billion invoices every year. Since these documents exhibit many subtle differences in layout and text, extracting structured fields such as sender name, amount, and VAT rate from them automatically is an open research question. In this paper, existing work in template-based document extraction is extended, and a system is devised that is able to reliably extract all required fields for up to 70% of all documents in the data set, more than any other previously reported method. The approaches are described for 1) detecting through visual features which template a given document belongs to, 2) automatically generating extraction rules for a given new template by composing regular expressions from multiple components, and 3) computing confidence scores that indicate the accuracy of the automatic extractions. The system can generate templates with as little as one training sample and only requires the ground truth field values instead of detailed annotations such as bounding boxes that are hard to obtain. The system is deployed and used inside a commercial accounting software.Keywords: data mining, information retrieval, business, feature extraction, layout, business data processing, document handling, end-user trained information extraction, document archiving, scanned business documents, automated document processing, F1-measure, commercial accounting software
Procedia PDF Downloads 1303815 Time Series Forecasting (TSF) Using Various Deep Learning Models
Authors: Jimeng Shi, Mahek Jain, Giri Narasimhan
Abstract:
Time Series Forecasting (TSF) is used to predict the target variables at a future time point based on the learning from previous time points. To keep the problem tractable, learning methods use data from a fixed-length window in the past as an explicit input. In this paper, we study how the performance of predictive models changes as a function of different look-back window sizes and different amounts of time to predict the future. We also consider the performance of the recent attention-based Transformer models, which have had good success in the image processing and natural language processing domains. In all, we compare four different deep learning methods (RNN, LSTM, GRU, and Transformer) along with a baseline method. The dataset (hourly) we used is the Beijing Air Quality Dataset from the UCI website, which includes a multivariate time series of many factors measured on an hourly basis for a period of 5 years (2010-14). For each model, we also report on the relationship between the performance and the look-back window sizes and the number of predicted time points into the future. Our experiments suggest that Transformer models have the best performance with the lowest Mean Average Errors (MAE = 14.599, 23.273) and Root Mean Square Errors (RSME = 23.573, 38.131) for most of our single-step and multi-steps predictions. The best size for the look-back window to predict 1 hour into the future appears to be one day, while 2 or 4 days perform the best to predict 3 hours into the future.Keywords: air quality prediction, deep learning algorithms, time series forecasting, look-back window
Procedia PDF Downloads 1543814 Highly Skilled Migrants Trapped in the Brain Waste: The Eastern European Graduates in the Western European Underemployment
Authors: Katalin Bándy
Abstract:
The European emigration of highly educated immigrants draws attention to the problem of brain drain. Due to the Eastern European countries joining the EU and the opening of the Western European labour market the west-wards migration brisked up. By now another problem has been intensified correlated to migration: the migration of highly skilled workers related to brain waste tendencies. With some exceptions, educated immigrants from Eastern European countries are more likely to end up in unskilled jobs than residents. This paper is about to reveal the above-mentioned problems and this study is supported by the results of secondary pieces of research and the own survey made in the EU-15 among the Hungarian highly skilled (especially economics graduated) migrants, and it also examines the causes and in the focus there are the migrant motivations of the high-skilled young generation after the crisis.Keywords: brain drain, brain waste, migration of highly-skilled, underemployment
Procedia PDF Downloads 3413813 Flood Hazard and Risk Mapping to Assess Ice-Jam Flood Mitigation Measures
Authors: Karl-Erich Lindenschmidt, Apurba Das, Joel Trudell, Keanne Russell
Abstract:
In this presentation, we explore options for mitigating ice-jam flooding along the Athabasca River in western Canada. Not only flood hazard, expressed in this case as the probability of flood depths and extents being exceeded, but also flood risk, in which annual expected damages are calculated. Flood risk is calculated, which allows a cost-benefit analysis to be made so that decisions on the best mitigation options are not based solely on flood hazard but also on the costs related to flood damages and the benefits of mitigation. The river ice model is used to simulate extreme ice-jam flood events with which scenarios are run to determine flood exposure and damages in flood-prone areas along the river. We will concentrate on three mitigation options – the placement of a dike, artificial breakage of the ice cover along the river, the installation of an ice-control structure, and the construction of a reservoir. However, any mitigation option is not totally failsafe. For example, dikes can still be overtopped and breached, and ice jams may still occur in areas of the river where ice covers have been artificially broken up. Hence, for all options, it is recommended that zoning of building developments away from greater flood hazard areas be upheld. Flood mitigation can have a negative effect of giving inhabitants a false sense of security that flooding may not happen again, leading to zoning policies being relaxed. (Text adapted from Lindenschmidt [2022] "Ice Destabilization Study - Phase 2", submitted to the Regional Municipality of Wood Buffalo, Alberta, Canada)Keywords: ice jam, flood hazard, flood risk river ice modelling, flood risk
Procedia PDF Downloads 1853812 The Impact of Reducing Road Traffic Speed in London on Noise Levels: A Comparative Study of Field Measurement and Theoretical Calculation
Authors: Jessica Cecchinelli, Amer Ali
Abstract:
The continuing growth in road traffic and the resultant impact on the level of pollution and safety especially in urban areas have led local and national authorities to reduce traffic speed and flow in major towns and cities. Various boroughs of London have recently reduced the in-city speed limit from 30mph to 20mph mainly to calm traffic, improve safety and reduce noise and vibration. This paper reports the detailed field measurements using noise sensor and analyser and the corresponding theoretical calculations and analysis of the noise levels on a number of roads in the central London Borough of Camden where speed limit was reduced from 30mph to 20mph in all roads except the major routes of the ‘Transport for London (TfL)’. The measurements, which included the key noise levels and scales at residential streets and main roads, were conducted during weekdays and weekends normal and rush hours. The theoretical calculations were done according to the UK procedure ‘Calculation of Road Traffic Noise 1988’ and with conversion to the European L-day, L-evening, L-night, and L-den and other important levels. The current study also includes comparable data and analysis from previously measured noise in the Borough of Camden and other boroughs of central London. Classified traffic flow and speed on the roads concerned were observed and used in the calculation part of the study. Relevant data and description of the weather condition are reported. The paper also reports a field survey in the form of face-to-face interview questionnaires, which was carried out in parallel with the field measurement of noise, in order to ascertain the opinions and views of local residents and workers in the reduced speed zones of 20mph. The main findings are that the reduction in speed had reduced the noise pollution on the studied zones and that the measured and calculated noise levels for each speed zone are closely matched. Among the other findings was that of the field survey of the opinions and views of the local residents and workers in the reduced speed 20mph zones who supported the scheme and felt that it had improved the quality of life in their areas giving a sense of calmness and safety particularly for families with children, the elderly, and encouraged pedestrians and cyclists. The key conclusions are that lowering the speed limit in built-up areas would not just reduce the number of serious accidents but it would also reduce the noise pollution and promote clean modes of transport particularly walking and cycling. The details of the site observations and the corresponding calculations together with critical comparative analysis and relevant conclusions will be reported in the full version of the paper.Keywords: noise calculation, noise field measurement, road traffic noise, speed limit in london, survey of people satisfaction
Procedia PDF Downloads 4243811 Journey to Inclusive School: Description of Crucial Sensitive Concepts in the Context of Situational Analysis
Authors: Denisa Denglerova, Radim Sip
Abstract:
Academic sources as well as international agreements and national documents define inclusion in terms of several criteria: equal opportunities, fulfilling individual needs, development of human resources, community participation. In order for these criteria to be met, the community must be cohesive. Community cohesion, which is a relatively new concept, is not determined by homogeneity, but by the acceptance of diversity among the community members and utilisation of its positive potential. This brings us to a central category of inclusion - appreciating diversity and using it to a positive effect. However, school diversity is a real phenomenon, which schools need to tackle more and more often. This is also indicated by the number of publications focused on diversity in schools. These sources present recent analyses of using identity as a tool of coping with the demands of a diversified society. The aim of this study is to identify and describe in detail the processes taking place in selected schools, which contribute to their pro-inclusive character. The research is designed around a multiple case study of three pro-inclusive schools. Paradigmatically speaking, the research is rooted in situational epistemology. This is also related to the overall framework of interpretation, for which we are going to use innovative methods of situational analysis. In terms of specific research outcomes this will manifest itself in replacing the idea of “objective theory” by the idea of “detailed cartography of a social world”. The cartographic approach directs both the logic of data collection and the choice of methods of their analysis and interpretation. The research results include detection of the following sensitive concepts: Key persons. All participants can contribute to promoting an inclusion-friendly environment; however, some do so with greater motivation than others. These could include school management, teachers with a strong vision of equality, or school counsellors. They have a significant effect on the transformation of the school, and are themselves deeply convinced that inclusion is necessary. Accordingly, they select suitable co-workers; they also inspire some of the other co-workers to make changes, leading by example. Employees with strongly opposing views gradually leave the school, and new members of staff are introduced to the concept of inclusion and openness from the beginning. Manifestations of school openness in working with diversity on all important levels. By this we mean positive manipulation with diversity both in the relationships between “traditional” school participants (directors, teachers, pupils) and school-parent relationships, or relationships between schools and the broader community, in terms of teaching methods as well as ways how the school culture affects the school environment. Other important detected concepts significantly helping to form a pro-inclusive environment in the school are individual and parallel classes; freedom and responsibility of both pupils and teachers, manifested on the didactic level by tendencies towards an open curriculum; ways of asserting discipline in the school environment.Keywords: inclusion, diversity, education, sensitive concept, situational analysis
Procedia PDF Downloads 1983810 Evaluation of Cellulase and Xylanase Production by Micrococcus Sp. Isolated from Decaying Lignocellulosic Biomass Obtained from Alice Environment in the Eastern Cape of South Africa
Authors: Z. Mmango, U. Nwodo, L. V. Mabinya, A. I. Okoh
Abstract:
Cellulose and hemicellulose account for a large portion of the world‘s plant biomass. In nature, these polysaccharides are intertwined forming complex materials that requires multiple and expensive treatment processes to free up the raw materials trapped in the matrix. Enzymatic degradation remains as the preferred technique as it is inexpensive and eco-friendly. However, the insufficiencies of enzyme battery systems in the degradation of lignocellulosic complex motivate the search for effective degrading enzymes from bacterial isolates from uncommon environment. The study aimed at the evaluation of actinomycetes isolated from saw dust samples collected from wood factory under bed. Cellulase and xylanase production was screened through organism culture on carboxyl methyl cellulose agar and Birchwood xylan. Halo zone indicating lignocellose utilization was shown by an isolate identified through 16S rRNA gene as Micrococcus luteus. The optimum condition for the production of cellulase and xylanase were incubation temperature of 25 °C, fermentation medium pH 5 and 10, agitation speed of 50 and 200 (rpm) and fermentation incubation time of 96 and 84 (h) respectively. The high cellulose and xylanase activity obtained from this isolate portends industrial relevance.Keywords: carboxyl methyl cellulose, birchwood xylan, optimization, cellulase, xylanase, micrococcus, DNS method
Procedia PDF Downloads 3543809 Soil Organic Carbon and Nutrients in Smallholding Land Uses in Southern Ethiopia
Authors: Mekdes Lulu
Abstract:
This study assessed the soil organic C (SOC) and soil nutrients in smallholding home garden, woodlot, grazing land, and cropland at two soil depths and two sites in Wolaita Zone, southern Ethiopia. The results showed that soil properties were significantly influenced by land use. The home garden had significantly higher (p < 0.05) SOC and soil nutrients when compared to the cropland. When the home garden was compared to the woodlot and grazing land uses, it had significantly higher (p < 0.05) values except in SOC, total N (TN), cation exchange capacity (CEC), and exchangeable Ca. Cropland, in comparison with grazing land and woodlot, had a non-significant difference except TN. The SOC stock (0–40 cm) in the home garden, woodlot, grazing land and cropland was 79.5, 68.0, 65.0, and 58.1 Mg ha–1, respectively. Home garden significantly differed (p <0.05) in SOC only from cropland, and this was attributed not only to the relatively higher organic input in the home garden but also to the little organic matter input and frequently tillage of the cropland. The similar SOC among the home garden, woodlot and grazing lands may imply that the balance between inputs and outputs could be nearly similar for the land uses. Soil TN and CEC had a nearly similar pattern of difference as in SOC among the land uses because of their close relationship with SOC. In general, the land use influence on soil nutrients can be in the order: home garden > wood land » grazing land » cropland, with home garden showing the least difference from the woodlot and the greatest from the cropland. In the agroecosystem, in general, the influence of smallholding home garden on SOC and soil nutrient was marginally different from Eucalyptus woodlot and grazing lands but evidently different from cropland.Keywords: cropland, grazing land, home garden, soc stock, soil nutrients, woodlot
Procedia PDF Downloads 263808 A Retrospective Cohort Study on an Outbreak of Gastroenteritis Linked to a Buffet Lunch Served during a Conference in Accra
Authors: Benjamin Osei Tutu, Sharon Annison
Abstract:
On 21st November, 2016, an outbreak of foodborne illness occurred after a buffet lunch served during a stakeholders’ consultation meeting held in Accra. An investigation was conducted to characterise the affected people, determine the etiologic food, the source of contamination and the etiologic agent and to implement appropriate public health measures to prevent future occurrences. A retrospective cohort study was conducted via telephone interviews, using a structured questionnaire developed from the buffet menu. A case was defined as any person suffering from symptoms of foodborne illness e.g. diarrhoea and/or abdominal cramps after eating food served during the stakeholder consultation meeting in Accra on 21st November, 2016. The exposure status of all the members of the cohort was assessed by taking the food history of each respondent during the telephone interview. The data obtained was analysed using Epi Info 7. An environmental risk assessment was conducted to ascertain the source of the food contamination. Risks of foodborne infection from the foods eaten were determined using attack rates and odds ratios. Data was obtained from 54 people who consumed food served during the stakeholders’ meeting. Out of this population, 44 people reported with symptoms of food poisoning representing 81.45% (overall attack rate). The peak incubation period was seven hours with a minimum and maximum incubation periods of four and 17 hours, respectively. The commonly reported symptoms were diarrhoea (97.73%, 43/44), vomiting (84.09%, 37/44) and abdominal cramps (75.00%, 33/44). From the incubation period, duration of illness and the symptoms, toxin-mediated food poisoning was suspected. The environmental risk assessment of the implicated catering facility indicated a lack of time/temperature control, inadequate knowledge on food safety among workers and sanitation issues. Limited number of food samples was received for microbiological analysis. Multivariate analysis indicated that illness was significantly associated with the consumption of the snacks served (OR 14.78, P < 0.001). No stool and blood or samples of etiologic food were available for organism isolation; however, the suspected etiologic agent was Staphylococcus aureus or Clostridium perfringens. The outbreak could probably be due to the consumption of unwholesome snack (tuna sandwich or chicken. The contamination and/or growth of the etiologic agent in the snack may be due to the breakdown in cleanliness, time/temperature control and good food handling practices. Training of food handlers in basic food hygiene and safety is recommended.Keywords: Accra, buffet, conference, C. perfringens, cohort study, food poisoning, gastroenteritis, office workers, Staphylococcus aureus
Procedia PDF Downloads 2303807 Quantitative, Preservative Methodology for Review of Interview Transcripts Using Natural Language Processing
Authors: Rowan P. Martnishn
Abstract:
During the execution of a National Endowment of the Arts grant, approximately 55 interviews were collected from professionals across various fields. These interviews were used to create deliverables – historical connections for creations that began as art and evolved entirely into computing technology. With dozens of hours’ worth of transcripts to be analyzed by qualitative coders, a quantitative methodology was created to sift through the documents. The initial step was to both clean and format all the data. First, a basic spelling and grammar check was applied, as well as a Python script for normalized formatting which used an open-source grammatical formatter to make the data as coherent as possible. 10 documents were randomly selected to manually review, where words often incorrectly translated during the transcription were recorded and replaced throughout all other documents. Then, to remove all banter and side comments, the transcripts were spliced into paragraphs (separated by change in speaker) and all paragraphs with less than 300 characters were removed. Secondly, a keyword extractor, a form of natural language processing where significant words in a document are selected, was run on each paragraph for all interviews. Every proper noun was put into a data structure corresponding to that respective interview. From there, a Bidirectional and Auto-Regressive Transformer (B.A.R.T.) summary model was then applied to each paragraph that included any of the proper nouns selected from the interview. At this stage the information to review had been sent from about 60 hours’ worth of data to 20. The data was further processed through light, manual observation – any summaries which proved to fit the criteria of the proposed deliverable were selected, as well their locations within the document. This narrowed that data down to about 5 hours’ worth of processing. The qualitative researchers were then able to find 8 more connections in addition to our previous 4, exceeding our minimum quota of 3 to satisfy the grant. Major findings of the study and subsequent curation of this methodology raised a conceptual finding crucial to working with qualitative data of this magnitude. In the use of artificial intelligence there is a general trade off in a model between breadth of knowledge and specificity. If the model has too much knowledge, the user risks leaving out important data (too general). If the tool is too specific, it has not seen enough data to be useful. Thus, this methodology proposes a solution to this tradeoff. The data is never altered outside of grammatical and spelling checks. Instead, the important information is marked, creating an indicator of where the significant data is without compromising the purity of it. Secondly, the data is chunked into smaller paragraphs, giving specificity, and then cross-referenced with the keywords (allowing generalization over the whole document). This way, no data is harmed, and qualitative experts can go over the raw data instead of using highly manipulated results. Given the success in deliverable creation as well as the circumvention of this tradeoff, this methodology should stand as a model for synthesizing qualitative data while maintaining its original form.Keywords: B.A.R.T.model, keyword extractor, natural language processing, qualitative coding
Procedia PDF Downloads 293806 Automatic Fluid-Structure Interaction Modeling and Analysis of Butterfly Valve Using Python Script
Authors: N. Guru Prasath, Sangjin Ma, Chang-Wan Kim
Abstract:
A butterfly valve is a quarter turn valve which is used to control the flow of a fluid through a section of pipe. Generally, butterfly valve is used in wide range of applications such as water distribution, sewage, oil and gas plants. In particular, butterfly valve with larger diameter finds its immense applications in hydro power plants to control the fluid flow. In-lieu with the constraints in cost and size to run laboratory setup, analysis of large diameter values will be mostly studied by computational method which is the best and inexpensive solution. For fluid and structural analysis, CFD and FEM software is used to perform large scale valve analyses, respectively. In order to perform above analysis in butterfly valve, the CAD model has to recreate and perform mesh in conventional software’s for various dimensions of valve. Therefore, its limitation is time consuming process. In-order to overcome that issue, python code was created to outcome complete pre-processing setup automatically in Salome software. Applying dimensions of the model clearly in the python code makes the running time comparatively lower and easier way to perform analysis of the valve. Hence, in this paper, an attempt was made to study the fluid-structure interaction (FSI) of butterfly valves by varying the valve angles and dimensions using python code in pre-processing software, and results are produced.Keywords: butterfly valve, flow coefficient, automatic CFD analysis, FSI analysis
Procedia PDF Downloads 2413805 Music Reading Expertise Facilitates Implicit Statistical Learning of Sentence Structures in a Novel Language: Evidence from Eye Movement Behavior
Authors: Sara T. K. Li, Belinda H. J. Chung, Jeffery C. N. Yip, Janet H. Hsiao
Abstract:
Music notation and text reading both involve statistical learning of music or linguistic structures. However, it remains unclear how music reading expertise influences text reading behavior. The present study examined this issue through an eye-tracking study. Chinese-English bilingual musicians and non-musicians read English sentences, Chinese sentences, musical phrases, and sentences in Tibetan, a language novel to the participants, with their eye movement recorded. Each set of stimuli consisted of two conditions in terms of structural regularity: syntactically correct and syntactically incorrect musical phrases/sentences. They then completed a sentence comprehension (for syntactically correct sentences) or a musical segment/word recognition task afterwards to test their comprehension/recognition abilities. The results showed that in reading musical phrases, as compared with non-musicians, musicians had a higher accuracy in the recognition task, and had shorter reading time, fewer fixations, and shorter fixation duration when reading syntactically correct (i.e., in diatonic key) than incorrect (i.e., in non-diatonic key/atonal) musical phrases. This result reflects their expertise in music reading. Interestingly, in reading Tibetan sentences, which was novel to both participant groups, while non-musicians did not show any behavior differences between reading syntactically correct or incorrect Tibetan sentences, musicians showed a shorter reading time and had marginally fewer fixations when reading syntactically correct sentences than syntactically incorrect ones. However, none of the musicians reported discovering any structural regularities in the Tibetan stimuli after the experiment when being asked explicitly, suggesting that they may have implicitly acquired the structural regularities in Tibetan sentences. This group difference was not observed when they read English or Chinese sentences. This result suggests that music reading expertise facilities reading texts in a novel language (i.e., Tibetan), but not in languages that the readers are already familiar with (i.e., English and Chinese). This phenomenon may be due to the similarities between reading music notations and reading texts in a novel language, as in both cases the stimuli follow particular statistical structures but do not involve semantic or lexical processing. Thus, musicians may transfer their statistical learning skills stemmed from music notation reading experience to implicitly discover structures of sentences in a novel language. This speculation is consistent with a recent finding showing that music reading expertise modulates the processing of English nonwords (i.e., words that do not follow morphological or orthographic rules) but not pseudo- or real words. These results suggest that the modulation of music reading expertise on language processing depends on the similarities in the cognitive processes involved. It also has important implications for the benefits of music education on language and cognitive development.Keywords: eye movement behavior, eye-tracking, music reading expertise, sentence reading, structural regularity, visual processing
Procedia PDF Downloads 3803804 Unstructured-Data Content Search Based on Optimized EEG Signal Processing and Multi-Objective Feature Extraction
Authors: Qais M. Yousef, Yasmeen A. Alshaer
Abstract:
Over the last few years, the amount of data available on the globe has been increased rapidly. This came up with the emergence of recent concepts, such as the big data and the Internet of Things, which have furnished a suitable solution for the availability of data all over the world. However, managing this massive amount of data remains a challenge due to their large verity of types and distribution. Therefore, locating the required file particularly from the first trial turned to be a not easy task, due to the large similarities of names for different files distributed on the web. Consequently, the accuracy and speed of search have been negatively affected. This work presents a method using Electroencephalography signals to locate the files based on their contents. Giving the concept of natural mind waves processing, this work analyses the mind wave signals of different people, analyzing them and extracting their most appropriate features using multi-objective metaheuristic algorithm, and then classifying them using artificial neural network to distinguish among files with similar names. The aim of this work is to provide the ability to find the files based on their contents using human thoughts only. Implementing this approach and testing it on real people proved its ability to find the desired files accurately within noticeably shorter time and retrieve them as a first choice for the user.Keywords: artificial intelligence, data contents search, human active memory, mind wave, multi-objective optimization
Procedia PDF Downloads 1753803 Automatic Early Breast Cancer Segmentation Enhancement by Image Analysis and Hough Transform
Authors: David Jurado, Carlos Ávila
Abstract:
Detection of early signs of breast cancer development is crucial to quickly diagnose the disease and to define adequate treatment to increase the survival probability of the patient. Computer Aided Detection systems (CADs), along with modern data techniques such as Machine Learning (ML) and Neural Networks (NN), have shown an overall improvement in digital mammography cancer diagnosis, reducing the false positive and false negative rates becoming important tools for the diagnostic evaluations performed by specialized radiologists. However, ML and NN-based algorithms rely on datasets that might bring issues to the segmentation tasks. In the present work, an automatic segmentation and detection algorithm is described. This algorithm uses image processing techniques along with the Hough transform to automatically identify microcalcifications that are highly correlated with breast cancer development in the early stages. Along with image processing, automatic segmentation of high-contrast objects is done using edge extraction and circle Hough transform. This provides the geometrical features needed for an automatic mask design which extracts statistical features of the regions of interest. The results shown in this study prove the potential of this tool for further diagnostics and classification of mammographic images due to the low sensitivity to noisy images and low contrast mammographies.Keywords: breast cancer, segmentation, X-ray imaging, hough transform, image analysis
Procedia PDF Downloads 833802 Technology Impact on the Challenge between Human Rights and Cyber Terrorism
Authors: Abanoub Zare Zakaria Herzalla
Abstract:
The link between terrorism and human rights has become a major challenge in the fight against terrorism around the world. This is based on the fact that terrorism and human rights are so closely linked that when the former starts, the latter are violated. This direct connection was recognized in the Vienna Declaration and Program of Action adopted by the World Conference on Human Rights in Vienna on June 25, 1993, which recognizes that acts of terrorism in all their forms and manifestations aim to destroy the human rights of people. Terrorism therefore represents an attack on our most basic human rights. To this end, the first part of this article focuses on the connections between terrorism and human rights and seeks to highlight the interdependence between these two concepts. The second part discusses the emerging concept of cyberterrorism and its manifestations. An analysis of the fight against cyberterrorism in the context of human rights is also carried out.Keywords: sustainable development, human rights, the right to development, the human rights-based approach to development, environmental rights, economic development, social sustainability human rights protection, human rights violations, workers’ rights, justice, security.
Procedia PDF Downloads 473801 Enhancing the Performance of Automatic Logistic Centers by Optimizing the Assignment of Material Flows to Workstations and Flow Racks
Authors: Sharon Hovav, Ilya Levner, Oren Nahum, Istvan Szabo
Abstract:
In modern large-scale logistic centers (e.g., big automated warehouses), complex logistic operations performed by human staff (pickers) need to be coordinated with the operations of automated facilities (robots, conveyors, cranes, lifts, flow racks, etc.). The efficiency of advanced logistic centers strongly depends on optimizing picking technologies in synch with the facility/product layout, as well as on optimal distribution of material flows (products) in the system. The challenge is to develop a mathematical operations research (OR) tool that will optimize system cost-effectiveness. In this work, we propose a model that describes an automatic logistic center consisting of a set of workstations located at several galleries (floors), with each station containing a known number of flow racks. The requirements of each product and the working capacity of stations served by a given set of workers (pickers) are assumed as predetermined. The goal of the model is to maximize system efficiency. The proposed model includes two echelons. The first is the setting of the (optimal) number of workstations needed to create the total processing/logistic system, subject to picker capacities. The second echelon deals with the assignment of the products to the workstations and flow racks, aimed to achieve maximal throughputs of picked products over the entire system given picker capacities and budget constraints. The solutions to the problems at the two echelons interact to balance the overall load in the flow racks and maximize overall efficiency. We have developed an operations research model within each echelon. In the first echelon, the problem of calculating the optimal number of workstations is formulated as a non-standard bin-packing problem with capacity constraints for each bin. The problem arising in the second echelon is presented as a constrained product-workstation-flow rack assignment problem with non-standard mini-max criteria in which the workload maximum is calculated across all workstations in the center and the exterior minimum is calculated across all possible product-workstation-flow rack assignments. The OR problems arising in each echelon are proved to be NP-hard. Consequently, we find and develop heuristic and approximation solution algorithms based on exploiting and improving local optimums. The LC model considered in this work is highly dynamic and is recalculated periodically based on updated demand forecasts that reflect market trends, technological changes, seasonality, and the introduction of new items. The suggested two-echelon approach and the min-max balancing scheme are shown to work effectively on illustrative examples and real-life logistic data.Keywords: logistics center, product-workstation, assignment, maximum performance, load balancing, fast algorithm
Procedia PDF Downloads 2283800 Future-Proofing the Workforce: A Case Study of Integrated Human Capability Frameworks to Support Business Success
Authors: Penelope Paliadelis, Asheley Jones, Glenn Campbell
Abstract:
This paper discusses the development of co-designed capability frameworks for two large multinational organizations led by a university department. The aim was to create evidence-based, integrated capability frameworks that could define, identify, and measure human skill capabilities independent of specific work roles. The frameworks capture and cluster human skills required in the workplace and capture their application at various levels of mastery. Identified capability gaps inform targeted learning opportunities for workers to enhance their employability skills. The paper highlights the value of this evidence-based framework development process in capturing, defining, and assessing desired human-focused capabilities for organizational growth and success.Keywords: capability framework, human skills, work-integrated learning, credentialing, digital badging
Procedia PDF Downloads 783799 Contribution of Remote Sensing and GIS to the Study of the Impact of the Salinity of Sebkhas on the Quality of Groundwater: Case of Sebkhet Halk El Menjel (Sousse)
Authors: Gannouni Sonia, Hammami Asma, Saidi Salwa, Rebai Noamen
Abstract:
Water resources in Tunisia have experienced quantitative and qualitative degradation, especially when talking about wetlands and Sbekhas. Indeed, the objective of this work is to study the spatio-temporal evolution of salinity for 29 years (from 1987 to 2016). A study of the connection between surface water and groundwater is necessary to know the degree of influence of the Sebkha brines on the water table. The evolution of surface salinity is determined by remote sensing based on Landsat TM and OLI/TIRS satellite images of the years 1987, 2007, 2010, and 2016. The processing of these images allowed us to determine the NDVI(Normalized Difference Vegetation Index), the salinity index, and the surface temperature around Sebkha. In addition, through a geographic information system(GIS), we could establish a map of the distribution of salinity in the subsurface of the water table of Chott Mariem and Hergla/SidiBouAli/Kondar. The results of image processing and the calculation of the index and surface temperature show an increase in salinity downstream of in addition to the sebkha and the development of vegetation cover upstream and the western part of the sebkha. This richness may be due both to contamination by seawater infiltration from the barrier beach of Hergla as well as the passage of groundwater to the sebkha.Keywords: spatio-temporal monitoring, salinity, satellite images, NDVI, sebkha
Procedia PDF Downloads 1333798 Recovery of the Demolition and Construction Waste, Casablanca (Morocco)
Authors: Morsli Mourad, Tahiri Mohamed, Samdi Azzeddine
Abstract:
Casablanca is the biggest city in Morocco. It concentrates more than 60% of the economic and industrial activity of the kingdom. Its building and public works (BTP) sector is the leading source of inert waste scattered in open areas. This inert waste is a major challenge for the city of Casablanca, as it is not properly managed, thus causing a significant nuisance for the environment and the health of the population. Hence the vision of our project is to recycle and valorize concrete waste. In this work, we present concrete results in the exploitation of this abundant and permanent deposit. Typical wastes are concrete, clay and concrete bricks, ceramic tiles, marble panels, gypsum, scrap metal, wood . The work performed included: geolocation with a combination of artificial intelligence and Google Earth, estimation of the amount of waste per site, sorting, crushing, grinding, and physicochemical characterization of the samples. Then, we proceeded to the exploitation of the types of substrates to be developed: light cement, coating, and glue for ceramics... The said products were tested and characterized by X-ray fluorescence, specific surface, resistance to bending and crushing, etc. We will present in detail the main results of our research work and also describe the specific properties of each material developed.Keywords: déchets de démolition et des chantiers de construction, logiciels de combinaison SIG, valorisation de déchets inertes, enduits, ciment leger, casablanca
Procedia PDF Downloads 1123797 Alphabet Recognition Using Pixel Probability Distribution
Authors: Vaidehi Murarka, Sneha Mehta, Dishant Upadhyay
Abstract:
Our project topic is “Alphabet Recognition using pixel probability distribution”. The project uses techniques of Image Processing and Machine Learning in Computer Vision. Alphabet recognition is the mechanical or electronic translation of scanned images of handwritten, typewritten or printed text into machine-encoded text. It is widely used to convert books and documents into electronic files etc. Alphabet Recognition based OCR application is sometimes used in signature recognition which is used in bank and other high security buildings. One of the popular mobile applications includes reading a visiting card and directly storing it to the contacts. OCR's are known to be used in radar systems for reading speeders license plates and lots of other things. The implementation of our project has been done using Visual Studio and Open CV (Open Source Computer Vision). Our algorithm is based on Neural Networks (machine learning). The project was implemented in three modules: (1) Training: This module aims “Database Generation”. Database was generated using two methods: (a) Run-time generation included database generation at compilation time using inbuilt fonts of OpenCV library. Human intervention is not necessary for generating this database. (b) Contour–detection: ‘jpeg’ template containing different fonts of an alphabet is converted to the weighted matrix using specialized functions (contour detection and blob detection) of OpenCV. The main advantage of this type of database generation is that the algorithm becomes self-learning and the final database requires little memory to be stored (119kb precisely). (2) Preprocessing: Input image is pre-processed using image processing concepts such as adaptive thresholding, binarizing, dilating etc. and is made ready for segmentation. “Segmentation” includes extraction of lines, words, and letters from the processed text image. (3) Testing and prediction: The extracted letters are classified and predicted using the neural networks algorithm. The algorithm recognizes an alphabet based on certain mathematical parameters calculated using the database and weight matrix of the segmented image.Keywords: contour-detection, neural networks, pre-processing, recognition coefficient, runtime-template generation, segmentation, weight matrix
Procedia PDF Downloads 3893796 Fault Detection and Diagnosis of Broken Bar Problem in Induction Motors Base Wavelet Analysis and EMD Method: Case Study of Mobarakeh Steel Company in Iran
Authors: M. Ahmadi, M. Kafil, H. Ebrahimi
Abstract:
Nowadays, induction motors have a significant role in industries. Condition monitoring (CM) of this equipment has gained a remarkable importance during recent years due to huge production losses, substantial imposed costs and increases in vulnerability, risk, and uncertainty levels. Motor current signature analysis (MCSA) is one of the most important techniques in CM. This method can be used for rotor broken bars detection. Signal processing methods such as Fast Fourier transformation (FFT), Wavelet transformation and Empirical Mode Decomposition (EMD) are used for analyzing MCSA output data. In this study, these signal processing methods are used for broken bar problem detection of Mobarakeh steel company induction motors. Based on wavelet transformation method, an index for fault detection, CF, is introduced which is the variation of maximum to the mean of wavelet transformation coefficients. We find that, in the broken bar condition, the amount of CF factor is greater than the healthy condition. Based on EMD method, the energy of intrinsic mode functions (IMF) is calculated and finds that when motor bars become broken the energy of IMFs increases.Keywords: broken bar, condition monitoring, diagnostics, empirical mode decomposition, fourier transform, wavelet transform
Procedia PDF Downloads 1503795 Human Rights Impact on Citizens Evolution
Authors: Joseph Marzouk Gerais Abdelmalak
Abstract:
The interface between development and human rights has long been the subject of academic debate. Therefore, to understand the dynamics between the two concepts, a number of principles have been adopted, ranging from the right to development to a human rights-based approach to development. Despite these attempts, the exact connection between development and human rights is not yet fully understood. However, the inherent interdependence between these two concepts and the idea that development efforts should be undertaken with respect for human rights guarantees have gained momentum in recent years. It will then be examined whether the right to sustainable development is recognized.The article therefore concludes that the principles of sustainable development are recognized, directly or indirectly, in various human rights instruments, which represents a positive answer to the question posed above. Therefore, this work discusses international and regional human rights instruments as well as case law and interpretative guidelines from human rights bodies to demonstrate this hypothesis.Keywords: sustainable development, human rights, the right to development, the human rights-based approach to development, environmental rights, economic development, social sustainability human rights protection, human rights violations, workers’ rights, justice, security
Procedia PDF Downloads 733794 Modernization of Garri-Frying Technologies with Respect to Women Anthromophic Quality in Nigeria
Authors: Adegbite Bashiru Adeniyi, Olaniyi Akeem Olawale, Ayobamidele Sinatu Juliet
Abstract:
The study was carried out in the 6 South Western states of Nigeria to analyze socio-economic characteristic of garri processors and their anthropometric qualities with respect to modern technologies used in garri processing. About 20 respondents were randomly selected from each of the 6 workstations purposively considered for the study due to their daily processing activities already attracted high patronage of customers. These include Oguntolu village (Ogun State), Igoba-Akure (Ondo State), Imo-Ilesa (Osun State), Odo Oba-Ileri (Oyo State), Irasa village (Ekiti State) and Epe in Lagos state. Interview schedule was conducted for 120 respondents to elicit information. Data were analyzed using descriptive statistical tools. It was observed from the findings that respondents were in their most productive age range (36-45 years) except Ogun state where majority (45%) were relatively older than 45 years. A fewer processors were much younger than 26 years old. It furthers revealed that not less than 55% have body weight greater than 50.0 kilogram, also not less than 70% were taller than 1.5 meter. So also, the hand length and hand thickness of the majority were long and bulky which are considered suitable for operating some modern and improved technologies in garri-frying process. This information could be used by various technological developers to enhance production of modern equipment and tools for a greater efficiency.Keywords: agro-business, anthromorphic, modernization, proficiency
Procedia PDF Downloads 5123793 Instant Location Detection of Objects Moving at High Speed in C-OTDR Monitoring Systems
Authors: Andrey V. Timofeev
Abstract:
The practical efficient approach is suggested to estimate the high-speed objects instant bounds in C-OTDR monitoring systems. In case of super-dynamic objects (trains, cars) is difficult to obtain the adequate estimate of the instantaneous object localization because of estimation lag. In other words, reliable estimation coordinates of monitored object requires taking some time for data observation collection by means of C-OTDR system, and only if the required sample volume will be collected the final decision could be issued. But it is contrary to requirements of many real applications. For example, in rail traffic management systems we need to get data off the dynamic objects localization in real time. The way to solve this problem is to use the set of statistical independent parameters of C-OTDR signals for obtaining the most reliable solution in real time. The parameters of this type we can call as 'signaling parameters' (SP). There are several the SP’s which carry information about dynamic objects instant localization for each of C-OTDR channels. The problem is that some of these parameters are very sensitive to dynamics of seismoacoustic emission sources but are non-stable. On the other hand, in case the SP is very stable it becomes insensitive as a rule. This report contains describing the method for SP’s co-processing which is designed to get the most effective dynamic objects localization estimates in the C-OTDR monitoring system framework.Keywords: C-OTDR-system, co-processing of signaling parameters, high-speed objects localization, multichannel monitoring systems
Procedia PDF Downloads 4703792 Drivers of Farmers' Contract Compliance Behaviour: Evidence from a Case Study of Dangote Tomato Processing Plant in Northern Nigeria.
Authors: Umar Shehu Umar
Abstract:
Contract farming is a viable strategy agribusinesses rely on to strengthen vertical coordination. However, low contract compliance remains a significant setback to agribusinesses' contract performance. The present study aims to understand what drives smallholder farmers’ contract compliance behaviour. Qualitative information was collected through Focus Group Discussions to enrich the design of the survey questionnaire administered on a sample of 300 randomly selected farmers contracted by the Dangote Tomato Processing Plant (DTPP) in four regions of northern Nigeria. Novel transaction level data of tomato sales covering one season were collected in addition to socio-economic information of the sampled farmers. Binary logistic model results revealed that open fresh market tomato prices and payment delays negatively affect farmers' compliance behaviour while quantity harvested, education level and input provision correlated positively with compliance. The study suggests that contract compliance will increase if contracting firms devise a reliable and timely payment plan (e.g., digital payment), continue input and service provisions (e.g., improved seeds, extension services) and incentives (e.g., loyalty rewards, bonuses) in the contract.Keywords: contract farming, compliance, farmers and processors., smallholder
Procedia PDF Downloads 563791 Study on the Suppression of Hydrogen Generation by Aluminum-Containing Waste Incineration Ash and Water
Authors: Hideyuki Onodera, Ryoji Imai, Masahiro Sakai
Abstract:
Explosions have occurred in incineration plants in conveyors, ash pits, and other locations. The cause of such explosions is thought to be the reaction of metallic aluminum contained in the ash with water used to cool the ash and prevent scattering, resulting in the generation of hydrogen. Given this background, conveyors and other equipment have been damaged by explosions, which has hindered the stable operation of incineration plants. In addition, workers may be injured by equipment explosions, creating an unsafe situation. To remedy these problems, it is necessary to devise a way to prevent hydrogen explosions from occurring. To overcome this problem, we conducted a hydrogen generation reaction experiment using simulated incinerator ash powder containing aluminum, calcium oxide, and water and confirmed that conditions exist to stop the hydrogen generation reaction. The results of this research may contribute to the suppression of hydrogen explosions at incineration plants.Keywords: waste incinerated ash, aluminum, water, hydrogen, suppression of hydrogen generation, incineration plant
Procedia PDF Downloads 28