Search results for: digital forensics process and procedures
1021 Study of Lanthanoide Organic Frameworks Properties and Synthesis: Multicomponent Ligands
Authors: Ayla Roberta Galaco, Juliana Fonseca De Lima, Osvaldo Antonio Serra
Abstract:
Coordination polymers, also known as metal-organic frameworks (MOFs) or lanthanoide organic frameworks (LOFs) have been reported due of their promising applications in gas storage, separation, catalysis, luminescence, magnetism, drug delivery, and so on. As a type of organic–inorganic hybrid materials, the properties of coordination polymers could be chosen by deliberately selecting the organic and inorganic components. LOFs have received considerable attention because of their properties such as porosity, luminescence, and magnetism. Methods such as solvothermal synthesis are important as a strategy to control the structural and morphological properties as well as the composition of the target compounds. In this work the first solvothermal synthesis was employed to obtain the compound [Y0.4,Yb0.4,Er0.2(dmf)(for)(H2O)(tft)], by using terephthalic acid (tft) and oxalic acid, decomposed in formate (for), as ligands; Yttrium, Ytterbium and, Erbium as metal centers, in DMF and water for 4 days under 160 °C. The semi-rigid terephthalic acid (dicarboxylic) coordinates with Ln3+ ions and also is possible to form a polyfunctional bridge. On the other hand, oxalate anion has no high-energy vibrational groups, which benefits the excitation of Yb3+ in upconversion process. It was observed that the compounds with water molecules in the coordination sphere of the lanthanoide ions cause lower crystalline properties and change the structure of the LOF (1D, 2D, 3D). In the FTIR, the bands at 1589 and 1500 cm-1 correspond to the asymmetric stretching vibration of –COO. The band at 1383 cm-1 is assigned to the symmetric stretching vibration of –COO. Single crystal X-ray diffraction study reveals an infinite 3D coordination framework that crystalizes in space group P21/c. The other three products, [TR(chel)(ofd)0,5(H2O)2], where TR= Eu3+, Y3, and Yb3+/Er3+ were obtained by using 1, 2-phenylenedioxydiacetic acid (ofd) and chelidonic acid (chel) as organic ligands. Thermal analysis shows that the lanthanoide organic frameworks do not collapse at temperatures below 250 °C. By the polycrystalline X-ray diffraction patterns (PXRD) it was observed that the compounds with Eu3+, Y3+, and Yb3+/Er3+ ions are isostructural. From PXRD patterns, high crystallinity can be noticed for the complexes. The final products were characterized by single X-ray diffraction (XRD), Fourier transform infrared spectroscopy (FTIR), energy dispersive spectroscopy (EDS) and thermogravimetric analysis (TGA). The X-ray diffraction (XRD) is an effective method to investigate crystalline properties of synthesized materials. The solid crystal obtained in the synthesis show peaks at 2θ < 10°, indicating the MOF formation. The chemical composition of LOFs was also confirmed by EDS.Keywords: isostructural, lanthanoids, lanthanoids organic frameworks (LOFs), metal organic frameworks (MOFs), thermogravimetry, X-Ray diffraction
Procedia PDF Downloads 2601020 [Keynote Talk]: Production Flow Coordination on Supply Chains: Brazilian Case Studies
Authors: Maico R. Severino, Laura G. Caixeta, Nadine M. Costa, Raísa L. T. Napoleão, Éverton F. V. Valle, Diego D. Calixto, Danielle Oliveira
Abstract:
One of the biggest barriers that companies find nowadays is the coordination of production flow in their Supply Chains (SC). In this study, coordination is understood as a mechanism for incorporating the entire production channel, with everyone involved focused on achieving the same goals. Sometimes, this coordination is attempted by the use of logistics practices or production plan and control methods. No papers were found in the literature that presented the combined use of logistics practices and production plan and control methods. The main objective of this paper is to propose solutions for six case studies combining logistics practices and Ordering Systems (OS). The methodology used in this study was a conceptual model of decision making. This model contains six phases: a) the analysis the types and characteristics of relationships in the SC; b) the choice of the OS; c) the choice of the logistics practices; d) the development of alternative proposals of combined use; e) the analysis of the consistency of the chosen alternative; f) the qualitative and quantitative assessment of the impact on the coordination of the production flow and the verification of applicability of the proposal in the real case. This study was conducted on six Brazilian SC of different sectors: footwear, food and beverages, garment, sugarcane, mineral and metal mechanical. The results from this study showed that there was improvement in the coordination of the production flow through the following proposals: a) for the footwear industry the use of Period Bath Control (PBC), Quick Response (QR) and Enterprise Resource Planning (ERP); b) for the food and beverage sector firstly the use of Electronic Data Interchange (EDI), ERP, Continuous Replenishment (CR) and Drum-Buffer-Rope Order (DBR) (for situations in which the plants of both companies are distant), and secondly EDI, ERP, Milk-Run and Review System Continues (for situations in which the plants of both companies are close); c) for the garment industry the use of Collaborative Planning, Forecasting, and Replenishment (CPFR) and Constant Work-In-Process (CONWIP) System; d) for the sugarcane sector the use of EDI, ERP and CONWIP System; e) for the mineral processes industry the use of Vendor Managed Inventory (VMI), EDI and MaxMin Control System; f) for the metal mechanical sector the use of CONWIP System and Continuous Replenishment (CR). It should be emphasized that the proposals are exclusively recommended for the relationship between client and supplier studied. Therefore, it cannot be generalized to other cases. However, what can be generalized is the methodology used to choose the best practices for each case. Based on the study, it can be concluded that the combined use of OS and logistics practices enable a better coordination of flow production on SC.Keywords: supply chain management, production flow coordination, logistics practices, ordering systems
Procedia PDF Downloads 2081019 A Mixed Integer Linear Programming Model for Container Collection
Authors: J. Van Engeland, C. Lavigne, S. De Jaeger
Abstract:
In the light of the transition towards a more circular economy, recovery of products, parts or materials will gain in importance. Additionally, the EU proximity principle related to waste management and emissions generated by transporting large amounts of end-of-life products, shift attention to local recovery networks. The Flemish inter-communal cooperation for municipal solid waste management Meetjesland (IVM) is currently investigating the set-up of such a network. More specifically, the network encompasses the recycling of polyvinyl chloride (PVC), which is collected in separate containers. When these containers are full, a truck should transport them to the processor which can recycle the PVC into new products. This paper proposes a model to optimize the container collection. The containers are located at different Civic Amenity sites (CA sites) in a certain region. Since people can drop off their waste at these CA sites, the containers will gradually fill up during a planning horizon. If a certain container is full, it has to be collected and replaced by an empty container. The collected waste is then transported to a single processor. To perform this collection and transportation of containers, the responsible firm has a set of vehicles stationed at a single depot and different personnel crews. A vehicle can load exactly one container. If a trailer is attached to the vehicle, it can load an additional container. Each day of the planning horizon, the different crews and vehicles leave the depot to collect containers at the different sites. After loading one or two containers, the crew has to drive to the processor for unloading the waste and to pick up empty containers. Afterwards, the crew can again visit sites or it can return to the depot to end its collection work for that day. All along the collection process, the crew has to respect the opening hours of the sites. In order to allow for some flexibility, a crew is allowed to wait a certain amount of time at the gate of a site until it opens. The problem described can be modelled as a variant to the PVRP-TW (Periodic Vehicle Routing Problem with Time Windows). However, a vehicle can at maximum load two containers, hence only two subsequent site visits are possible. For that reason, we will refer to the model as a model for building tactical waste collection schemes. The goal is to a find a schedule describing which crew should visit which CA site on which day to minimize the number of trucks and the routing costs. The model was coded in IBM CPLEX Optimization studio and applied to a number of test instances. Good results were obtained, and specific suggestions concerning route and truck costs could be made. For a large range of input parameters, collection schemes using two trucks are obtained.Keywords: container collection, crew scheduling, mixed integer linear programming, waste management
Procedia PDF Downloads 1341018 Definition of Aerodynamic Coefficients for Microgravity Unmanned Aerial System
Authors: Gamaliel Salazar, Adriana Chazaro, Oscar Madrigal
Abstract:
The evolution of Unmanned Aerial Systems (UAS) has made it possible to develop new vehicles capable to perform microgravity experiments which due its cost and complexity were beyond the reach for many institutions. In this study, the aerodynamic behavior of an UAS is studied through its deceleration stage after an initial free fall phase (where the microgravity effect is generated) using Computational Fluid Dynamics (CFD). Due to the fact that the payload would be analyzed under a microgravity environment and the nature of the payload itself, the speed of the UAS must be reduced in a smoothly way. Moreover, the terminal speed of the vehicle should be low enough to preserve the integrity of the payload and vehicle during the landing stage. The UAS model is made by a study pod, control surfaces with fixed and mobile sections, landing gear and two semicircular wing sections. The speed of the vehicle is decreased by increasing the angle of attack (AoA) of each wing section from 2° (where the airfoil S1091 has its greatest aerodynamic efficiency) to 80°, creating a circular wing geometry. Drag coefficients (Cd) and forces (Fd) are obtained employing CFD analysis. A simplified 3D model of the vehicle is analyzed using Ansys Workbench 16. The distance between the object of study and the walls of the control volume is eight times the length of the vehicle. The domain is discretized using an unstructured mesh based on tetrahedral elements. The refinement of the mesh is made by defining an element size of 0.004 m in the wing and control surfaces in order to figure out the fluid behavior in the most important zones, as well as accurate approximations of the Cd. The turbulent model k-epsilon is selected to solve the governing equations of the fluids while a couple of monitors are placed in both wing and all-body vehicle to visualize the variation of the coefficients along the simulation process. Employing a statistical approximation response surface methodology the case of study is parametrized considering the AoA of the wing as the input parameter and Cd and Fd as output parameters. Based on a Central Composite Design (CCD), the Design Points (DP) are generated so the Cd and Fd for each DP could be estimated. Applying a 2nd degree polynomial approximation the drag coefficients for every AoA were determined. Using this values, the terminal speed at each position is calculated considering a specific Cd. Additionally, the distance required to reach the terminal velocity at each AoA is calculated, so the minimum distance for the entire deceleration stage without comprising the payload could be determine. The Cd max of the vehicle is 1.18, so its maximum drag will be almost like the drag generated by a parachute. This guarantees that aerodynamically the vehicle can be braked, so it could be utilized for several missions allowing repeatability of microgravity experiments.Keywords: microgravity effect, response surface, terminal speed, unmanned system
Procedia PDF Downloads 1731017 Machine Learning Techniques for Estimating Ground Motion Parameters
Authors: Farid Khosravikia, Patricia Clayton
Abstract:
The main objective of this study is to evaluate the advantages and disadvantages of various machine learning techniques in forecasting ground-motion intensity measures given source characteristics, source-to-site distance, and local site condition. Intensity measures such as peak ground acceleration and velocity (PGA and PGV, respectively) as well as 5% damped elastic pseudospectral accelerations at different periods (PSA), are indicators of the strength of shaking at the ground surface. Estimating these variables for future earthquake events is a key step in seismic hazard assessment and potentially subsequent risk assessment of different types of structures. Typically, linear regression-based models, with pre-defined equations and coefficients, are used in ground motion prediction. However, due to the restrictions of the linear regression methods, such models may not capture more complex nonlinear behaviors that exist in the data. Thus, this study comparatively investigates potential benefits from employing other machine learning techniques as a statistical method in ground motion prediction such as Artificial Neural Network, Random Forest, and Support Vector Machine. The algorithms are adjusted to quantify event-to-event and site-to-site variability of the ground motions by implementing them as random effects in the proposed models to reduce the aleatory uncertainty. All the algorithms are trained using a selected database of 4,528 ground-motions, including 376 seismic events with magnitude 3 to 5.8, recorded over the hypocentral distance range of 4 to 500 km in Oklahoma, Kansas, and Texas since 2005. The main reason of the considered database stems from the recent increase in the seismicity rate of these states attributed to petroleum production and wastewater disposal activities, which necessities further investigation in the ground motion models developed for these states. Accuracy of the models in predicting intensity measures, generalization capability of the models for future data, as well as usability of the models are discussed in the evaluation process. The results indicate the algorithms satisfy some physically sound characteristics such as magnitude scaling distance dependency without requiring pre-defined equations or coefficients. Moreover, it is shown that, when sufficient data is available, all the alternative algorithms tend to provide more accurate estimates compared to the conventional linear regression-based method, and particularly, Random Forest outperforms the other algorithms. However, the conventional method is a better tool when limited data is available.Keywords: artificial neural network, ground-motion models, machine learning, random forest, support vector machine
Procedia PDF Downloads 1221016 The Affordable Housing Problems of Elderly Households in the Istanbul Metropolitan Area
Authors: Elifsu Sahin
Abstract:
In the world and in Turkey, approximately 1 in 10 people is 65 years of age or older. The age group of 65 and over is the fastest-growing age group since 1990. This demographic aging trend and demographic transformation have spread over a long period in Western Europe and North America, while in Turkey, they have occurred over a relatively short period. The aging of the population poses many challenges in terms of housing supply, housing satisfaction, and economic access to housing, due to factors such as a decrease in the number of people in households, low incomes, and increased time spent in housing and housing neighborhoods. On the other hand, since 2000, neoliberal economic policies and government policies have led to serious growth in the construction and housing sectors in Turkey. During this process, the housing market in Turkey generally produced housing for high-income groups and foreigners. Housing has become an investment instrument, and rising housing prices and rents have seriously reduced both the affordability of housing and households' chances of living in healthy housing. Housing has become a growing problem for vulnerable groups such as low- and middle-income households, students, refugees, and the elderly. Moreover, in recent years, international migration, pandemics, economic crises, inflation, and the expected Istanbul earthquake have raised housing prices and rent in Turkey as a whole, especially in Istanbul. The aim of the study is to investigate how elderly households that don't own homes deal with the economic accessibility of housing and other affordability-related housing problems in the Istanbul Metropolitan Area today, when housing becomes an investment instrument, the issue of social housing is not on the agenda, and households can be added to the market according to their ability to pay. A complex method was adopted in the research, using a combination of various statistical data and interview findings. Based on household income, in-depth interviews were conducted with 100 elderly households who don't own their own homes and were randomly selected in identified neighborhoods, analyzing the micro-area within the districts in the Istanbul Metropolitan Area, where middle- and low-income households are concentrated. The study found that more than 50% of the net income of elderly households was spent on rent and other housing expenses. Some of the households said that they restrict spending on food, health, and entertainment because of their housing expenses. Among the findings of the study is that households receive financial support from their children or move into their children’s house for economic reasons. Due to the decrease in household income, especially after the loss of a spouse, the single individual moves into their children’s house. Moreover, some of the interviewed households had to change their house and move to a smaller, lower-rent house on the urban periphery for economic reasons after retirement, especially after 2020, despite their unwillingness.Keywords: affordable housing, elderly households, housing policy, istanbul metropolitan area
Procedia PDF Downloads 331015 Scenario of Some Minerals and Impact of Promoter Hypermethylation of DAP-K Gene in Gastric Carcinoma Patients of Kashmir Valley
Authors: Showkat Ahmad Bhat, Iqra Reyaz, Falaque ul Afshan, Ahmad Arif Reshi, Muneeb U. Rehman, Manzoor R. Mir, Sabhiya Majid, Sonallah, Sheikh Bilal, Ishraq Hussain
Abstract:
Background: Gastric cancer is the fourth most common cancer and the second leading cause of worldwide cancer-related deaths, with a wide variation in incidence rates across different geographical areas. The current view of cancer is that a malignancy arises from a transformation of the genetic material of a normal cell, followed by successive mutations and by chain of alterations in genes such as DNA repair genes, oncogenes, Tumor suppressor genes. Minerals are necessary for the functioning of several transcriptional factors, proteins that recognize certain DNA sequences and have been found to play a role in gastric cancer. Material Methods:The present work was a case control study and its aim was to ascertain the role of minerals and promoter hypermethylation of CpG islands of DAP-K gene in Gastric cancer patients among the Kashmiri population. Serum was extracted from all the samples and mineral estimation was done by AAS from serum, DNA was also extracted and was modified using bisulphite modification kit. Methylation-specific PCR was used for the analysis of the promoter hypermethylation status of DAP-K gene. The epigenetic analysis revealed that unlike other high risk regions, Kashmiri population has a different promoter hypermethylation profile of DAP-K gene and has different mineral profile. Results: In our study mean serum copper levels were significantly different for the two genders (p<0.05), while as no significant differences were observed for iron and zinc levels. In Methylation-specific PCR the methylation status of the promoter region of DAP-K gene was as 67.50% (27/40) of the gastric cancer tissues showed methylated DAP-K promoter and 32.50% (13/40) of the cases however showed unmethylated DAP-K promoter. Almost all 85% (17/20) of the histopathologically confirmed normal tissues showed unmethylated DAP-K promoter except only in 3 cases where DAP-K promoter was found to be methylated. The association of promoter hypermethylation with gastric cancer was evaluated by χ2 (Chi square) test and was found to be significant (P=0.0006). Occurrence of DAP-K methylation was found to be unequally distributed in males and females with more frequency in males than in females but the difference was not statistically significant (P =0.7635, Odds ratio=1.368 and 95% C.I=0.4197 to 4.456). When the frequency of DAP-K promoter methylation was compared with clinical staging of the disease, DAP-K promoter methylation was found to be certainly higher in Stage III/IV (85.71%) compared to Stage I/ II (57.69%) but the difference was not statistically significant (P =0.0673). These results suggest that DAP-K aberrant promoter hypermethylation in Kashmiri population contributes to the process of carcinogenesis in Gastric cancer and is reportedly one of the commonest epigenetic changes in the development of Gastric cancer.Keywords: gastric cancer, minerals, AAS, hypermethylation, CpG islands, DAP-K gene
Procedia PDF Downloads 5171014 Intelligent Control of Agricultural Farms, Gardens, Greenhouses, Livestock
Authors: Vahid Bairami Rad
Abstract:
The intelligentization of agricultural fields can control the temperature, humidity, and variables affecting the growth of agricultural products online and on a mobile phone or computer. Smarting agricultural fields and gardens is one of the best and best ways to optimize agricultural equipment and has a 100 percent direct effect on the growth of plants and agricultural products and farms. Smart farms are the topic that we are going to discuss today, the Internet of Things and artificial intelligence. Agriculture is becoming smarter every day. From large industrial operations to individuals growing organic produce locally, technology is at the forefront of reducing costs, improving results and ensuring optimal delivery to market. A key element to having a smart agriculture is the use of useful data. Modern farmers have more tools to collect intelligent data than in previous years. Data related to soil chemistry also allows people to make informed decisions about fertilizing farmland. Moisture meter sensors and accurate irrigation controllers have made the irrigation processes to be optimized and at the same time reduce the cost of water consumption. Drones can apply pesticides precisely on the desired point. Automated harvesting machines navigate crop fields based on position and capacity sensors. The list goes on. Almost any process related to agriculture can use sensors that collect data to optimize existing processes and make informed decisions. The Internet of Things (IoT) is at the center of this great transformation. Internet of Things hardware has grown and developed rapidly to provide low-cost sensors for people's needs. These sensors are embedded in IoT devices with a battery and can be evaluated over the years and have access to a low-power and cost-effective mobile network. IoT device management platforms have also evolved rapidly and can now be used securely and manage existing devices at scale. IoT cloud services also provide a set of application enablement services that can be easily used by developers and allow them to build application business logic. Focus on yourself. These development processes have created powerful and new applications in the field of Internet of Things, and these programs can be used in various industries such as agriculture and building smart farms. But the question is, what makes today's farms truly smart farms? Let us put this question in another way. When will the technologies associated with smart farms reach the point where the range of intelligence they provide can exceed the intelligence of experienced and professional farmers?Keywords: food security, IoT automation, wireless communication, hybrid lifestyle, arduino Uno
Procedia PDF Downloads 561013 Principal Well-Being at Hong Kong: A Quantitative Investigation
Authors: Junjun Chen, Yingxiu Li
Abstract:
The occupational well-being of school principals has played a vital role in the pursuit of individual and school wellness and success. However, principals’ well-being worldwide is under increasing threat because of the challenging and complex nature of their work and growing demands for school standardisation and accountability. Pressure is particularly acute in the post-pandemicfuture as principals attempt to deal with the impact of the pandemic on top of more regular demands. This is particularly true in Hong Kong, as school principals are increasingly wedged between unparalleled political, social, and academic responsibilities. Recognizing the semantic breadth of well-being, scholars have not determined a single, mutually agreeable definition but agreed that the concept of well-being has multiple dimensions across various disciplines. The multidimensional approach promises more precise assessments of the relationships between well-being and other concepts than the ‘affect-only’ approach or other single domains for capturing the essence of principal well-being. The multiple-dimension well-being concept is adopted in this project to understand principal well-being in this study. This study aimed to understand the situation of principal well-being and its influential drivers with a sample of 670 principals from Hong Kong and Mainland China. An online survey was sent to the participants after the breakout of COVID-19 by the researchers. All participants were well informed about the purposes and procedure of the project and the confidentiality of the data prior to filling in the questionnaire. Confirmatory factor analysis and structural equation modelling performed with Mplus were employed to deal with the dataset. The data analysis procedure involved the following three steps. First, the descriptive statistics (e.g., mean and standard deviation) were calculated. Second, confirmatory factor analysis (CFA) was used to trim principal well-being measurement performed with maximum likelihood estimation. Third, structural equation modelling (SEM) was employed to test the influential factors of principal well-being. The results of this study indicated that the overall of principal well-being were above the average mean score. The highest ranking in this study given by the principals was to their psychological and social well-being (M = 5.21). This was followed by spiritual (M = 5.14; SD = .77), cognitive (M = 5.14; SD = .77), emotional (M = 4.96; SD = .79), and physical well-being (M = 3.15; SD = .73). Participants ranked their physical well-being the lowest. Moreover, professional autonomy, supervisor and collegial support, school physical conditions, professional networking, and social media have showed a significant impact on principal well-being. The findings of this study will potentially enhance not only principal well-being, but also the functioning of an individual principal and a school without sacrificing principal well-being for quality education in the process. This will eventually move one step forward for a new future - a wellness society advocated by OECD. Importantly, well-being is an inside job that begins with choosing to have wellness, whilst supports to become a wellness principal are also imperative.Keywords: well-being, school principals, quantitative, influential factors
Procedia PDF Downloads 831012 The Use of Punctuation by Primary School Students Writing Texts Collaboratively: A Franco-Brazilian Comparative Study
Authors: Cristina Felipeto, Catherine Bore, Eduardo Calil
Abstract:
This work aims to analyze and compare the punctuation marks (PM) in school texts of Brazilian and French students and the comments on these PM made spontaneously by the students during the ongoing text. Assuming textual genetics as an investigative field within a dialogical and enunciative approach, we defined a common methodological design in two 1st year classrooms (7 years old) of the primary school, one classroom in Brazil (Maceio) and the other one in France (Paris). Through a multimodal capture system of writing processes in real time and space (Ramos System), we recorded the collaborative writing proposal in dyads in each of the classrooms. This system preserves the classroom’s ecological characteristics and provides a video recording synchronized with dialogues, gestures and facial expressions of the students, the stroke of the pen’s ink on the sheet of paper and the movement of the teacher and students in the classroom. The multimodal register of the writing process allowed access to the text in progress and the comments made by the students on what was being written. In each proposed text production, teachers organized their students in dyads and requested that they should talk, combine and write a fictional narrative. We selected a Dyad of Brazilian students (BD) and another Dyad of French students (FD) and we have filmed 6 proposals for each of the dyads. The proposals were collected during the 2nd Term of 2013 (Brazil) and 2014 (France). In 6 texts written by the BD there were identified 39 PMs and 825 written words (on average, a PM every 23 words): Of these 39 PMs, 27 were highlighted orally and commented by either student. In the texts written by the FD there were identified 48 PMs and 258 written words (on average, 1 PM every 5 words): Of these 48 PM, 39 were commented by the French students. Unlike what the studies on punctuation acquisition point out, the PM that occurred the most were hyphens (BD) and commas (FD). Despite the significant difference between the types and quantities of PM in the written texts, the recognition of the need for writing PM in the text in progress and the comments have some common characteristics: i) the writing of the PM was not anticipated in relation to the text in progress, then they were added after the end of a sentence or after the finished text itself; ii) the need to add punctuation marks in the text came after one of the students had ‘remembered’ that a particular sign was needed; iii) most of the PM inscribed were not related to their linguistic functions, but the graphic-visual feature of the text; iv) the comments justify or explain the PM, indicating metalinguistic reflections made by the students. Our results indicate how the comments of the BD and FD express the dialogic and subjective nature of knowledge acquisition. Our study suggests that the initial learning of PM depends more on its graphic features and interactional conditions than on its linguistic functions.Keywords: collaborative writing, erasure, graphic marks, learning, metalinguistic awareness, textual genesis
Procedia PDF Downloads 1621011 Biorefinery as Extension to Sugar Mills: Sustainability and Social Upliftment in the Green Economy
Authors: Asfaw Gezae Daful, Mohsen Alimandagari, Kathleen Haigh, Somayeh Farzad, Eugene Van Rensburg, Johann F. Görgens
Abstract:
The sugar industry has to 're-invent' itself to ensure long-term economic survival and opportunities for job creation and enhanced community-level impacts, given increasing pressure from fluctuating and low global sugar prices, increasing energy prices and sustainability demands. We propose biorefineries for re-vitalisation of the sugar industry using low value lignocellulosic biomass (sugarcane bagasse, leaves, and tops) annexed to existing sugar mills, producing a spectrum of high value platform chemicals along with biofuel, bioenergy, and electricity. Opportunity is presented for greener products, to mitigate climate change and overcome economic challenges. Xylose from labile hemicellulose remains largely underutilized and the conversion to value-add products a major challenge. Insight is required on pretreatment and/or extraction to optimize production of cellulosic ethanol together with lactic acid, furfural or biopolymers from sugarcane bagasse, leaves, and tops. Experimental conditions for alkaline and pressurized hot water extraction dilute acid and steam explosion pretreatment of sugarcane bagasse and harvest residues were investigated to serve as a basis for developing various process scenarios under a sugarcane biorefinery scheme. Dilute acid and steam explosion pretreatment were optimized for maximum hemicellulose recovery, combined sugar yield and solids digestibility. An optimal range of conditions for alkaline and liquid hot water extraction of hemicellulosic biopolymers, as well as conditions for acceptable enzymatic digestibility of the solid residue, after such extraction was established. Using data from the above, a series of energy efficient biorefinery scenarios are under development and modeled using Aspen Plus® software, to simulate potential factories to better understand the biorefinery processes and estimate the CAPEX and OPEX, environmental impacts, and overall viability. Rigorous and detailed sustainability assessment methodology was formulated to address all pillars of sustainability. This work is ongoing and to date, models have been developed for some of the processes which can ultimately be combined into biorefinery scenarios. This will allow systematic comparison of a series of biorefinery scenarios to assess the potential to reduce negative impacts on and maximize the benefits of social, economic, and environmental factors on a lifecycle basis.Keywords: biomass, biorefinery, green economy, sustainability
Procedia PDF Downloads 5141010 Corporate Life Cycle and Corporate Social Responsibility Performance: Empirical Evidence from Pharmaceutical Industry in China
Authors: Jing (Claire) LI
Abstract:
The topic of corporate social responsibility (CSR) is significant for pharmaceutical companies in China at this current stage. This is because, as a rapid growth industry in China in recent years, the pharmaceutical industry in China has been undergone continuous and terrible incidents relating to CSR. However, there is limited research and practice of CSR in Chinese pharmaceutical companies. Also, there is an urgent call for more research in an international context to understand the implications of corporate life cycle on CSR performance. To respond to the research need and research call, this study examines the relationship between corporate life cycle and CSR performance of Chinese listed companies in pharmaceutical industry. This research studies Chinese listed companies in pharmaceutical industry for the period of 2010-2017, where the data is available in database. Following the literature, this study divides CSR performance with regards to CSR dimensions, including shareholders, creditors, employees, customers, suppliers, the government, and the society. This study uses CSR scores of HEXUN database and financial measures of these CSR dimensions to measure the CSR performance. This study performed regression analysis to examine the relationship between corporate life cycle stages and CSR performance with regards to CSR dimensions for pharmaceutical listed companies in China. Using cash flow pattern as proxy of corporate life cycle to classify corporate life cycle stages, this study found that most (least) pharmaceutical companies in China are in maturity (decline) stage. This study found that CSR performance for most dimensions are highest (lowest) in maturity (decline) stage as well. Among these CSR dimensions, performing responsibilities for shareholder is the most important among all CSR responsibilities for pharmaceutical companies. This study is the first to provide important empirical evidence from Chinese pharmaceutical industry on the association between life cycle and CSR performance, supporting that corporate life cycle is a key factor in CSR performance. The study expands corporate life cycle and CSR literatures and has both empirical and theoretical contributions to the literature. From perspective of empirical contributions, the findings contribute to the argument that whether there is a relationship between CSR performance and various corporate life cycle stages in the literature. This study also provides empirical evidence that companies in different corporate life cycles have difference in CSR performance. From perspective of theoretical contributions, this study relates CSR and stakeholders to corporate life cycle stages and complements the corporate life cycle and CSR literature. This study has important implications for managers and policy makers. First, the results will be helpful for managers to have an understanding in the essence of CSR, and their company’s current and future CSR focus over corporate life cycle. This study provides a reference for their actions and may help them make more wise resources allocation decisions of CSR investment. Second, policy makers (in the government, stock exchanges, and securities commission) may consider corporate life cycle as an important factor in formulating future regulations for companies. Future research can explore the "process-based" differences in CSR performance and more industries.Keywords: China, corporate life cycle, corporate social responsibility, pharmaceutical industry
Procedia PDF Downloads 1051009 The Effects of Stoke's Drag, Electrostatic Force and Charge on Penetration of Nanoparticles through N95 Respirators
Authors: Jacob Schwartz, Maxim Durach, Aniruddha Mitra, Abbas Rashidi, Glen Sage, Atin Adhikari
Abstract:
NIOSH (National Institute for Occupational Safety and Health) approved N95 respirators are commonly used by workers in construction sites where there is a large amount of dust being produced from sawing, grinding, blasting, welding, etc., both electrostatically charged and not. A significant portion of airborne particles in construction sites could be nanoparticles created beside coarse particles. The penetration of the particles through the masks may differ depending on the size and charge of the individual particle. In field experiments relevant to this current study, we found that nanoparticles of medium size ranges are penetrating more frequently than nanoparticles of smaller and larger sizes. For example, penetration percentages of nanoparticles of 11.5 – 27.4 nm into a sealed N95 respirator on a manikin head ranged from 0.59 to 6.59%, whereas nanoparticles of 36.5 – 86.6 nm ranged from 7.34 to 16.04%. The possible causes behind this increased penetration of mid-size nanoparticles through mask filters are not yet explored. The objective of this study is to identify causes behind this unusual behavior of mid-size nanoparticles. We have considered such physical factors as Boltzmann distribution of the particles in thermal equilibrium with the air, kinetic energy of the particles at impact on the mask, Stoke’s drag force, and electrostatic forces in the mask stopping the particles. When the particles collide with the mask, only the particles that have enough kinetic energy to overcome the energy loss due to the electrostatic forces and the Stokes’ drag in the mask can pass through the mask. To understand this process, the following assumptions were made: (1) the effect of Stoke’s drag depends on the particles’ velocity at entry into the mask; (2) the electrostatic force is proportional to the charge on the particles, which in turn is proportional to the surface area of the particles; (3) the general dependence on electrostatic charge and thickness means that for stronger electrostatic resistance in the masks and thicker the masks’ fiber layers the penetration of particles is reduced, which is a sensible conclusion. In sampling situations where one mask was soaked in alcohol eliminating electrostatic interaction the penetration was much larger in the mid-range than the same mask with electrostatic interaction. The smaller nanoparticles showed almost zero penetration most likely because of the small kinetic energy, while the larger sized nanoparticles showed almost negligible penetration most likely due to the interaction of the particle with its own drag force. If there is no electrostatic force the fraction for larger particles grows. But if the electrostatic force is added the fraction for larger particles goes down, so diminished penetration for larger particles should be due to increased electrostatic repulsion, may be due to increased surface area and therefore larger charge on average. We have also explored the effect of ambient temperature on nanoparticle penetrations and determined that the dependence of the penetration of particles on the temperature is weak in the range of temperatures in the measurements 37-42°C, since the factor changes in the range from 3.17 10-3K-1 to 3.22 10-3K-1.Keywords: respiratory protection, industrial hygiene, aerosol, electrostatic force
Procedia PDF Downloads 1941008 A New Approach for Preparation of Super Absorbent Polymers: In-Situ Surface Cross-Linking
Authors: Reyhan Özdoğan, Mithat Çelebi, Özgür Ceylan, Mehmet Arif Kaya
Abstract:
Super absorbent polymers (SAPs) are defined as materials that can absorb huge amount of water or aqueous solution in comparison to their own mass and retain in their lightly cross-linked structure. SAPs were produced from water soluble monomers via polymerization subsequently controlled crosslinking. SAPs are generally used for water absorbing applications such as baby diapers, patient or elder pads and other hygienic product industries. Crosslinking density (CD) of SAP structure is an essential factor for water absortion capacity (WAC). Low internal CD leads to high WAC values and vice versa. However, SAPs have low CD and high swelling capacities and tend to disintegrate when pressure is applied upon them, so SAPs under load cannot absorb liquids effectively. In order to prevent this undesired situation and to obtain suitable SAP structures having high swelling capacity and ability to work under load, surface crosslinking can be the answer. In industry, these superabsorbent gels are mostly produced via solution polymerization and then they need to be dried, grinded, sized, post polymerized and finally surface croslinked (involves spraying of a crosslinking solution onto dried and grinded SAP particles, and then curing by heat). It can easily be seen that these steps are time consuming and should be handled carefully for the desired final product. If we could synthesize desired final SAPs using less processes it will help reducing time and production costs which are very important for any industries. In this study, synthesis of SAPs were achieved successfully by inverse suspension (Pickering type) polymerization and subsequently in-situ surface cross-linking via using proper surfactants in high boiling point solvents. Our one-pot synthesis of surface cross-linked SAPs invovles only one-step for preparation, thus it can be said that this technique exhibits more preferable characteristic for the industry in comparison to conventional methods due to its one-step easy process. Effects of different surface crosslinking agents onto properties of poly(acrylic acid-co-sodium acrylate) based SAPs are investigated. Surface crosslink degrees are evaluated by swelling under load (SUL) test. It was determined water absorption capacities of obtained SAPs decrease with the increasing surface crosslink density while their mechanic properties are improved.Keywords: inverse suspension polymerization, polyacrylic acid, super absorbent polymers (SAPs), surface crosslinking, sodium polyacrylate
Procedia PDF Downloads 3231007 Grain Size Statistics and Depositional Pattern of the Ecca Group Sandstones, Karoo Supergroup in the Eastern Cape Province, South Africa
Authors: Christopher Baiyegunhi, Kuiwu Liu, Oswald Gwavava
Abstract:
Grain size analysis is a vital sedimentological tool used to unravel the hydrodynamic conditions, mode of transportation and deposition of detrital sediments. In this study, detailed grain-size analysis was carried out on thirty-five sandstone samples from the Ecca Group in the Eastern Cape Province of South Africa. Grain-size statistical parameters, bivariate analysis, linear discriminate functions, Passega diagrams and log-probability curves were used to reveal the depositional processes, sedimentation mechanisms, hydrodynamic energy conditions and to discriminate different depositional environments. The grain-size parameters show that most of the sandstones are very fine to fine grained, moderately well sorted, mostly near-symmetrical and mesokurtic in nature. The abundance of very fine to fine grained sandstones indicates the dominance of low energy environment. The bivariate plots that the samples are mostly grouped, except for the Prince Albert samples that show scattered trend, which is due to the either mixture of two modes in equal proportion in bimodal sediments or good sorting in unimodal sediments. The linear discriminant function (LDF) analysis is dominantly indicative of turbidity current deposits under shallow marine environments for samples from the Prince Albert, Collingham and Ripon Formations, while those samples from the Fort Brown Formation are fluvial (deltaic) deposits. The graphic mean value shows the dominance of fine sand-size particles, which point to relatively low energy conditions of deposition. In addition, the LDF results point to low energy conditions during the deposition of the Prince Albert, Collingham and part of the Ripon Formation (Pluto Vale and Wonderfontein Shale Members), whereas the Trumpeters Member of the Ripon Formation and the overlying Fort Brown Formation accumulated under high energy conditions. The CM pattern shows a clustered distribution of sediments in the PQ and QR segments, indicating that the sediments were deposited mostly by suspension and rolling/saltation, and graded suspension. Furthermore, the plots also show that the sediments are mainly deposited by turbidity currents. Visher diagrams show the variability of hydraulic depositional conditions for the Permian Ecca Group sandstones. Saltation is the major process of transportation, although suspension and traction also played some role during deposition of the sediments. The sediments were mainly in saltation and suspension before being deposited.Keywords: grain size analysis, hydrodynamic condition, depositional environment, Ecca Group, South Africa
Procedia PDF Downloads 4811006 Exploration of the Psychological Aspect of Empowerment of Marginalized Women Working in the Unorganized Sector
Authors: Sharmistha Chanda, Anindita Choudhuri
Abstract:
This exploratory study highlights the psychological aspects of women's empowerment to find the importance of the psychological dimension of empowerment, such as; meaning, competence, self-determination, impact, and assumption, especially in the weaker marginalized section of women. A large proportion of rural, suburban, and urban poor survive by working in unorganized sectors of metropolitan cities. Relative Poverty and lack of employment in rural areas and small towns drive many people to the metropolitan city for work and livelihood. Women working in that field remain unrecognized as people of low socio-economic status. They are usually willing to do domestic work as daily wage workers, single wage earners, street vendors, family businesses like agricultural activities, domestic workers, and self-employed. Usually, these women accept such jobs because they do not have such an opportunity as they lack the basic level of education that is required for better-paid jobs. The unorganized sector, on the other hand, has no such clear-cut employer-employee relationships and lacks most forms of social protection. Having no fixed employer, these workers are casual, contractual, migrant, home-based, own-account workers who attempt to earn a living from whatever meager assets and skills they possess. Women have become more empowered both financially and individually through small-scale business ownership or entrepreneurship development and in household-based work. In-depth interviews have been done with 10 participants in order to understand their living styles, habits, self-identity, and empowerment in their society in order to evaluate the key challenges that they may face following by qualitative research approach. Transcription has been done from the collected data. The three-layer coding technique guides the data analysis process, encompassing – open coding, axial coding, and selective coding. Women’s Entrepreneurship is one of the foremost concerns as the Government, and non-government institutions are readily serving this domain with the primary objectives of promoting self-employment opportunities in general and empowering women in specific. Thus, despite hardship and unrecognition unorganized sector provides a huge array of opportunities for rural and sub-urban poor to earn. Also, the upper section of society tends to depend on this working force. This study gave an idea about the well-being, and meaning in life, life satisfaction on the basis of their lived experience.Keywords: marginalized women, psychological empowerment, relative poverty, unorganized sector
Procedia PDF Downloads 621005 Reflections of Narrative Architecture in Transformational Representations on the Architectural Design Studio
Authors: M. Mortas, H. Asar, P. Dursun Cebi
Abstract:
The visionary works of architectural representation in the 21st century's present situation, are practiced through the methodologies which try to expose the intellectual and theoretical essences of futurologist positions that are revealed with this era's interactions. Expansions of conceptual and contextual inputs related to one architectural design representation, depend on its deepness of critical attitudes, its interactions with the concepts such as experience, meaning, affection, psychology, perception and aura, as well as its communication with spatial, cultural and environmental factors. The purpose of this research study is to be able to offer methodological application areas for the design dimensions of experiential practices into architectural design studios, by focusing on the architectural representative narrations of 'transformation,' 'metamorphosis,' 'morphogenesis,' 'in-betweenness', 'superposition' and 'intertwine’ in which they affect and are affected by the today’s spatiotemporal hybridizations of architecture. The narrative representations and the visual theory paradigms of the designers are chosen under the main title of 'transformation' for the investigation of these visionary and critical representations' dismantlings and decodings. Case studies of this research area are chosen from Neil Spiller, Bryan Cantley, Perry Kulper and Dan Slavinsky’s transformative, morphogenetic representations. The theoretical dismantlings and decodings which are obtained from these artists’ contemporary architectural representations are tried to utilize and practice in the structural design studios as alternative methodologies when to approach architectural design processes, for enriching, differentiating, diversifying and 'transforming' the applications of so far used design process precedents. The research aims to indicate architectural students about how they can reproduce, rethink and reimagine their own representative lexicons and so languages of their architectural imaginations, regarding the newly perceived tectonics of prosthetic, biotechnology, synchronicity, nanotechnology or machinery into various experiential design workshops. The methodology of this work can be thought as revealing the technical and theoretical tools, lexicons and meanings of contemporary-visionary architectural representations of our decade, with the essential contents and components of hermeneutics, etymology, existentialism, post-humanism, phenomenology and avant-gardism disciplines to re-give meanings the architectural visual theorists’ transformative representations of our decade. The value of this study may be to emerge the superposed and overlapped atmospheres of futurologist architectural representations for the students who need to rethink on the transcultural, deterritorialized and post-humanist critical theories to create and use the representative visual lexicons of themselves for their architectural soft machines and beings by criticizing the now, to be imaginative for the future of architecture.Keywords: architectural design studio, visionary lexicon, narrative architecture, transformative representation
Procedia PDF Downloads 1411004 A Visualization Classification Method for Identifying the Decayed Citrus Fruit Infected by Fungi Based on Hyperspectral Imaging
Authors: Jiangbo Li, Wenqian Huang
Abstract:
Early detection of fungal infection in citrus fruit is one of the major problems in the postharvest commercialization process. The automatic and nondestructive detection of infected fruits is still a challenge for the citrus industry. At present, the visual inspection of rotten citrus fruits is commonly performed by workers through the ultraviolet induction fluorescence technology or manual sorting in citrus packinghouses to remove fruit subject with fungal infection. However, the former entails a number of problems because exposing people to this kind of lighting is potentially hazardous to human health, and the latter is very inefficient. Orange is used as a research object. This study would focus on this problem and proposed an effective method based on Vis-NIR hyperspectral imaging in the wavelength range of 400-1000 nm with a spectroscopic resolution of 2.8 nm. In this work, three normalization approaches are applied prior to analysis to reduce the effect of sample curvature on spectral profiles, and it is found that mean normalization was the most effective pretreatment for decreasing spectral variability due to curvature. Then, principal component analysis (PCA) was applied to a dataset composing of average spectra from decayed and normal tissue to reduce the dimensionality of data and observe the ability of Vis-NIR hyper-spectra to discriminate data from two classes. In this case, it was observed that normal and decayed spectra were separable along the resultant first principal component (PC1) axis. Subsequently, five wavelengths (band) centered at 577, 702, 751, 808, and 923 nm were selected as the characteristic wavelengths by analyzing the loadings of PC1. A multispectral combination image was generated based on five selected characteristic wavelength images. Based on the obtained multispectral combination image, the intensity slicing pseudocolor image processing method is used to generate a 2-D visual classification image that would enhance the contrast between normal and decayed tissue. Finally, an image segmentation algorithm for detection of decayed fruit was developed based on the pseudocolor image coupled with a simple thresholding method. For the investigated 238 independent set samples including infected fruits infected by Penicillium digitatum and normal fruits, the total success rate is 100% and 97.5%, respectively, and, the proposed algorithm also used to identify the orange infected by penicillium italicum with a 100% identification accuracy, indicating that the proposed multispectral algorithm here is an effective method and it is potential to be applied in citrus industry.Keywords: citrus fruit, early rotten, fungal infection, hyperspectral imaging
Procedia PDF Downloads 2991003 Effects of Temperature and the Use of Bacteriocins on Cross-Contamination from Animal Source Food Processing: A Mathematical Model
Authors: Benjamin Castillo, Luis Pastenes, Fernando Cerdova
Abstract:
The contamination of food by microbial agents is a common problem in the industry, especially regarding the elaboration of animal source products. Incorrect manipulation of the machinery or on the raw materials can cause a decrease in production or an epidemiological outbreak due to intoxication. In order to improve food product quality, different methods have been used to reduce or, at least, to slow down the growth of the pathogens, especially deteriorated, infectious or toxigenic bacteria. These methods are usually carried out under low temperatures and short processing time (abiotic agents), along with the application of antibacterial substances, such as bacteriocins (biotic agents). This, in a controlled and efficient way that fulfills the purpose of bacterial control without damaging the final product. Therefore, the objective of the present study is to design a secondary mathematical model that allows the prediction of both the biotic and abiotic factor impact associated with animal source food processing. In order to accomplish this objective, the authors propose a three-dimensional differential equation model, whose components are: bacterial growth, release, production and artificial incorporation of bacteriocins and changes in pH levels of the medium. These three dimensions are constantly being influenced by the temperature of the medium. Secondly, this model adapts to an idealized situation of cross-contamination animal source food processing, with the study agents being both the animal product and the contact surface. Thirdly, the stochastic simulations and the parametric sensibility analysis are compared with referential data. The main results obtained from the analysis and simulations of the mathematical model were to discover that, although bacterial growth can be stopped in lower temperatures, even lower ones are needed to eradicate it. However, this can be not only expensive, but counterproductive as well in terms of the quality of the raw materials and, on the other hand, higher temperatures accelerate bacterial growth. In other aspects, the use and efficiency of bacteriocins are an effective alternative in the short and medium terms. Moreover, an indicator of bacterial growth is a low-level pH, since lots of deteriorating bacteria are lactic acids. Lastly, the processing times are a secondary agent of concern when the rest of the aforementioned agents are under control. Our main conclusion is that when acclimating a mathematical model within the context of the industrial process, it can generate new tools that predict bacterial contamination, the impact of bacterial inhibition, and processing method times. In addition, the mathematical modeling proposed logistic input of broad application, which can be replicated on non-meat food products, other pathogens or even on contamination by crossed contact of allergen foods.Keywords: bacteriocins, cross-contamination, mathematical model, temperature
Procedia PDF Downloads 1441002 Numerical Analysis and Parametric Study of Granular Anchor Pile on Expansive Soil Using Finite Element Method: Case of Addis Ababa, Bole Sub-City
Authors: Abdurahman Anwar Shfa
Abstract:
Addis Ababa is among the fastest-growing urban areas in the country. There are many new constructions of public and private condominiums and large new low rising residential buildings for residents. But the wide range of heaving problems of expansive soil in the city become a major difficulty for the construction sector, especially in low rising buildings, by causing different problems such as distortion and cracking of floor slabs, cracks in grade beams, and walls, jammed or misaligned Doors and Windows; failure of blocks supporting grade beams. Hence an attractive and economical design solution may be required for such type of problem. Therefore, this research works to publicize a recent innovation called the Granular Anchor Pile system for the reduction of the heave effect of expansive soil. This research is written for the objective of numerical investigation of the behavior of Granular Anchor Pile under the heave using Finite element analysis PLAXIS 3D program by means of studying the effect of different parameters like length of the pile, diameter of pile, and pile group by applying prescribed displacement of 10% of pile diameter at the center of granular pile anchor. An additional objective is examining the suitability of Granular Anchor Pile as an alternative solution for heave problems in expansive soils mostly for low rising buildings found in Addis Ababa City, especially in Bole Sub-City, by considering different factors such as the local availability of construction materials, economy for the construction, installation process condition, environmental benefit, time consumption and performance of the pile. Accordingly, the performance of the pile improves when the length of the pile increases. This is due to an increase in the self-weight of the pile and friction mobilized between the pile and soil interface. Additionally, the uplift capacity of the pile decreases when increasing the pile diameter and spacing between the piles in the group due to a reduction in the number of piles in the group. But, few cases show that the uplift capacity of the pile increases with increasing the pile diameter for a constant number of piles in the group and increasing the spacing between the pile and in the case of single pile capacity. This is due to the increment of piles' self-weight and surface area of the pile group and also the decrement of stress overlap in the soil caused by piles respectively. According to the suitability analysis, it is observed that Granular Anchor Pile is sensible or practical to apply for the actual problem of Expansive soil in a low rising building constructed in the country because of its convenience for all considerations.Keywords: expansive soil, granular anchor pile, PLAXIS, suitability analysis
Procedia PDF Downloads 351001 Creating Standards to Define the Role of Employment Specialists: A Case Study
Authors: Joseph Ippolito, David Megenhardt
Abstract:
In the United States, displaced workers, the unemployed and those seeking to build additional work skills are provided employment training and job placement services through a system of One-Stop Career Centers that are sponsored by the country’s 593 local Workforce Boards. During the period 2010-2015, these centers served roughly 8 million individuals each year. The quality of services provided at these centers rests upon professional employment specialists who work closely with clients to identify their job interests, to connect them to appropriate training opportunities, to match them with needed supportive social services and to guide them to eventual employment. Despite the crucial role these Employment Specialists play, currently there are no broadly accepted standards that establish what these individuals are expected to do in the workplace, nor are there indicators to assess how well an individual performs these responsibilities. Education Development Center (EDC) and the United Labor Agency (ULA) have partnered to create a foundation upon which curriculum can be developed that addresses the skills, knowledge and behaviors that Employment Specialists must master in order to serve their clients effectively. EDC is a non-profit, education research and development organization that designs, implements, and evaluates programs to improve education, health and economic opportunity worldwide. ULA is the social action arm of organized labor in Greater Cleveland, Ohio. ULA currently operates One-Stop Career Centers in both Cleveland and Pittsburgh, Pennsylvania. This case study outlines efforts taken to create standards that define the work of Employment Specialists and to establish indicators that can guide assessment of work performance. The methodology involved in the study has engaged a panel of expert Employment Specialists in rigorous, structured dialogues that analyze and identify the characteristics that enable them to be effective in their jobs. It has also drawn upon and integrated reviews of the panel’s work by more than 100 other Employment Specialists across the country. The results of this process are two documents that provide resources for developing training curriculum for future Employment Specialists, namely: an occupational profile of an Employment Specialist that offers a detailed articulation of the skills, knowledge and behaviors that enable individuals to be successful at this job, and; a collection of performance based indicators, aligned to the profile, which illustrate what the work responsibilities of an Employment Specialist 'look like' a four levels of effectiveness ranging from novice to expert. The method of occupational analysis used by the study has application across a broad number of fields.Keywords: assessment, employability, job standards, workforce development
Procedia PDF Downloads 2341000 Team Teaching versus Traditional Pedagogical Method
Authors: L. M. H. Mustonen, S. A. Heikkilä
Abstract:
The focus of the paper is to describe team teaching as a HAMK’s pedagogical method, and its impacts to the teachers work. Background: Traditionally it is thought that teaching is a job where one mostly works alone. More and more teachers feel that their work is getting more stressful. Solutions to these problems have been sought in Häme University of Applied sciences’ (From now on referred to as HAMK). HAMK has made a strategic change to move to the group oriented working of teachers. Instead of isolated study courses, there are now larger 15 credits study modules. Implementation: As examples of the method, two cases are presented: technical project module and summer studies module, which was integrated into the EU development project called Energy Efficiency with Precise Control. In autumn 2017, technical project will be implemented third time. There are at least three teachers involved in it and it is the first module of the new students. Main focus is to learn the basic skills of project working. From communicational viewpoint, they learn the basics of written and oral reporting and the basics of video reporting skills. According to our quality control system, the need for the development is evaluated in the end of the module. There are always some differences in each implementation but the basics are the same. The other case summer studies 2017 is new and part of a larger EU project. For the first time, we took a larger group of first to third year students from different study programmes to the summer studies. The students learned professional skills and also skills from different fields of study, international cooperation, and communication skills. Benefits and challenges: After three years, it is possible to consider what the changes mean in the everyday work of the teachers - and of course – what it means to students and the learning process. The perspective is HAMK’s electrical and automation study programme: At first, the change always means more work. The routines born after many years and the course material used for years may not be valid anymore. Teachers are teaching in modules simultaneously and often with some subjects overlapping. Finding the time to plan the modules together is often difficult. The essential benefit is that the learning outcomes have improved. This can be seen in the feedback given by both the teachers and the students. Conclusions: A new type of working environment is being born. A team of teachers designs a module that matches the objectives and ponders the answers to such questions as what are the knowledge-based targets of the module? Which pedagogical solutions will achieve the desired results? At what point do multiple teachers instruct the class together? How is the module evaluated? How can the module be developed further for the next execution? The team discusses openly and finds the solutions. Collegiate responsibility and support are always present. These are strengthening factors of the new communal university teaching culture. They are also strong sources of pleasure of work.Keywords: pedagogical development, summer studies, team teaching, well-being at work
Procedia PDF Downloads 109999 Examining Smallholder Farmers’ Perceptions of Climate Change and Barriers to Strategic Adaptation in Todee District, Liberia
Authors: Joe Dorbor Wuokolo
Abstract:
Thousands of smallholder farmers in Todee District, Montserrado county, are currently vulnerable to the negative impact of climate change. The district, which is the agricultural hot spot for the county, is faced with unfavorable changes in the daily temperature due to climate change. Farmers in the district have observed a dramatic change in the ratio of rainfall to sunshine, which has caused a chilling effect on their crop yields. However, there is a lack of documentation regarding how farmers perceive and respond to these changes and challenges. A study was conducted in the region to examine the perceptions of smallholder farmers regarding the negative impact of climate change, the adaptation strategies practice, and the barriers that hinder the process of advancing adaptation strategy. On purpose, a sample of 41 respondents from five towns was selected, including five town chiefs, five youth leaders, five women leaders, and sixteen community members. Women and youth leaders were specifically chosen to provide gender balance and enhance the quality of the investigation. Additionally, to validate the barriers farmers face during adaptation to climate change, this study interviewed eight experts from local and international organizations and government ministries and agencies involved in climate change and agricultural programs on what they perceived as the major barrier in both local and national level that impede farmers adaptation to climate change impact. SPSS was used to code the data, and descriptive statistics were used to analyze the data. The weighted average index (WAI) was used to rank adaptation strategies and the perceived importance of adaptation practices among farmers. On a scale from 0 to 3, 0 indicates the least important technique, and 3 indicates the most effective technique. In addition, the Problem Confrontation Index (PCI) was used to rank the barriers that prevented farmers from implementing adaptation measures. According to the findings, approximately 60% of all respondents considered the use of irrigation systems to be the most effective adaptation strategy, with drought-resistant varieties making up 30% of the total. Additionally, 80% of respondents placed a high value on drought-resistant varieties, while 63% percent placed it on irrigation practices. In addition, 78% of farmers ranked and indicated that unpredictability of the weather is the most significant barrier to their adaptation strategies, followed by the high cost of farm inputs and lack of access to financing facilities. 80% of respondents believe that the long-term changes in precipitation (rainfall) and temperature (hotness) are accelerating. This suggests that decision-makers should adopt policies and increase the capacity of smallholder farmers to adapt to the negative impact of climate change in order to ensure sustainable food production.Keywords: adaptation strategies, climate change, farmers’ perception, smallholder farmers
Procedia PDF Downloads 82998 Distinguishing Substance from Spectacle in Violent Extremist Propaganda through Frame Analysis
Authors: John Hardy
Abstract:
Over the last decade, the world has witnessed an unprecedented rise in the quality and availability of violent extremist propaganda. This phenomenon has been fueled primarily by three interrelated trends: rapid adoption of online content mediums by creators of violent extremist propaganda, increasing sophistication of violent extremist content production, and greater coordination of content and action across violent extremist organizations. In particular, the self-styled ‘Islamic State’ attracted widespread attention from its supporters and detractors alike by mixing shocking video and imagery content in with substantive ideological and political content. Although this practice was widely condemned for its brutality, it proved to be effective at engaging with a variety of international audiences and encouraging potential supporters to seek further information. The reasons for the noteworthy success of this kind of shock-value propaganda content remain unclear, despite many governments’ attempts to produce counterpropaganda. This study examines violent extremist propaganda distributed by five terrorist organizations between 2010 and 2016, using material released by the Al Hayat Media Center of the Islamic State, Boko Haram, Al Qaeda, Al Qaeda in the Arabian Peninsula, and Al Qaeda in the Islamic Maghreb. The time period covers all issues of the infamous publications Inspire and Dabiq, as well as the most shocking video content released by the Islamic State and its affiliates. The study uses frame analysis to distinguish thematic from symbolic content in violent extremist propaganda by contrasting the ways that substantive ideology issues were framed against the use of symbols and violence to garner attention and to stylize propaganda. The results demonstrate that thematic content focuses significantly on diagnostic frames, which explain violent extremist groups’ causes, and prognostic frames, which propose solutions to addressing or rectifying the cause shared by groups and their sympathizers. Conversely, symbolic violence is primarily stylistic and rarely linked to thematic issues or motivational framing. Frame analysis provides a useful preliminary tool in disentangling substantive ideological and political content from stylistic brutality in violent extremist propaganda. This provides governments and researchers a method for better understanding the framing and content used to design narratives and propaganda materials used to promote violent extremism around the world. Increased capacity to process and understand violent extremist narratives will further enable governments and non-governmental organizations to develop effective counternarratives which promote non-violent solutions to extremists’ grievances.Keywords: countering violent extremism, counternarratives, frame analysis, propaganda, terrorism, violent extremism
Procedia PDF Downloads 173997 Economic Decision Making under Cognitive Load: The Role of Numeracy and Financial Literacy
Authors: Vânia Costa, Nuno De Sá Teixeira, Ana C. Santos, Eduardo Santos
Abstract:
Financial literacy and numeracy have been regarded as paramount for rational household decision making in the increasing complexity of financial markets. However, financial decisions are often made under sub-optimal circumstances, including cognitive overload. The present study aims to clarify how financial literacy and numeracy, taken as relevant expert knowledge for financial decision-making, modulate possible effects of cognitive load. Participants were required to perform a choice between a sure loss or a gambling pertaining a financial investment, either with or without a competing memory task. Two experiments were conducted varying only the content of the competing task. In the first, the financial choice task was made while maintaining on working memory a list of five random letters. In the second, cognitive load was based upon the retention of six random digits. In both experiments, one of the items in the list had to be recalled given its serial position. Outcomes of the first experiment revealed no significant main effect or interactions involving cognitive load manipulation and numeracy and financial literacy skills, strongly suggesting that retaining a list of random letters did not interfere with the cognitive abilities required for financial decision making. Conversely, and in the second experiment, a significant interaction between the competing mnesic task and level of financial literacy (but not numeracy) was found for the frequency of choice of a gambling option. Overall, and in the control condition, both participants with high financial literacy and high numeracy were more prone to choose the gambling option. However, and when under cognitive load, participants with high financial literacy were as likely as their illiterate counterparts to choose the gambling option. This outcome is interpreted as evidence that financial literacy prevents intuitive risk-aversion reasoning only under highly favourable conditions, as is the case when no other task is competing for cognitive resources. In contrast, participants with higher levels of numeracy were consistently more prone to choose the gambling option in both experimental conditions. These results are discussed in the light of the opposition between classical dual-process theories and fuzzy-trace theories for intuitive decision making, suggesting that while some instances of expertise (as numeracy) are prone to support easily accessible gist representations, other expert skills (as financial literacy) depend upon deliberative processes. It is furthermore suggested that this dissociation between types of expert knowledge might depend on the degree to which they are generalizable across disparate settings. Finally, applied implications of the present study are discussed with a focus on how it informs financial regulators and the importance and limits of promoting financial literacy and general numeracy.Keywords: decision making, cognitive load, financial literacy, numeracy
Procedia PDF Downloads 182996 Influence of Intra-Yarn Permeability on Mesoscale Permeability of Plain Weave and 3D Fabrics
Authors: Debabrata Adhikari, Mikhail Matveev, Louise Brown, Andy Long, Jan Kočí
Abstract:
A good understanding of mesoscale permeability of complex architectures in fibrous porous preforms is of particular interest in order to achieve efficient and cost-effective resin impregnation of liquid composite molding (LCM). Fabrics used in structural reinforcements are typically woven or stitched. However, 3D fabric reinforcement is of particular interest because of the versatility in the weaving pattern with the binder yarn and in-plain yarn arrangements to manufacture thick composite parts, overcome the limitation in delamination, improve toughness etc. To predict the permeability based on the available pore spaces between the inter yarn spaces, unit cell-based computational fluid dynamics models have been using the Stokes Darcy model. Typically, the preform consists of an arrangement of yarns with spacing in the order of mm, wherein each yarn consists of thousands of filaments with spacing in the order of μm. The fluid flow during infusion exchanges the mass between the intra and inter yarn channels, meaning there is no dead-end of flow between the mesopore in the inter yarn space and the micropore in the yarn. Several studies have employed the Brinkman equation to take into account the flow through dual-scale porosity reinforcement to estimate their permeability. Furthermore, to reduce the computational effort of dual scale flow, scale separation criteria based on the ratio between yarn permeability to the yarn spacing was also proposed to quantify the dual scale and negligible micro-scale flow regime for the prediction of mesoscale permeability. In the present work, the key parameter to identify the influence of intra yarn permeability on the mesoscale permeability has been investigated with the systematic study of weft and warp yarn spacing on the plane weave as well as the position of binder yarn and number of in-plane yarn layers on 3D weave fabric. The permeability tensor has been estimated using an OpenFOAM-based model for the various weave pattern with idealized geometry of yarn implemented using open-source software TexGen. Additionally, scale separation criterion has been established based on the various configuration of yarn permeability for the 3D fabric with both the isotropic and anisotropic yarn from Gebart’s model. It was observed that the variation of mesoscale permeability Kxx within 30% when the isotropic porous yarn is considered for a 3D fabric with binder yarn. Furthermore, the permeability model developed in this study will be used for multi-objective optimizations of the preform mesoscale geometry in terms of yarn spacing, binder pattern, and a number of layers with an aim to obtain improved permeability and reduced void content during the LCM process.Keywords: permeability, 3D fabric, dual-scale flow, liquid composite molding
Procedia PDF Downloads 96995 Calibration of Residential Buildings Energy Simulations Using Real Data from an Extensive in situ Sensor Network – A Study of Energy Performance Gap
Authors: Mathieu Bourdeau, Philippe Basset, Julien Waeytens, Elyes Nefzaoui
Abstract:
As residential buildings account for a third of the overall energy consumption and greenhouse gas emissions in Europe, building energy modeling is an essential tool to reach energy efficiency goals. In the energy modeling process, calibration is a mandatory step to obtain accurate and reliable energy simulations. Nevertheless, the comparison between simulation results and the actual building energy behavior often highlights a significant performance gap. The literature discusses different origins of energy performance gaps, from building design to building operation. Then, building operation description in energy models, especially energy usages and users’ behavior, plays an important role in the reliability of simulations but is also the most accessible target for post-occupancy energy management and optimization. Therefore, the present study aims to discuss results on the calibration ofresidential building energy models using real operation data. Data are collected through a sensor network of more than 180 sensors and advanced energy meters deployed in three collective residential buildings undergoing major retrofit actions. The sensor network is implemented at building scale and in an eight-apartment sample. Data are collected for over one year and half and coverbuilding energy behavior – thermal and electricity, indoor environment, inhabitants’ comfort, occupancy, occupants behavior and energy uses, and local weather. Building energy simulations are performed using a physics-based building energy modeling software (Pleaides software), where the buildings’features are implemented according to the buildingsthermal regulation code compliance study and the retrofit project technical files. Sensitivity analyses are performed to highlight the most energy-driving building features regarding each end-use. These features are then compared with the collected post-occupancy data. Energy-driving features are progressively replaced with field data for a step-by-step calibration of the energy model. Results of this study provide an analysis of energy performance gap on an existing residential case study under deep retrofit actions. It highlights the impact of the different building features on the energy behavior and the performance gap in this context, such as temperature setpoints, indoor occupancy, the building envelopeproperties but also domestic hot water usage or heat gains from electric appliances. The benefits of inputting field data from an extensive instrumentation campaign instead of standardized scenarios are also described. Finally, the exhaustive instrumentation solution provides useful insights on the needs, advantages, and shortcomings of the implemented sensor network for its replicability on a larger scale and for different use cases.Keywords: calibration, building energy modeling, performance gap, sensor network
Procedia PDF Downloads 159994 Examining Kokugaku as a Pattern of Defining Identity in Global Comparison
Authors: Mária Ildikó Farkas
Abstract:
Kokugaku of the Edo period can be seen as a key factor of defining cultural (and national) identity in the 18th and early 19th century based on Japanese cultural heritage. Kokugaku focused on Japanese classics, on exploring, studying and reviving (or even inventing) ancient Japanese language, literature, myths, history and also political ideology. ‘Japanese culture’ as such was distinguished from Chinese (and all other) cultures, ‘Japanese identity’ was thus defined. Meiji scholars used kokugaku conceptions of Japan to construct a modern national identity based on the premodern and culturalist conceptions of community. The Japanese cultural movement of the 18-19th centuries (kokugaku) of defining cultural and national identity before modernization can be compared not to the development of Western Europe (where national identity strongly attached to modern nation states) or other parts of Asia (where these emerged after the Western colonization), but rather with the ‘national awakening’ movements of the peoples of East Central Europe, a comparison which have not been dealt with in the secondary literature yet. The role of a common language, culture, history and myths in the process of defining cultural identity – following mainly Miroslav Hroch’s comparative and interdisciplinary theory of national development – can be examined compared to the movements of defining identity of the peoples of East Central Europe (18th-19th c). In the shadow of a cultural and/or political ‘monolith’ (China for Japan and Germany for Central Europe), before modernity, ethnic groups or communities started to evolve their own identities with cultural movements focusing on their own language and culture, thus creating their cultural identity, and in the end, a new sense of community, the nation. Comparing actual texts (‘narratives’) of the kokugaku scholars and Central European writers of the nation building period (18th and early 19th centuries) can reveal the similarities of the discourses of deliberate searches for identity. Similar motives of argument can be identified in these narratives: ‘language’ as the primary bearer of collective identity, the role of language in culture, ‘culture’ as the main common attribute of the community; and similar aspirations to explore, search and develop native language, ‘genuine’ culture, ‘original’ traditions. This comparative research offering ‘development patterns’ for interpretation can help us understand processes that may be ambiguously considered ‘backward’ or even ‘deleterious’ (e.g. cultural nationalism) or just ‘unique’. ‘Cultural identity’ played a very important role in the formation of national identity during modernization especially in the case of non-Western communities, who had to face the danger of losing their identities in the course of ‘Westernization’ accompanying modernization.Keywords: cultural identity, Japanese modernization, kokugaku, national awakening
Procedia PDF Downloads 271993 Assessment of Rainfall Erosivity, Comparison among Methods: Case of Kakheti, Georgia
Authors: Mariam Tsitsagi, Ana Berdzenishvili
Abstract:
Rainfall intensity change is one of the main indicators of climate change. It has a great influence on agriculture as one of the main factors causing soil erosion. Splash and sheet erosion are one of the most prevalence and harmful for agriculture. It is invisible for an eye at first stage, but the process will gradually move to stream cutting erosion. Our study provides the assessment of rainfall erosivity potential with the use of modern research methods in Kakheti region. The region is the major provider of wheat and wine in the country. Kakheti is located in the eastern part of Georgia and characterized quite a variety of natural conditions. The climate is dry subtropical. For assessment of the exact rate of rainfall erosion potential several year data of rainfall with short intervals are needed. Unfortunately, from 250 active metro stations running during the Soviet period only 55 of them are active now and 5 stations in Kakheti region respectively. Since 1936 we had data on rainfall intensity in this region, and rainfall erosive potential is assessed, in some old papers, but since 1990 we have no data about this factor, which in turn is a necessary parameter for determining the rainfall erosivity potential. On the other hand, researchers and local communities suppose that rainfall intensity has been changing and the number of haily days has also been increasing. However, finding a method that will allow us to determine rainfall erosivity potential as accurate as possible in Kakheti region is very important. The study period was divided into three sections: 1936-1963; 1963-1990 and 1990-2015. Rainfall erosivity potential was determined by the scientific literature and old meteorological stations’ data for the first two periods. And it is known that in eastern Georgia, at the boundary between steppe and forest zones, rainfall erosivity in 1963-1990 was 20-75% higher than that in 1936-1963. As for the third period (1990-2015), for which we do not have data of rainfall intensity. There are a variety of studies, where alternative ways of calculating the rainfall erosivity potential based on lack of data are discussed e.g.based on daily rainfall data, average annual rainfall data and the elevation of the area, etc. It should be noted that these methods give us a totally different results in case of different climatic conditions and sometimes huge errors in some cases. Three of the most common methods were selected for our research. Each of them was tested for the first two sections of the study period. According to the outcomes more suitable method for regional climatic conditions was selected, and after that, we determined rainfall erosivity potential for the third section of our study period with use of the most successful method. Outcome data like attribute tables and graphs was specially linked to the database of Kakheti, and appropriate thematic maps were created. The results allowed us to analyze the rainfall erosivity potential changes from 1936 to the present and make the future prospect. We have successfully implemented a method which can also be use for some another region of Georgia.Keywords: erosivity potential, Georgia, GIS, Kakheti, rainfall
Procedia PDF Downloads 224992 A Methodology Based on Image Processing and Deep Learning for Automatic Characterization of Graphene Oxide
Authors: Rafael do Amaral Teodoro, Leandro Augusto da Silva
Abstract:
Originated from graphite, graphene is a two-dimensional (2D) material that promises to revolutionize technology in many different areas, such as energy, telecommunications, civil construction, aviation, textile, and medicine. This is possible because its structure, formed by carbon bonds, provides desirable optical, thermal, and mechanical characteristics that are interesting to multiple areas of the market. Thus, several research and development centers are studying different manufacturing methods and material applications of graphene, which are often compromised by the scarcity of more agile and accurate methodologies to characterize the material – that is to determine its composition, shape, size, and the number of layers and crystals. To engage in this search, this study proposes a computational methodology that applies deep learning to identify graphene oxide crystals in order to characterize samples by crystal sizes. To achieve this, a fully convolutional neural network called U-net has been trained to segment SEM graphene oxide images. The segmentation generated by the U-net is fine-tuned with a standard deviation technique by classes, which allows crystals to be distinguished with different labels through an object delimitation algorithm. As a next step, the characteristics of the position, area, perimeter, and lateral measures of each detected crystal are extracted from the images. This information generates a database with the dimensions of the crystals that compose the samples. Finally, graphs are automatically created showing the frequency distributions by area size and perimeter of the crystals. This methodological process resulted in a high capacity of segmentation of graphene oxide crystals, presenting accuracy and F-score equal to 95% and 94%, respectively, over the test set. Such performance demonstrates a high generalization capacity of the method in crystal segmentation, since its performance considers significant changes in image extraction quality. The measurement of non-overlapping crystals presented an average error of 6% for the different measurement metrics, thus suggesting that the model provides a high-performance measurement for non-overlapping segmentations. For overlapping crystals, however, a limitation of the model was identified. To overcome this limitation, it is important to ensure that the samples to be analyzed are properly prepared. This will minimize crystal overlap in the SEM image acquisition and guarantee a lower error in the measurements without greater efforts for data handling. All in all, the method developed is a time optimizer with a high measurement value, considering that it is capable of measuring hundreds of graphene oxide crystals in seconds, saving weeks of manual work.Keywords: characterization, graphene oxide, nanomaterials, U-net, deep learning
Procedia PDF Downloads 160