Search results for: automated document processing
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 4947

Search results for: automated document processing

1167 Exploring Subjective Simultaneous Mixed Emotion Experiences in Middle Childhood

Authors: Esther Burkitt

Abstract:

Background: Evidence is mounting that mixed emotions can be experienced simultaneously in different ways across the lifespan. Four types of patterns of simultaneously mixed emotions (sequential, prevalent, highly parallel, and inverse types) have been identified in middle childhood and adolescence. Moreover, the recognition of these experiences tends to develop firstly when children consider peers rather than the self. This evidence from children and adolescents is based on examining the presence of experiences specified in adulthood. The present study, therefore, applied an exhaustive coding scheme to investigate whether children experience types of previously unidentified simultaneous mixed emotional experiences. Methodology: One hundred and twenty children (60 girls) aged 7 years 1 month - 9 years 2 months (X=8 years 1 month; SD = 10 months) were recruited from mainstream schools across the UK. Two age groups were formed (youngest, n = 61, 7 years 1 month- 8 years 1 months: oldest, n = 59, 8 years 2 months – 9 years 2 months) and allocated to one of two conditions hearing vignettes describing happy and sad mixed emotion events in age and gender-matched protagonist or themselves. Results: Loglinear analyses identified new types of flexuous, vertical, and other experiences along with established sequential, prevalent, highly parallel, and inverse types of experience. Older children recognised more complex experiences other than the self-condition. Conclusion: Several additional types of simultaneously mixed emotions are recognised in middle childhood. The theoretical relevance of simultaneous mixed emotion processing in childhood is considered, and the potential utility of the findings in emotion assessments is discussed.

Keywords: emotion, childhood, self, other

Procedia PDF Downloads 61
1166 Artificial Intelligence in Art and Other Sectors: Selected Aspects of Mutual Impact

Authors: Justyna Minkiewicz

Abstract:

Artificial Intelligence (AI) applied in the arts may influence the development of AI knowledge in other sectors and then also impact mutual collaboration with the artistic environment. Hence this collaboration may also impact the development of art projects. The paper will reflect the qualitative research outcomes based on in-depth (IDI) interviews within the marketing sector in Poland and desk research. Art is a reflection of the spirit of our times. Moreover, now we are experiencing a significant acceleration in the development of technologies and their use in various sectors. The leading technologies that contribute to the development of the economy, including the creative sector, embrace technologies such as artificial intelligence, blockchain, extended reality, voice processing, and virtual beings. Artificial intelligence is one of the leading technologies developed for several decades, which is currently reaching a high level of interest and use in various sectors. However, the conducted research has shown that there is still low awareness of artificial intelligence and its wide application in various sectors. The study will show how artists use artificial intelligence in their art projects and how it can be translated into practice within the business. At the same time, the paper will raise awareness of the need for businesses to be inspired by the artistic environment. The research proved that there is still a need to popularize knowledge about this technology which is crucial for many sectors. Art projects are tools to develop knowledge and awareness of society and also various sectors. At the same time, artists may benefit from such collaboration. The paper will include selected aspects of mutual relations, areas of possible inspiration, and possible transfers of technological solutions. Those are AI applications in creative industries such as advertising and film, image recognition in art, and projects from different sectors.

Keywords: artificial intelligence, business, art, creative industry, technology

Procedia PDF Downloads 76
1165 Starchy Wastewater as Raw Material for Biohydrogen Production by Dark Fermentation: A Review

Authors: Tami A. Ulhiza, Noor I. M. Puad, Azlin S. Azmi, Mohd. I. A. Malek

Abstract:

High amount of chemical oxygen demand (COD) in starchy waste can be harmful to the environment. In common practice, starch processing wastewater is discharged to the river without proper treatment. However, starchy waste still contains complex sugars and organic acids. By the right pretreatment method, the complex sugar can be hydrolyzed into more readily digestible sugars which can be utilized to be converted into more valuable products. At the same time, the global demand of energy is inevitable. The continuous usage of fossil fuel as the main source of energy can lead to energy scarcity. Hydrogen is a renewable form of energy which can be an alternative energy in the future. Moreover, hydrogen is clean and carries the highest energy compared to other fuels. Biohydrogen produced from waste has significant advantages over chemical methods. One of the major problems in biohydrogen production is the raw material cost. The carbohydrate-rich starchy wastes such as tapioca, maize, wheat, potato, and sago wastes is a promising candidate to be used as a substrate in producing biohydrogen. The utilization of those wastes for biohydrogen production can provide cheap energy generation with simultaneous waste treatment. Therefore this paper aims to review variety source of starchy wastes that has been widely used to synthesize biohydrogen. The scope includes the source of waste, the performance in yielding hydrogen, the pretreatment method and the type of culture that is suitable for starchy waste.

Keywords: biohydrogen, dark fermentation, renewable energy, starchy waste

Procedia PDF Downloads 210
1164 Comparative Study on Sensory Profiles of Liquor from Different Dried Cocoa Beans

Authors: Khairul Bariah Sulaiman, Tajul Aris Yang

Abstract:

Malaysian dried cocoa beans have been reported to have low quality flavour and are often sold at discounted prices. Various efforts have been made to improve the Malaysian beans quality. Among these efforts is introduction of the shallow box fermentation technique and pulp preconditioned through pods storage. However, after nearly four decades of the effort was done, Malaysian cocoa farmers still received lower prices for their beans. So, this study was carried out in order to assess the flavour quality of dried cocoa beans produced by shallow box fermentation techniques, combination of shallow box fermentation with pods storage and compared to dried cocoa beans obtained from Ghana. A total of eight samples of dried cocoa was used in this study, which one of the samples was Ghanaian beans (coded with no.8), while the rest were Malaysian cocoa beans with different post-harvest processing (coded with no. 1, 2, 3, 4, 5, 6 and 7). Cocoa liquor was prepared from all samples in the prescribed techniques and sensory evaluation was carried out using Quantitative Descriptive Analysis (QDA) Method with 0-10 scale by Malaysian Cocoa Board trained panelist. Sensory evaluation showed that cocoa attributes for all cocoa liquors ranging from 3.5 to 5.3, whereas bitterness was ranging from 3.4 to 4.6 and astringent attribute ranging from 3.9 to 5.5, respectively. Meanwhile, all cocoa liquors were having acid or sourness attribute ranging from 1.6 to 3.6, respectively. In general cocoa liquor prepared from sample coded no 4 has almost similar flavour profile and no significantly different at p < 0.05 with Ghana, in term of most flavour attributes as compared to the other six samples.

Keywords: cocoa beans, flavour, fermentation, shallow box, pods storage

Procedia PDF Downloads 369
1163 Functional Yoghurt Enriched with Microencapsulated Olive Leaves Extract Powder Using Polycaprolactone via Double Emulsion/Solvent Evaporation Technique

Authors: Tamer El-Messery, Teresa Sanchez-Moya, Ruben Lopez-Nicolas, Gaspar Ros, Esmat Aly

Abstract:

Olive leaves (OLs), the main by-product of the olive oil industry, have a considerable amount of phenolic compounds. The exploitation of these compounds represents the current trend in food processing. In this study, OLs polyphenols were microencapsulated with polycaprolactone (PCL) and utilized in formulating novel functional yoghurt. PCL-microcapsules were characterized by scanning electron microscopy, and Fourier transform infrared spectrometry analysis. Their total phenolic (TPC), total flavonoid (TFC) contents, and antioxidant activities (DPPH, FRAP, ABTS), and polyphenols bioaccessibility were measured after oral, gastric, and intestinal steps of in vitro digestion. The four yoghurt formulations (containing 0, 25, 50, and 75 mg of PCL-microsphere/100g yoghurt) were evaluated for their pH, acidity, syneresis viscosity, and color during storage. In vitro digestion significantly affected the phenolic composition in non-encapsulated extract while had a lower impact on encapsulated phenolics. Higher protection was provided for encapsulated OLs extract, and their higher release was observed at the intestinal phase. Yoghurt with PCL-microsphere had lower viscosity, syneresis, and color parameters, as compared to control yoghurt. Thus, OLs represent a valuable and cheap source of polyphenols which can be successfully applied, in microencapsulated form, to formulate functional yoghurt.

Keywords: yoghurt quality attributes, olive leaves, phenolic and flavonoids compounds, antioxidant activity, polycaprolactone as microencapsulant

Procedia PDF Downloads 123
1162 Light-Weight Network for Real-Time Pose Estimation

Authors: Jianghao Hu, Hongyu Wang

Abstract:

The effective and efficient human pose estimation algorithm is an important task for real-time human pose estimation on mobile devices. This paper proposes a light-weight human key points detection algorithm, Light-Weight Network for Real-Time Pose Estimation (LWPE). LWPE uses light-weight backbone network and depthwise separable convolutions to reduce parameters and lower latency. LWPE uses the feature pyramid network (FPN) to fuse the high-resolution, semantically weak features with the low-resolution, semantically strong features. In the meantime, with multi-scale prediction, the predicted result by the low-resolution feature map is stacked to the adjacent higher-resolution feature map to intermediately monitor the network and continuously refine the results. At the last step, the key point coordinates predicted in the highest-resolution are used as the final output of the network. For the key-points that are difficult to predict, LWPE adopts the online hard key points mining strategy to focus on the key points that hard predicting. The proposed algorithm achieves excellent performance in the single-person dataset selected in the AI (artificial intelligence) challenge dataset. The algorithm maintains high-precision performance even though the model only contains 3.9M parameters, and it can run at 225 frames per second (FPS) on the generic graphics processing unit (GPU).

Keywords: depthwise separable convolutions, feature pyramid network, human pose estimation, light-weight backbone

Procedia PDF Downloads 138
1161 Business Continuity Risk Review for a Large Petrochemical Complex

Authors: Michel A. Thomet

Abstract:

A discrete-event simulation model was used to perform a Reliability-Availability-Maintainability (RAM) study of a large petrochemical complex which included sixteen process units, and seven feeds and intermediate streams. All the feeds and intermediate streams have associated storage tanks, so that if a processing unit fails and shuts down, the downstream units can keep producing their outputs. This also helps the upstream units which do not have to reduce their outputs, but can store their excess production until the failed unit restart. Each process unit and each pipe section carrying the feeds and intermediate streams has a probability of failure with an associated distribution and a Mean Time Between Failure (MTBF), as well as a distribution of the time to restore and a Mean Time To Restore (MTTR). The utilities supporting the process units can also fail and have their own distributions with specific MTBF and MTTR. The model runs are for ten years or more and the runs are repeated several times to obtain statistically relevant results. One of the main results is the On-Stream factor (OSF) of each process unit (percent of hours in a year when the unit is running in nominal conditions). One of the objectives of the study was to investigate if the storage capacity of each of the feeds and the intermediate stream was adequate. This was done by increasing the storage capacities in several steps and through running the simulation to see if the OSF were improved and by how much. Other objectives were to see if the failure of the utilities were an important factor in the overall OSF, and what could be done to reduce their failure rates through redundant equipment.

Keywords: business continuity, on-stream factor, petrochemical, RAM study, simulation, MTBF

Procedia PDF Downloads 197
1160 Space Time Adaptive Algorithm in Bi-Static Passive Radar Systems for Clutter Mitigation

Authors: D. Venu, N. V. Koteswara Rao

Abstract:

Space – time adaptive processing (STAP) is an effective tool for detecting a moving target in spaceborne or airborne radar systems. Since airborne passive radar systems utilize broadcast, navigation and excellent communication signals to perform various surveillance tasks and also has attracted significant interest from the distinct past, therefore the need of the hour is to have cost effective systems as compared to conventional active radar systems. Moreover, requirements of small number of secondary samples for effective clutter suppression in bi-static passive radar offer abundant illuminator resources for passive surveillance radar systems. This paper presents a framework for incorporating knowledge sources directly in the space-time beam former of airborne adaptive radars. STAP algorithm for clutter mitigation for passive bi-static radar has better quantitation of the reduction in sample size thereby amalgamating the earlier data bank with existing radar data sets. Also, we proposed a novel method to estimate the clutter matrix and perform STAP for efficient clutter suppression based on small sample size. Furthermore, the effectiveness of the proposed algorithm is verified using MATLAB simulations in order to validate STAP algorithm for passive bi-static radar. In conclusion, this study highlights the importance for various applications which augments traditional active radars using cost-effective measures.

Keywords: bistatic radar, clutter, covariance matrix passive radar, STAP

Procedia PDF Downloads 280
1159 Developing an Out-of-Distribution Generalization Model Selection Framework through Impurity and Randomness Measurements and a Bias Index

Authors: Todd Zhou, Mikhail Yurochkin

Abstract:

Out-of-distribution (OOD) detection is receiving increasing amounts of attention in the machine learning research community, boosted by recent technologies, such as autonomous driving and image processing. This newly-burgeoning field has called for the need for more effective and efficient methods for out-of-distribution generalization methods. Without accessing the label information, deploying machine learning models to out-of-distribution domains becomes extremely challenging since it is impossible to evaluate model performance on unseen domains. To tackle this out-of-distribution detection difficulty, we designed a model selection pipeline algorithm and developed a model selection framework with different impurity and randomness measurements to evaluate and choose the best-performing models for out-of-distribution data. By exploring different randomness scores based on predicted probabilities, we adopted the out-of-distribution entropy and developed a custom-designed score, ”CombinedScore,” as the evaluation criterion. This proposed score was created by adding labeled source information into the judging space of the uncertainty entropy score using harmonic mean. Furthermore, the prediction bias was explored through the equality of opportunity violation measurement. We also improved machine learning model performance through model calibration. The effectiveness of the framework with the proposed evaluation criteria was validated on the Folktables American Community Survey (ACS) datasets.

Keywords: model selection, domain generalization, model fairness, randomness measurements, bias index

Procedia PDF Downloads 108
1158 Bottleneck Modeling in Information Technology Service Management

Authors: Abhinay Puvvala, Veerendra Kumar Rai

Abstract:

A bottleneck situation arises when the outflow is lesser than the inflow in a pipe-like setup. A more practical interpretation of bottlenecks emphasizes on the realization of Service Level Objectives (SLOs) at given workloads. Our approach detects two key aspects of bottlenecks – when and where. To identify ‘when’ we continuously poll on certain key metrics such as resource utilization, processing time, request backlog and throughput at a system level. Further, when the slope of the expected sojourn time at a workload is greater than ‘K’ times the slope of expected sojourn time at the previous step of the workload while the workload is being gradually increased in discrete steps, a bottleneck situation arises. ‘K’ defines the threshold condition and is computed based on the system’s service level objectives. The second aspect of our approach is to identify the location of the bottleneck. In multi-tier systems with a complex network of layers, it is a challenging problem to locate bottleneck that affects the overall system performance. We stage the system by varying workload incrementally to draw a correlation between load increase and system performance to the point where Service Level Objectives are violated. During the staging process, multiple metrics are monitored at hardware and application levels. The correlations are drawn between metrics and the overall system performance. These correlations along with the Service Level Objectives are used to arrive at the threshold conditions for each of these metrics. Subsequently, the same method used to identify when a bottleneck occurs is used on metrics data with threshold conditions to locate bottlenecks.

Keywords: bottleneck, workload, service level objectives (SLOs), throughput, system performance

Procedia PDF Downloads 214
1157 Visual Inspection of Road Conditions Using Deep Convolutional Neural Networks

Authors: Christos Theoharatos, Dimitris Tsourounis, Spiros Oikonomou, Andreas Makedonas

Abstract:

This paper focuses on the problem of visually inspecting and recognizing the road conditions in front of moving vehicles, targeting automotive scenarios. The goal of road inspection is to identify whether the road is slippery or not, as well as to detect possible anomalies on the road surface like potholes or body bumps/humps. Our work is based on an artificial intelligence methodology for real-time monitoring of road conditions in autonomous driving scenarios, using state-of-the-art deep convolutional neural network (CNN) techniques. Initially, the road and ego lane are segmented within the field of view of the camera that is integrated into the front part of the vehicle. A novel classification CNN is utilized to identify among plain and slippery road textures (e.g., wet, snow, etc.). Simultaneously, a robust detection CNN identifies severe surface anomalies within the ego lane, such as potholes and speed bumps/humps, within a distance of 5 to 25 meters. The overall methodology is illustrated under the scope of an integrated application (or system), which can be integrated into complete Advanced Driver-Assistance Systems (ADAS) systems that provide a full range of functionalities. The outcome of the proposed techniques present state-of-the-art detection and classification results and real-time performance running on AI accelerator devices like Intel’s Myriad 2/X Vision Processing Unit (VPU).

Keywords: deep learning, convolutional neural networks, road condition classification, embedded systems

Procedia PDF Downloads 110
1156 Development of Latent Fingerprints on Non-Porous Surfaces Recovered from Fresh and Sea Water

Authors: A. Somaya Madkour, B. Abeer sheta, C. Fatma Badr El Dine, D. Yasser Elwakeel, E. Nermine AbdAllah

Abstract:

Criminal offenders have a fundamental goal not to leave any traces at the crime scene. Some may suppose that items recovered underwater will have no forensic value, therefore, they try to destroy the traces by throwing items in water. These traces are subjected to the destructive environmental effects. This can represent a challenge for Forensic experts investigating finger marks. Accordingly, the present study was conducted to determine the optimal method for latent fingerprints development on non-porous surfaces submerged in aquatic environments at different time interval. The two factors analyzed in this study were the nature of aquatic environment and length of submerged time. In addition, the quality of developed finger marks depending on the used method was also assessed. Therefore, latent fingerprints were deposited on metallic, plastic and glass objects and submerged in fresh or sea water for one, two, and ten days. After recovery, the items were subjected to cyanoacrylate fuming, black powder and small particle reagent processing and the prints were examined. Each print was evaluated according to fingerprint quality assessment scale. The present study demonstrated that the duration of submersion affects the quality of finger marks; the longer the duration, the worse the quality.The best results of visualization were achieved using cyanoacrylate either in fresh or sea water. This study has also revealed that the exposure to sea water had more destructive influence on the quality of detected finger marks.

Keywords: fingerprints, fresh water, sea, non-porous

Procedia PDF Downloads 436
1155 Review of Microstructure, Mechanical and Corrosion Behavior of Aluminum Matrix Composite Reinforced with Agro/Industrial Waste Fabricated by Stir Casting Process

Authors: Mehari Kahsay, Krishna Murthy Kyathegowda, Temesgen Berhanu

Abstract:

Aluminum matrix composites have gained focus on research and industrial use, especially those not requiring extreme loading or thermal conditions, for the last few decades. Their relatively low cost, simple processing and attractive properties are the reasons for the widespread use of aluminum matrix composites in the manufacturing of automobiles, aircraft, military, and sports goods. In this article, the microstructure, mechanical, and corrosion behaviors of the aluminum metal matrix were reviewed, focusing on the stir casting fabrication process and usage of agro/industrial waste reinforcement particles. The results portrayed that mechanical properties like tensile strength, ultimate tensile strength, hardness, percentage of elongation, impact, and fracture toughness are highly dependent on the amount, kind, and size of reinforcing particles. Additionally, uniform distribution, wettability of reinforcement particles, and the porosity level of the resulting composite also affect the mechanical and corrosion behaviors of aluminum matrix composites. The two-step stir-casting process resulted in better wetting characteristics, a lower porosity level, and a uniform distribution of particles with proper handling of process parameters. On the other hand, the inconsistent and contradicting results on corrosion behavior regarding monolithic and hybrid aluminum matrix composites need further study.

Keywords: microstructure, mechanical behavior, corrosion, aluminum matrix composite

Procedia PDF Downloads 49
1154 Artificial Neural Network in Ultra-High Precision Grinding of Borosilicate-Crown Glass

Authors: Goodness Onwuka, Khaled Abou-El-Hossein

Abstract:

Borosilicate-crown (BK7) glass has found broad application in the optic and automotive industries and the growing demands for nanometric surface finishes is becoming a necessity in such applications. Thus, it has become paramount to optimize the parameters influencing the surface roughness of this precision lens. The research was carried out on a 4-axes Nanoform 250 precision lathe machine with an ultra-high precision grinding spindle. The experiment varied the machining parameters of feed rate, wheel speed and depth of cut at three levels for different combinations using Box Behnken design of experiment and the resulting surface roughness values were measured using a Taylor Hobson Dimension XL optical profiler. Acoustic emission monitoring technique was applied at a high sampling rate to monitor the machining process while further signal processing and feature extraction methods were implemented to generate the input to a neural network algorithm. This paper highlights the training and development of a back propagation neural network prediction algorithm through careful selection of parameters and the result show a better classification accuracy when compared to a previously developed response surface model with very similar machining parameters. Hence artificial neural network algorithms provide better surface roughness prediction accuracy in the ultra-high precision grinding of BK7 glass.

Keywords: acoustic emission technique, artificial neural network, surface roughness, ultra-high precision grinding

Procedia PDF Downloads 290
1153 Internet of Assets: A Blockchain-Inspired Academic Program

Authors: Benjamin Arazi

Abstract:

Blockchain is the technology behind cryptocurrencies like Bitcoin. It revolutionizes the meaning of trust in the sense of offering total reliability without relying on any central entity that controls or supervises the system. The Wall Street Journal states: “Blockchain Marks the Next Step in the Internet’s Evolution”. Blockchain was listed as #1 in Linkedin – The Learning Blog “most in-demand hard skills needed in 2020”. As stated there: “Blockchain’s novel way to store, validate, authorize, and move data across the internet has evolved to securely store and send any digital asset”. GSMA, a leading Telco organization of mobile communications operators, declared that “Blockchain has the potential to be for value what the Internet has been for information”. Motivated by these seminal observations, this paper presents the foundations of a Blockchain-based “Internet of Assets” academic program that joins under one roof leading application areas that are characterized by the transfer of assets over communication lines. Two such areas, which are pillars of our economy, are Fintech – Financial Technology and mobile communications services. The next application in line is Healthcare. These challenges are met based on available extensive professional literature. Blockchain-based assets communication is based on extending the principle of Bitcoin, starting with the basic question: If digital money that travels across the universe can ‘prove its own validity’, can this principle be applied to digital content. A groundbreaking positive answer here led to the concept of “smart contract” and consequently to DLT - Distributed Ledger Technology, where the word ‘distributed’ relates to the non-existence of reliable central entities or trusted third parties. The terms Blockchain and DLT are frequently used interchangeably in various application areas. The World Bank Group compiled comprehensive reports, analyzing the contribution of DLT/Blockchain to Fintech. The European Central Bank and Bank of Japan are engaged in Project Stella, “Balancing confidentiality and auditability in a distributed ledger environment”. 130 DLT/Blockchain focused Fintech startups are now operating in Switzerland. Blockchain impact on mobile communications services is treated in detail by leading organizations. The TM Forum is a global industry association in the telecom industry, with over 850 member companies, mainly mobile operators, that generate US$2 trillion in revenue and serve five billion customers across 180 countries. From their perspective: “Blockchain is considered one of the digital economy’s most disruptive technologies”. Samples of Blockchain contributions to Fintech (taken from a World Bank document): Decentralization and disintermediation; Greater transparency and easier auditability; Automation & programmability; Immutability & verifiability; Gains in speed and efficiency; Cost reductions; Enhanced cyber security resilience. Samples of Blockchain contributions to the Telco industry. Establishing identity verification; Record of transactions for easy cost settlement; Automatic triggering of roaming contract which enables near-instantaneous charging and reduction in roaming fraud; Decentralized roaming agreements; Settling accounts per costs incurred in accordance with agreement tariffs. This clearly demonstrates an academic education structure where fundamental technologies are studied in classes together with these two application areas. Advanced courses, treating specific implementations then follow separately. All are under the roof of “Internet of Assets”.

Keywords: blockchain, education, financial technology, mobile telecommunications services

Procedia PDF Downloads 162
1152 Criminal Attitude vs Transparency in the Arab World

Authors: Keroles Akram Saed Ghatas

Abstract:

The political violence that characterized 1992 continued into 1993, creating a major security crisis for President Hosni Mubarak's government as the death toll and human rights abuses soared. Increasingly sensitive to criticism of 's human rights activities, the government established human rights departments in key ministries, beginning with the Foreign Office in February. Similar offices have been set up in the Justice and Agriculture Ministries, and plans to set up an office in the Home Office have been announced. It turned out that the main task of the law unit was to overturn the conclusions of international human rights organizations.President Mubarak was elected in a national referendum on October 4 for a third six-year term after being appointed on July 21 by the People's Assembly, an elected parliament overwhelmingly dominated by the in-power National Democratic Party will Mr. Mubarak ran unhindered. The Interior Ministry announced that nearly 16 million people cast their votes (84% of eligible voters), of which 96.28%. voted for presidential re-election.In 1993, armed Islamic extremists escalated their attacks on Christian citizens, government officials, police officers and senior security officials, resulting in casualties among the intended victims and bystanders. Sporadic attacks on buses, boats and tourist attractions also occurred throughout the year. From March 1992 to October 28, 1993, a total of 222 people lost their lives in the riots: 36 Coptic Christians and 38 other citizens; If one is a foreigner; sixty-six members of the Security Forces; and seventy-six known or suspected activists who were killed while resisting arrest. The latter was killed in airstrikes and firefights with security forces and at the site of planned attacks. On March 9-10, a series of airstrikes in Cairo, Giza, Qalyubiya province north of the capital and Aswan killed fifteen suspected militants and five members of the security forces.One of the airstrikes in Giza, part of Greater Cairo, killed the wife and son of Khalifa Mahmoud Ramadan, a suspected militant who was himself killed. The government agency Middle East News Agency reported on March 10 that the raids were part of a "broad confrontational plan aimed at ofterrorist elements"The state of emergency declared in October 1981 after the assassination of President Anwar el-Sadat was still in force in Egypt. The law, previously in effect continuously from June 1967 to May 1980, continued to grant the executive branch unique legal powers that effectively overrode the human rights guarantees of the Egyptian constitution. These provisions included wide discretionary powers in arresting and detaining individuals, as well as the ability to try civilians in military courts. The Cairo-based Independent Organization for Human Rights said so in a document sent to the United Nations in July 1993The human rights committee said the continued imposition of the state of emergency had resulted in "another constitution for the country" and "led to widespread misconduct by the security apparatus".

Keywords: constitution, human rights, legal power, president, anwar, el-sadat, assassination, state of emergency, middle east, news, agency, confrontational, arresting, fugitive, leaders, terrorist, elements, armed islamic extremists.

Procedia PDF Downloads 17
1151 A Recommender System for Job Seekers to Show up Companies Based on Their Psychometric Preferences and Company Sentiment Scores

Authors: A. Ashraff

Abstract:

The increasing importance of the web as a medium for electronic and business transactions has served as a catalyst or rather a driving force for the introduction and implementation of recommender systems. Recommender Systems play a major role in processing and analyzing thousands of data rows or reviews and help humans make a purchase decision of a product or service. It also has the ability to predict whether a particular user would rate a product or service based on the user’s profile behavioral pattern. At present, Recommender Systems are being used extensively in every domain known to us. They are said to be ubiquitous. However, in the field of recruitment, it’s not being utilized exclusively. Recent statistics show an increase in staff turnover, which has negatively impacted the organization as well as the employee. The reasons being company culture, working flexibility (work from home opportunity), no learning advancements, and pay scale. Further investigations revealed that there are lacking guidance or support, which helps a job seeker find the company that will suit him best, and though there’s information available about companies, job seekers can’t read all the reviews by themselves and get an analytical decision. In this paper, we propose an approach to study the available review data on IT companies (score their reviews based on user review sentiments) and gather information on job seekers, which includes their Psychometric evaluations. Then presents the job seeker with useful information or rather outputs on which company is most suitable for the job seeker. The theoretical approach, Algorithmic approach and the importance of such a system will be discussed in this paper.

Keywords: psychometric tests, recommender systems, sentiment analysis, hybrid recommender systems

Procedia PDF Downloads 89
1150 Iterative Segmentation and Application of Hausdorff Dilation Distance in Defect Detection

Authors: S. Shankar Bharathi

Abstract:

Inspection of surface defects on metallic components has always been challenging due to its specular property. Occurrences of defects such as scratches, rust, pitting are very common in metallic surfaces during the manufacturing process. These defects if unchecked can hamper the performance and reduce the life time of such component. Many of the conventional image processing algorithms in detecting the surface defects generally involve segmentation techniques, based on thresholding, edge detection, watershed segmentation and textural segmentation. They later employ other suitable algorithms based on morphology, region growing, shape analysis, neural networks for classification purpose. In this paper the work has been focused only towards detecting scratches. Global and other thresholding techniques were used to extract the defects, but it proved to be inaccurate in extracting the defects alone. However, this paper does not focus on comparison of different segmentation techniques, but rather describes a novel approach towards segmentation combined with hausdorff dilation distance. The proposed algorithm is based on the distribution of the intensity levels, that is, whether a certain gray level is concentrated or evenly distributed. The algorithm is based on extraction of such concentrated pixels. Defective images showed higher level of concentration of some gray level, whereas in non-defective image, there seemed to be no concentration, but were evenly distributed. This formed the basis in detecting the defects in the proposed algorithm. Hausdorff dilation distance based on mathematical morphology was used to strengthen the segmentation of the defects.

Keywords: metallic surface, scratches, segmentation, hausdorff dilation distance, machine vision

Procedia PDF Downloads 404
1149 Recycled Use of Solid Wastes in Building Material: A Review

Authors: Oriyomi M. Okeyinka, David A. Oloke, Jamal M. Khatib

Abstract:

Large quantities of solid wastes being generated worldwide from sources such as household, domestic, industrial, commercial and construction demolition activities, leads to environmental concerns. Utilization of these wastes in making building construction materials can reduce the magnitude of the associated problems. When these waste products are used in place of other conventional materials, natural resources and energy are preserved and expensive and/or potentially harmful waste disposal is avoided. Recycling which is regarded as the third most preferred waste disposal option, with its numerous environmental benefits, stand as a viable option to offset the environmental impact associated with the construction industry. This paper reviews the results of laboratory tests and important research findings, and the potential of using these wastes in building construction materials with focus on sustainable development. Research gaps, which includes; the need to develop standard mix design for solid waste based building materials; the need to develop energy efficient method of processing solid waste use in concrete; the need to study the actual behavior or performance of such building materials in practical application and the limited real life application of such building materials have also been identified. Therefore a research is being proposed to develop an environmentally friendly, lightweight building block from recycled waste paper, without the use of cement, and with properties suitable for use as walling unit. This proposed research intends to incorporate, laboratory experimentation and modeling to address the identified research gaps.

Keywords: recycling, solid wastes, construction, building materials

Procedia PDF Downloads 366
1148 Vehicle Gearbox Fault Diagnosis Based on Cepstrum Analysis

Authors: Mohamed El Morsy, Gabriela Achtenová

Abstract:

Research on damage of gears and gear pairs using vibration signals remains very attractive, because vibration signals from a gear pair are complex in nature and not easy to interpret. Predicting gear pair defects by analyzing changes in vibration signal of gears pairs in operation is a very reliable method. Therefore, a suitable vibration signal processing technique is necessary to extract defect information generally obscured by the noise from dynamic factors of other gear pairs. This article presents the value of cepstrum analysis in vehicle gearbox fault diagnosis. Cepstrum represents the overall power content of a whole family of harmonics and sidebands when more than one family of sidebands is present at the same time. The concept for the measurement and analysis involved in using the technique are briefly outlined. Cepstrum analysis is used for detection of an artificial pitting defect in a vehicle gearbox loaded with different speeds and torques. The test stand is equipped with three dynamometers; the input dynamometer serves as the internal combustion engine, the output dynamometers introduce the load on the flanges of the output joint shafts. The pitting defect is manufactured on the tooth side of a gear of the fifth speed on the secondary shaft. Also, a method for fault diagnosis of gear faults is presented based on order cepstrum. The procedure is illustrated with the experimental vibration data of the vehicle gearbox. The results show the effectiveness of cepstrum analysis in detection and diagnosis of the gear condition.

Keywords: cepstrum analysis, fault diagnosis, gearbox, vibration signals

Procedia PDF Downloads 361
1147 Engineering Strategies Towards Improvement in Energy Storage Performance of Ceramic Capacitors for Pulsed Power Applications

Authors: Abdul Manan

Abstract:

The necessity for efficient and cost-effective energy storage devices to intelligently store the inconsistent energy output from modern renewable energy sources is peaked today. The scientific community is struggling to identify the appropriate material system for energy storage applications. Countless contributions by researchers worldwide have now helped us identify the possible snags and limitations associated with each material/method. Energy storage has attracted great attention for its use in portable electronic devices military field. Different devices, such as dielectric capacitors, supercapacitors, and batteries, are used for energy storage. Of these, dielectric capacitors have high energy output, a long life cycle, fast charging and discharging capabilities, work at high temperatures, and excellent fatigue resistance. The energy storage characteristics have been studied to be highly affected by various factors, such as grain size, optimized compositions, grain orientation, energy band gap, processing techniques, defect engineering, core-shell formation, interface engineering, electronegativity difference, the addition of additives, density, secondary phases, the difference of Pmax-Pr, sample thickness, area of the electrode, testing frequency, and AC/DC conditions. The data regarding these parameters/factors are scattered in the literature, and the aim of this study is to gather the data into a single paper that will be beneficial for new researchers in the field of interest. Furthermore, control over and optimizing these parameters will lead to enhancing the energy storage properties.

Keywords: strategies, ceramics, energy storage, capacitors

Procedia PDF Downloads 61
1146 Effect of Manganese Doping on Ferrroelectric Properties of (K0.485Na0.5Li0.015)(Nb0.98V0.02)O3 Lead-Free Piezoceramic

Authors: Chongtham Jiten, Radhapiyari Laishram, K. Chandramani Singh

Abstract:

Alkaline niobate (Na0.5K0.5)NbO3 ceramic system has attracted major attention in view of its potential for replacing the highly toxic but superior lead zirconate titanate (PZT) system for piezoelectric applications. Recently, a more detailed study of this system reveals that the ferroelectric and piezoelectric properties are optimized in the Li- and V-modified system having the composition (K0.485Na0.5Li0.015)(Nb0.98V0.02)O3. In the present work, we further study the pyroelectric behaviour of this composition along with another doped with Mn4+. So, (K0.485Na0.5Li0.015)(Nb0.98V0.02)O3 + x MnO2 (x = 0, and 0.01 wt. %) ceramic compositions were synthesized by conventional ceramic processing route. X-ray diffraction study reveals that both the undoped and Mn4+-doped ceramic samples prepared crystallize into a perovskite structure having orthorhombic symmetry. Dielectric study indicates that Mn4+ doping has little effect on both the Curie temperature (Tc) and tetragonal-orthorhombic phase transition temperature (Tot). The bulk density, room-temperature dielectric constant (εRT), and room-c The room-temperature coercive field (Ec) is observed to be lower in Mn4+ doped sample. The detailed analysis of the P-E hysteresis loops over the range of temperature from about room temperature to Tot points out that enhanced ferroelectric properties exist in this temperature range with better thermal stability for the Mn4+ doped ceramic. The study reveals that small traces of Mn4+ can modify (K0.485Na0.5Li0.015)(Nb0.98V0.02)O3 system so as to improve its ferroelectric properties with good thermal stability over a wide range of temperature.

Keywords: ceramics, dielectric properties, ferroelectric properties, lead-free, sintering, thermal stability

Procedia PDF Downloads 224
1145 Encapsulation and Protection of Bioactive Nutrients Based on Ligand-Binding Property of Milk Proteins

Authors: Hao Cheng, Yingzhou Ni, Amr M. Bakry, Li Liang

Abstract:

Functional foods containing bioactive nutrients offer benefits beyond basic nutrition and hence the possibility of delaying and preventing chronic diseases. However, many bioactive nutrients degrade rapidly under food processing and storage conditions. Encapsulation can be used to overcome these limitations. Food proteins have been widely used as carrier materials for the preparation of nano/micro-particles because of their ability to form gels and emulsions and to interact with polysaccharides. The mechanisms of interaction between bioactive nutrients and proteins must be understood in order to develop protein-based lipid-free delivery systems. Beta-lactoglobulin, a small globular protein in milk whey, exhibits an affinity to a wide range of compounds. Alfa-tocopherol, resveratrol and folic acid were respectively bound to the central cavity, the outer surface near Trp19–Arg124 and the hydrophobic pocket in the groove between the alfa-helix and the beta-barrel of the protein. Beta-lactoglobulin could thus bind the three bioactive nutrients simultaneously to form protein-multi-ligand complexes. Beta-casein, an intrinsically unstructured but major milk protein, could also interact with resveratrol and folic acid to form complexes. These results suggest the potential to develop milk-protein-based complex carrier systems for encapsulation of multiple bioactive nutrients for functional food application and also pharmaceutical and medical uses.

Keywords: milk protein, bioactive nutrient, interaction, protection

Procedia PDF Downloads 391
1144 Rational Memory Therapy: The Counselling Technique to Control Psychological and Psychosomatic Illnesses

Authors: Sachin Deshmukh

Abstract:

Mind and body synchronization occurs through memory and sensation production. Sensations are the guiding language of subconscious mind for conscious mind to take a proper action. Mind-mechanism is based upon memories collected so far since intrauterine life. There are three universal triggers for memory creation; they are persons, situations and objects. Memory is created as sensations experienced by special senses. Based upon experiencing comfort or discomfort, the triggers are categorized as safe or unsafe triggers. A memory comprises of ‘safe or unsafe feeling for triggers, and actions taken for that feeling’. Memories for triggers are created slowly, thoughtfully and consciously by the conscious mind, and archived in the subconscious mind for future references. Later on, similar triggers can come in contact with the individual. Subconscious mind uses these stored feelings to decide whether these triggers are safe or unsafe. It produces comfort or discomfort sensations as emotions accordingly and reacts in the same way as has been recorded in memory. Speed of sensing and processing the triggers, and reacting by subconscious mind is that of the speed of bioelectricity. Hence, formula for human emotions has been designed in this paper as follows: Emotion (Stress or Peace) = Trigger (Person or Situation or object) x Mass of feelings (stressful or peaceful) associated with the Trigger x Speed of Light². We also establish modern medical scientific facts about relationship between reflex activity and memory. This research further develops the ‘Rational Memory Therapy’ focusing on therapeutic feelings conversion techniques, for stress prevention and management.

Keywords: memory, sensations, feelings, emotions, rational memory therapy

Procedia PDF Downloads 237
1143 Multimodal Optimization of Density-Based Clustering Using Collective Animal Behavior Algorithm

Authors: Kristian Bautista, Ruben A. Idoy

Abstract:

A bio-inspired metaheuristic algorithm inspired by the theory of collective animal behavior (CAB) was integrated to density-based clustering modeled as multimodal optimization problem. The algorithm was tested on synthetic, Iris, Glass, Pima and Thyroid data sets in order to measure its effectiveness relative to CDE-based Clustering algorithm. Upon preliminary testing, it was found out that one of the parameter settings used was ineffective in performing clustering when applied to the algorithm prompting the researcher to do an investigation. It was revealed that fine tuning distance δ3 that determines the extent to which a given data point will be clustered helped improve the quality of cluster output. Even though the modification of distance δ3 significantly improved the solution quality and cluster output of the algorithm, results suggest that there is no difference between the population mean of the solutions obtained using the original and modified parameter setting for all data sets. This implies that using either the original or modified parameter setting will not have any effect towards obtaining the best global and local animal positions. Results also suggest that CDE-based clustering algorithm is better than CAB-density clustering algorithm for all data sets. Nevertheless, CAB-density clustering algorithm is still a good clustering algorithm because it has correctly identified the number of classes of some data sets more frequently in a thirty trial run with a much smaller standard deviation, a potential in clustering high dimensional data sets. Thus, the researcher recommends further investigation in the post-processing stage of the algorithm.

Keywords: clustering, metaheuristics, collective animal behavior algorithm, density-based clustering, multimodal optimization

Procedia PDF Downloads 206
1142 Enhancer: An Effective Transformer Architecture for Single Image Super Resolution

Authors: Pitigalage Chamath Chandira Peiris

Abstract:

A widely researched domain in the field of image processing in recent times has been single image super-resolution, which tries to restore a high-resolution image from a single low-resolution image. Many more single image super-resolution efforts have been completed utilizing equally traditional and deep learning methodologies, as well as a variety of other methodologies. Deep learning-based super-resolution methods, in particular, have received significant interest. As of now, the most advanced image restoration approaches are based on convolutional neural networks; nevertheless, only a few efforts have been performed using Transformers, which have demonstrated excellent performance on high-level vision tasks. The effectiveness of CNN-based algorithms in image super-resolution has been impressive. However, these methods cannot completely capture the non-local features of the data. Enhancer is a simple yet powerful Transformer-based approach for enhancing the resolution of images. A method for single image super-resolution was developed in this study, which utilized an efficient and effective transformer design. This proposed architecture makes use of a locally enhanced window transformer block to alleviate the enormous computational load associated with non-overlapping window-based self-attention. Additionally, it incorporates depth-wise convolution in the feed-forward network to enhance its ability to capture local context. This study is assessed by comparing the results obtained for popular datasets to those obtained by other techniques in the domain.

Keywords: single image super resolution, computer vision, vision transformers, image restoration

Procedia PDF Downloads 87
1141 A Picture is worth a Billion Bits: Real-Time Image Reconstruction from Dense Binary Pixels

Authors: Tal Remez, Or Litany, Alex Bronstein

Abstract:

The pursuit of smaller pixel sizes at ever increasing resolution in digital image sensors is mainly driven by the stringent price and form-factor requirements of sensors and optics in the cellular phone market. Recently, Eric Fossum proposed a novel concept of an image sensor with dense sub-diffraction limit one-bit pixels (jots), which can be considered a digital emulation of silver halide photographic film. This idea has been recently embodied as the EPFL Gigavision camera. A major bottleneck in the design of such sensors is the image reconstruction process, producing a continuous high dynamic range image from oversampled binary measurements. The extreme quantization of the Poisson statistics is incompatible with the assumptions of most standard image processing and enhancement frameworks. The recently proposed maximum-likelihood (ML) approach addresses this difficulty, but suffers from image artifacts and has impractically high computational complexity. In this work, we study a variant of a sensor with binary threshold pixels and propose a reconstruction algorithm combining an ML data fitting term with a sparse synthesis prior. We also show an efficient hardware-friendly real-time approximation of this inverse operator. Promising results are shown on synthetic data as well as on HDR data emulated using multiple exposures of a regular CMOS sensor.

Keywords: binary pixels, maximum likelihood, neural networks, sparse coding

Procedia PDF Downloads 182
1140 Hybrid Advanced Oxidative Pretreatment of Complex Industrial Effluent for Biodegradability Enhancement

Authors: K. Paradkar, S. N. Mudliar, A. Sharma, A. B. Pandit, R. A. Pandey

Abstract:

The study explores the hybrid combination of Hydrodynamic Cavitation (HC) and Subcritical Wet Air Oxidation-based pretreatment of complex industrial effluent to enhance the biodegradability selectively (without major COD destruction) to facilitate subsequent enhanced downstream processing via anaerobic or aerobic biological treatment. Advanced oxidation based techniques can be less efficient as standalone options and a hybrid approach by combining Hydrodynamic Cavitation (HC), and Wet Air Oxidation (WAO) can lead to a synergistic effect since both the options are based on common free radical mechanism. The HC can be used for initial turbulence and generation of hotspots which can begin the free radical attack and this agitating mixture then can be subjected to less intense WAO since initial heat (to raise the activation energy) can be taken care by HC alone. Lab-scale venturi-based hydrodynamic cavitation and wet air oxidation reactor with biomethanated distillery wastewater (BMDWW) as a model effluent was examined for establishing the proof-of-concept. The results indicated that for a desirable biodegradability index (BOD: COD - BI) enhancement (up to 0.4), the Cavitation (standalone) pretreatment condition was: 5 bar and 88 min reaction time with a COD reduction of 36 % and BI enhancement of up to 0.27 (initial BI - 0.17). The optimum WAO condition (standalone) was: 150oC, 6 bar and 30 minutes with 31% COD reduction and 0.33 BI. The hybrid pretreatment (combined Cavitation + WAO) worked out to be 23.18 min HC (at 5 bar) followed by 30 min WAO at 150oC, 6 bar, at which around 50% COD was retained yielding a BI of 0.55. FTIR & NMR analysis of pretreated effluent indicated dissociation and/or reorientation of complex organic compounds in untreated effluent to simpler organic compounds post-pretreatment.

Keywords: hybrid, hydrodynamic cavitation, wet air oxidation, biodegradability index

Procedia PDF Downloads 595
1139 Ambivalence as Ethical Practice: Methodologies to Address Noise, Bias in Care, and Contact Evaluations

Authors: Anthony Townsend, Robyn Fasser

Abstract:

While complete objectivity is a desirable scientific position from which to conduct a care and contact evaluation (CCE), it is precisely the recognition that we are inherently incapable of operating objectively that is the foundation of ethical practice and skilled assessment. Drawing upon recent research from Daniel Kahneman (2021) on the differences between noise and bias, as well as different inherent biases collectively termed “The Elephant in the Brain” by Kevin Simler and Robin Hanson (2019) from Oxford University, this presentation addresses both the various ways in which our judgments, perceptions and even procedures can be distorted and contaminated while conducting a CCE, but also considers the value of second order cybernetics and the psychodynamic concept of ‘ambivalence’ as a conceptual basis to inform our assessment methodologies to limit such errors or at least better identify them. Both a conceptual framework for ambivalence, our higher-order capacity to allow for the convergence and consideration of multiple emotional experiences and cognitive perceptions to inform our reasoning, and a practical methodology for assessment relying on data triangulation, Bayesian inference and hypothesis testing is presented as a means of promoting ethical practice for health care professionals conducting CCEs. An emphasis on widening awareness and perspective, limiting ‘splitting’, is demonstrated both in how this form of emotional processing plays out in alienating dynamics in families as well as the assessment thereof. In addressing this concept, this presentation aims to illuminate the value of ambivalence as foundational to ethical practice for assessors.

Keywords: ambivalence, forensic, psychology, noise, bias, ethics

Procedia PDF Downloads 72
1138 Computer Countenanced Diagnosis of Skin Nodule Detection and Histogram Augmentation: Extracting System for Skin Cancer

Authors: S. Zith Dey Babu, S. Kour, S. Verma, C. Verma, V. Pathania, A. Agrawal, V. Chaudhary, A. Manoj Puthur, R. Goyal, A. Pal, T. Danti Dey, A. Kumar, K. Wadhwa, O. Ved

Abstract:

Background: Skin cancer is now is the buzzing button in the field of medical science. The cyst's pandemic is drastically calibrating the body and well-being of the global village. Methods: The extracted image of the skin tumor cannot be used in one way for diagnosis. The stored image contains anarchies like the center. This approach will locate the forepart of an extracted appearance of skin. Partitioning image models has been presented to sort out the disturbance in the picture. Results: After completing partitioning, feature extraction has been formed by using genetic algorithm and finally, classification can be performed between the trained and test data to evaluate a large scale of an image that helps the doctors for the right prediction. To bring the improvisation of the existing system, we have set our objectives with an analysis. The efficiency of the natural selection process and the enriching histogram is essential in that respect. To reduce the false-positive rate or output, GA is performed with its accuracy. Conclusions: The objective of this task is to bring improvisation of effectiveness. GA is accomplishing its task with perfection to bring down the invalid-positive rate or outcome. The paper's mergeable portion conflicts with the composition of deep learning and medical image processing, which provides superior accuracy. Proportional types of handling create the reusability without any errors.

Keywords: computer-aided system, detection, image segmentation, morphology

Procedia PDF Downloads 129