The prominence of AI algorithms has reached new heights in todays world. Algorithms are implemented in each and every aspect of our life with the primary intention of improving it. However, the working of most of these AIs is not entirely understood, this causes the problem. Since the models are not understood, the researchers do not know how to manually improve on it, creating an artificial ceiling. Understanding these algorithms is the key to realize how these machines think and not only learn from them but also teach it to understand better. The primary idea put forward in this paper is to explain a black-box model using a mimicking simulation, rather than the usual calculative approaches. The primary idea put forward is to understand how an AI works using a decision tree that will be designed to mimic the AI. It proposes the use of a Decision Tree based approach along with randomization using Monte Carlo simulations for a more precise simulation of the black-box.
In this information age, web crawling on the internet is a prime source for data collection. And with the surface web already being dominated by giants like Google and Microsoft, much attention has been on the Dark Web. While research on crawling approaches is generally available, a considerable gap is present for URL extraction on the dark web. With most literature using the regular expressions methodology or built-in parsers, the problem with these methods is the higher number of false positives generated with the Dark Web, which makes the crawler less efficient. This paper proposes the dedicated parsers methodology for extracting URLs from the dark web, which when compared proves to be better than the regular expression methodology. Factors that make link harvesting on the Dark Web a challenge are discussed in the paper.
2022
IGI
ML-enabled informed intervention for crowdsourcing-based optimization of medical resources
It is said that every adversity presents the opportunity to grow. The current pandemic is a lesson to all healthcare infrastructure stakeholders to look at existing setups with an open mind. This chapter’s proposed solution offers technology assistance to manage patient data effectively and extends the hospital data management system’s capability to predict the upcoming need for healthcare resources. Further, the authors intend to supplement the proposed solution with crowdsourcing to meet hospital demand and supply for unprecedented medical emergencies. The proposed approach would demonstrate its need in the current pandemic scenario and prepare the healthcare infrastructure with a more streamlined and cooperative approach than before.
Springer
Recognizing child unsafe apps through user reviews on the google play store
Google Play Store serves as a platform to host, download, and review android applications. Many researchers have explored the user review section and worked on approaches and solutions that would prove a more effective pipeline to enable developer feedback on application issues and praised features proving the section’s abundance of information. This work uses this same data to attempt a novel use case of determining child unsafe apps on Google Play Store. User reviews are collected using a crawler and categorized for selected keywords relating to child, media, and India. Since Google Play Store does not provide a definitive number of downloads, this work attempts to mitigate this challenge by instead calculating the user density for an application. The user density helps establish the engagement users have with an application and is calculated by the difference in the timestamps of the most and least recent reviews divided by the sum of total reviews and its upvotes for an application. 60,620 reviews from 1,600 applications were extracted to validate the proposed concept. This concept has proved effective in recognizing applications that present child unsafe content while also offering a novel concept of calculating user density.