2020
Autores
Guimaraes, N; Figueira, A; Torgo, L;
Publicação
Communications in Computer and Information Science
Abstract
The emergence of online social networks provided users with an easy way to publish and disseminate content, reaching broader audiences than previous platforms (such as blogs or personal websites) allowed. However, malicious users started to take advantage of these features to disseminate unreliable content through the network like false information, extremely biased opinions, or hate speech. Consequently, it becomes crucial to try to detect these users at an early stage to avoid the propagation of unreliable content in social networks’ ecosystems. In this work, we introduce a methodology to extract large corpus of unreliable posts using Twitter and two databases of unreliable websites (OpenSources and Media Bias Fact Check). In addition, we present an analysis of the content and users that publish and share several types of unreliable content. Finally, we develop supervised models to classify a twitter account according to its reliability. The experiments conducted using two different data sets show performance above 94% using Decision Trees as the learning algorithm. These experiments, although with some limitations, provide some encouraging results for future research on detecting unreliable accounts on social networks. © 2020, Springer Nature Switzerland AG.
2020
Autores
Figueira, A;
Publicação
EDULEARN20 Proceedings
Abstract
2020
Autores
Figueira, A;
Publicação
EDULEARN20 Proceedings
Abstract
2020
Autores
Guimaraes, N; Figueira, A; Torgo, L;
Publicação
PROCEEDINGS OF THE 16TH INTERNATIONAL CONFERENCE ON WEB INFORMATION SYSTEMS AND TECHNOLOGIES (WEBIST)
Abstract
The growth of social media as an information medium without restrictive measures on the creation of new accounts led to the rise of malicious agents with the intend to diffuse unreliable information in the network, ultimately affecting the perception of users in important topics such as political and health issues. Although the problem is being tackled within the domain of bot detection, the impact of studies in this area is still limited due to 1) not all accounts that spread unreliable content are bots, 2) human-operated accounts are also responsible for the diffusion of unreliable information and 3) bot accounts are not always malicious (e.g. news aggregators). Also, most of these methods are based on supervised models that required annotated data and updates to maintain their performance through time. In this work, we build a framework and develop knowledge-based metrics to complement the current research in bot detection and characterize the impact and behavior of a Twitter account, independently of the way it is operated (human or bot). We proceed to analyze a sample of the accounts using the metrics proposed and evaluate the necessity of these metrics by comparing them with the scores from a bot detection system. The results show that the metrics can characterize different degrees of unreliable accounts, from unreliable bot accounts with a high number of followers to human-operated accounts that also spread unreliable content (but with less impact on the network). Furthermore, evaluating a sample of the accounts with a bot detection system shown that bots compose around 11% of the sample of unreliable accounts extracted and that the bot score is not correlated with the proposed metrics. In addition, the accounts that achieve the highest values in our metrics present different characteristics than the ones that achieve the highest bot score. This provides evidence on the usefulness of our metrics in the evaluation of unreliable accounts in social networks. Copyright
2020
Autores
Silva, J; Aparicio, D; Ribeiro, P; Silva, F;
Publicação
PROCEEDINGS OF THE 35TH ANNUAL ACM SYMPOSIUM ON APPLIED COMPUTING (SAC'20)
Abstract
Scientific impact is commonly associated with the number of citations received. However, an author can easily boost his own citation count by (i) publishing articles that cite his own previous work (self-citations), (ii) having co-authors citing his work (co-author citations), or (iii) exchanging citations with authors from other research groups (reciprocated citations). Even though these friendly citations inflate an author's perceived scientific impact, author ranking algorithms do not normally address them. They, at most, remove self-citations. Here we present Friends-Only Citations AnalySer (FOCAS), a method that identifies friendly citations and reduces their negative effect in author ranking algorithms. FOCAS combines the author citation network with the co-authorship network in order to measure author proximity and penalises citations between friendly authors. FOCAS is general and can be regarded as an independent module applied while running (any) PageRank-like author ranking algorithm. FOCAS can be tuned to use three different criteria, namely authors' distance, citation frequency, and citation recency, or combinations of these. We evaluate and compare FOCAS against eight state-of-the-art author ranking algorithms. We compare their rankings with a ground-truth of best paper awards. We test our hypothesis on a citation and co-authorship network comprised of seven Information Retrieval top-conferences. We observed that FOCAS improved author rankings by 25% on average and, in one case, leads to a gain of 46%.
2020
Autores
Silva, J; Marques, ERB; Lopes, LMB; Silva, F;
Publicação
2020 FIFTH INTERNATIONAL CONFERENCE ON FOG AND MOBILE EDGE COMPUTING (FMEC)
Abstract
Edge computing is a hot research topic given the ever-increasing requirements of mobile applications in terms of computation and communication and the emerging Internet-of-Things with billions of devices. While ubiquitous and with considerable computational resources, devices at the edge may not be able to handle processing tasks on their own and thus resort to offloading to cloudlets, when available, or traditional cloud infrastructures. In this paper, we present JAY, a modular and extensible platform for mobile devices, cloudlets, and clouds that can manage computational tasks spawned by devices and make informed decisions about offloading to neighboring devices, cloudlets, or traditional clouds. JAY is parametric on the scheduling strategy and metrics used to make offloading decisions, providing a useful tool to study the impact of distinct offloading strategies. We illustrate the use of JAY with an evaluation of several offloading strategies in distinct cloud configurations using a real-world machine learning application, firing tasks can be dynamically executed on or offloaded to Android devices, cloudlet servers, or Google Cloud servers. The results obtained show that edge-clouds form competent computing platforms on their own and that they can effectively be meshed with cloudlets and traditional clouds when more demanding processing tasks are considered. In particular, edge computing is competitive with infrastructure clouds in scenarios where data is generated at the edge, high bandwidth is required, and a pool of computationally competent devices or an edge-server is available. The results also highlight JAY's ability of exposing the performance compromises in applications when they are deployed over distinct hybrid cloud configurations using distinct offloading strategies.
The access to the final selection minute is only available to applicants.
Please check the confirmation e-mail of your application to obtain the access code.