Multimodal Detection of Bots on X (Twitter) using Transformers (2024)

Loukas Ilias, Ioannis Michail Kazelidis, Dimitris AskounisThe authors are with the Decision Support Systems Laboratory, School of Electrical and Computer Engineering, National Technical University of Athens, 15780 Athens, Greece (e-mail: lilias@epu.ntua.gr; gkazelid@gmail.com; askous@epu.ntua.gr).

Abstract

Although not all bots are malicious, the vast majority of them are responsible for spreading misinformation and manipulating the public opinion about several issues, i.e., elections and many more. Therefore, the early detection of bots is crucial. Although there have been proposed methods for detecting bots in social media, there are still substantial limitations. For instance, existing research initiatives still extract a large number of features and train traditional machine learning algorithms or use GloVe embeddings and train LSTMs. However, feature extraction is a tedious procedure demanding domain expertise. Also, language models based on transformers have been proved to be better than LSTMs. Other approaches create large graphs and train graph neural networks requiring in this way many hours for training and access to computational resources. To tackle these limitations, this is the first study employing only the user description field and images of three channels denoting the type and content of tweets posted by the users. Firstly, we create digital DNA sequences, transform them to 3d images, and apply pretrained models of the vision domain, including EfficientNet, AlexNet, VGG16, etc. Next, we propose a multimodal approach, where we use TwHIN-BERT for getting the textual representation of the user description field and employ VGG16 for acquiring the visual representation for the image modality. We propose three different fusion methods, namely concatenation, gated multimodal unit, and crossmodal attention, for fusing the different modalities and compare their performances. Finally, we present a qualitative analysis of the behavior of our best performing model. Extensive experiments conducted on the Cresci’17 and TwiBot-20 datasets demonstrate valuable advantages of our introduced approaches over state-of-the-art ones.

Index Terms:

Bot detection, X (Twitter), digital DNA, image classification, transformers, gated multimodal unit, crossmodal attention

I Introduction

Social media platforms, such as Twitter (rebranded to X in 2023) and Reddit, constitute a valuable form of information, where the users have the opportunity to express their feelings, communicate with people around the world, and get informed about the news. For instance, social media have been used for the early detection of mental disorders [1]. Especially in Twitter, people can be informed about everything happening at the moment through the Twitter trends. However, social media are often manipulated by bots. Although not all bots are malicious and dangerous, the majority of them constitute the main form of misinformation [2]. Research has showed that bots can influence elections [3, 4], promote phishing attacks [5], and constitute an effective tool for manipulating social media by spreading articles of low-credibility scores [6]. According to [7], bots dominated hate speech during COVID-19. Therefore, the early detection of bots appears to be imperative nowadays.

Existing research initiatives train shallow machine learning classifiers for recognizing bots in Twitter and propose feature extraction approaches [8, 9]. However, feature extraction constitutes a time-consuming process, demands some level of domain expertise, while the optimal feature set for each specific task may not be found. In addition, there have been introduced deep learning approaches. However, these approaches either use as input a large number of features, including user and tweet metadata, or exploit GloVe embeddings and train LSTM neural networks [10, 11]. However, fine-tuning language models based on transformers requires less time for training and yields better evaluation results. Additionally, methods fine-tuning BERT, RoBERTa, etc. models cannot be applied to other languages except the english one. Moreover, graph-based approaches have been introduced for detecting bots. However, feature extraction techniques are also applied for creating feature vectors per users, which represent the nodes in graphs [12]. At the same time, large graphs are often created, which require a lot of time for training and require access to computational resources.

To address these limitations, we present the first study which exploits only the user description and the sequence of actions that an account performs for recognizing bots in Twitter. In the current study, we introduce both unimodal, i.e., neural networks exploiting either text or image, and multimodal approaches, i.e., neural networks utilizing both text and image. First, motivated by [13, 14], we design the digital DNA per user, which indicates the sequence of actions that an account performs. Next, we adopt the methodology proposed by [15] for converting the DNA sequence into an image. After that, we fine-tune several pretrained models, including AlexNet, ResNet, VGG16, etc., and compare their performances. For recognizing bots through user descriptions, this is the first study fine-tuning TwHIN-BERT [16] in the task of bot detection in Twitter. Regarding the multimodal approaches, we employ VGG16, which constitutes our best performing model, and extract the visual representation. In terms of the textual modality, we employ the TwHIN-BERT and extract the textual representation. Then, we propose three methods for fusing the representations of the different modalities. Specifically, first we concatenate the representation vectors of the two modalities. Second, we exploit a gated multimodal unit (GMU), which controls the importance of each modality towards the final classification. Thirdly, we use a cross-attention mechanism for capturing the inter-modal interactions. Experiments conducted on the Cresci’17 and TwiBot-20 datasets show that the introduced cross-modal model outperforms the competitive multimodal ones.

Our main contributions can be summarized as follows:

  • We create digital DNA sequences per user based on both the type and contents of the tweets and transform these sequences into 3d images.

  • We employ and compare many models of the vision domain, including EfficientNet, AlexNet, VGG16, etc. by utilizing the DNA sequence as an image consisting of three channels.

  • This is the first study fine-tuning the TwHIN-BERT model for recognizing bots utilizing only the user description field.

  • To the best of our knowledge, this is the first study introducing multimodal models employing the user description and the representation of the DNA sequence as an image.

  • We present the first study utilizing a gated multimodal unit and cross-attention mechanism, and comparing the fusion methods.

II Related Work

II-A Traditional Machine Learning Algorithms

In [8], the authors proposed two approaches for detecting bots in Twitter both at account and at tweet-level. In terms of the account level classification, the authors extracted a large set of features per user, applied feature selection algorithms, sampling techniques for dealing with the imbalanced datasets, and trained shallow machine learning algorithms.

Similarly, the authors in [10] adopted methods for recognizing bots both at account and tweet level. In terms of the account level, the authors extracted a set of features, including number of statuses (number of tweets and retweets), Followers Count, Friends Count, Favorites Count, Listed Count, Default Profile, etc. Next, the authors combined SMOTE with data enhancement via edited nearest neighbors (SMOTENN) and Tomek Links (SMOTOMEK), and trained traditional machine learning algorithms. In terms of the tweet level classification, the authors introduced a deep neural network consisting of LSTM and dense layers. The authors used as input GloVe embeddings of tweets and tweet metadata. Auxiliary output was also used, whose target was also the classification label.

The authors in [9] extracted a set of style-based features, including the number of punctuation marks, number of hashtags, number of retweets, number of user mentions, number of url links, and many more. They trained the following classification algorithms: Random Forest, Naive Bayes, Logistic Regression, Support Vector Machine with rbf and linear kernel, and Convolution Neural Network.

A wavelet-based approach was introduced by [17]. Specifically, the authors exploited the discrete wavelet transform and extracted a set of features, namely wavelet magnitude, wavelet phase, wavelet domain score, and so on. After this, a feature selection technique, namely a Correlation-based Feature Subset Selection, was proposed for reducing the dimensionality of the feature vector. Finally, a Random Forest Classifier was trained.

The authors in [18] proposed Botometer, which is a publicly available bot detection tool via the website, Python, or REST APIs. The authors use network, user, friends, and temporal features and train a Random Forest Classifier. The only input to this publicly available tool is the screen name.

II-B Deep Learning and Transformer-based Approaches

In [19], the authors exploited three types of LSTMs and four types of features, namely tweet content, tweet metadata, account metadata, and user description. In terms of the tweet content and the user description, the authors exploited the GloVe embeddings. Regarding account metadata, the authors extracted 30 features, while 15 tweet metadata features were also extracted. One limitation of this study is the feature extraction procedure and the usage of GloVe embeddings in conjunction with LSTM models instead of pretrained language models based on transformers, which capture better the context and require significantly less time for fine-tuning.

A similar approach was introduced by [11], where the authors utilized GloVe embeddings for representing the user description field as embeddings. Also, the authors extracted several types of features, namely number of followers and friends, number of tweets and retweets posted by the user, length of the name, entropy of the screen name, entropy of the description, etc. They passed the GloVe embeddings of the user description field into LSTM models and concatenated the output of the LSTMs with the rest of the features.

In [8], the authors proposed two approaches for detecting bots in Twitter both at account and at tweet-level. With regards to the tweet-level classification, the authors utilized a deep neural network. Specifically, the authors exploited GloVe embeddings and passed them through BiLSTM layers coupled with an attention mechanism.

Similarly, the authors in [10] adopted methods for recognizing bots both at account and tweet level. In terms of the tweet level classification, the authors introduced a deep neural network consisting of LSTM and dense layers. The authors used as input GloVe embeddings of tweets and tweet metadata. Auxiliary output was also used, whose target was also the classification label.

A different approach was proposed by [20], where the authors exploited Generative Adversarial Neural networks (GANs) for recognizing bots in Twitter. Specifically, the authors addressed the limitation of original SeqGAN [21] by introducing the GANBOT setting, where a shared LSTM was used between generator and classifer. Similar to [10], the authors used GloVe embeddings as input to LSTMs. GANs were also exploited by [22]. Specifically, the authors extracted first a set of features consisting of user meta-data, sentiment, friends, content, network, and timing. Finally, they proposed a new method based on Conditional GANs by introducing the Wasserstein distance with a gradient penalty and a condition-generation method through the modified density peak clustering algorithm. The authors compared their approaches with Random Oversampling, SMOTE, and ADASYN.

In [23], the authors use language models based on transformers, namely BERT and RoBERTa, for recognizing bots in Twitter. The authors concatenated the text embeddings along with additional metadata and trained a neural network consisting of dense layers.

In [15], the authors proposed a new method for recognizing bots in Twitter based on image recognition. Specifically, they introduced a new approach for converting the digital DNA sequence into an image consisting of three channels. Finally, the authors applied ResNet and stated that this method achieved competitive results to state-of-the-art approaches. However, the authors created a digital DNA sequence based only on the type of the tweets, while they fine-tuned only one pretrained model.

The authors in [24] utilized the concept of digital DNA sequences along with pretrained BERT models. First, the authors trained BERT on sentiment analysis classification tasks. Next, they passed each tweet through a BERT model and predicted the sentiment of each tweet, i.e., positive, negative, and neutral. In this way, they create a digital DNA sequence consisting of characters denoting the sentiment of each tweet.

A deep learning architecture consisting of two branches was proposed by [25]. Specifically, the authors used word2vec and passed the respective embeddings through CNNs. Next, they extracted features, including number of followers, number of tweets, account age/reputation, etc., and passed them through dense layers. After this, they concatenated the two branches and passed the resulting vector through a dense layer for getting the final prediction.

An active learning approach was adopted by [26]. Specifically, the authors extracted a set of features, including metadata-based, interaction-based, content-based, and timing-based. Next, active learning was employed for efficiently expanding the labeled data. Finally, a deep neural network consisting of ResNet, BiGRU, and attention layer was trained.

The study in [27] introduced a CNN and BiLSTM-based deep neural network model coupled with an attention mechanism. Specifically, the authors passed the profile, temporal, and activity information to a two-layers stacked BiLSTM, whereas they passed the contentinformation to a deep CNN. An attention mechanism was exploited at the top of the proposed architecture. Findings showed that the proposed approach outperformed the state-of-the-art approaches.

A CNN-LSTM network was also proposed by [28]. Specifically, the authors model the social behaviour per user, i.e., posting and retweeting, and exploit an LSTM network. Next, the authors consider users’ history tweets as temporal text data and use a CNN-LSTM network. Finally, they fuse the representation vectors obtained by the two networks and get the final prediction.

In [29], the authors extract word embeddings, character embeddings, part-of-speech embeddings, and named-entity embeddings. They pass these embeddings through BiLGRU [30] and get the final prediction. Findings stated that the proposed model performed better or comparably to state-of-the-art Twitter bot detection models.

In [31], the authors used GloVe embeddings as input to a 3-layer BiLSTM neural network for distinguishing Twitter bots from human accounts.

II-C Graph-based Approaches

Recently, graph convolutional networks have been also used for identifying bots in Twitter. Specifically, the study in [12] created an heterogeneous graph and employed Relational Graph Convolutional Networks. The authors represented each user as a vector consisting of user description, tweets, numerical and categorical properties. To be more precise, the authors employed RoBERTa to get a representation vector of the user description. Similarly, they used RoBERTa and averaged the representations of all the tweets posted by the user into one single vector. In terms of the numerical properties, the authors created a vector consisting of the number of friends, followers, favourites, etc. Regarding the categorical properties, the resulting vector was composed of binary variables, i.e., if the user has a profile image, if the user is verified, etc. However, this method cannot incorporate the intrinsic heterogeneity of relation. Therefore, the authors in [32] introduced propose a novel Twitter bot detection framework thatis graph-based and heterogeneity-aware. Specifically, the authors encoded each user adopting the procedure followed in [12]. Finally, the authors exploited relational graph transformers and semantic attention networks.

The authors in [33] proposed a text-graph interaction module along with a semantic consistency module. Specifically, the authors exploited the description and tweets and passed them through RoBERTa for getting the textual representation (embeddings). For graph-based approach, they adopted the method proposed by [12]. Next, they introduced a method for capturing the interactions between the text and graph modality. After that, a semantic consistency detection module was proposed, which exploits the attention weights obtained by the RoBERTa model. Finally, the respective vectors were concatenated and were passed through a dense layer for the final classification.

Multi-view graph attention networks were exploited in [34]. The authors used different datasets for training and evaluating their proposed approaches under a transfer learning scenario. Profile features were also utilized, including number of followers/friends, age in days, and many more.

Motivated by the fact that existing methods almost ignore the differences in bot behaviors in multiple domains, the study in [35] introduces a domain-aware approach for recognizing bots. Specifically, multi-relational graphs were exploited coupled with a user representation learning module, consisting of a series of graph embedding layers and semantic attention layers. Finally, domain-aware classifiers were exploited for detecting bots. Additionally, according to the authors this is the first study employing a federated learning framework.

The authors in [36] introduced a Graph Convolutional Neural Network for exploiting the characteristics of the accounts’ neighbours and identifying bots. Each user is represented by a node consisting of the following feature set: age, favourites_count, statuses_count, account length name, followers_count, and friends_count. Results showed that the proposed approach outperformed the state-of-the-art ones.

II-D Unsupervised Learning

Cresci et al. [14] introduced digital DNA for recognizing bots in Twitter. Specifically, the authors set the tweets of each user into chronological user and created a digital DNA sequence according to the type and content of each tweet. In terms of the type of the tweets, the resulting DNA sequence was composed of A, C, and T, where a "C" indicates a reply, a "T" denotes a retweet, and an "A" denotes a simple tweet. With regards to the content of the tweet, the authors examined whether the tweet contained hashtags, URLs, mentions, media, etc. Finally, the authors calculated similarities between these sequences using the longest common subsequence (LCS) similarity and clustered users based on their similarity scores.

In [37], the authors introduce MulBot, which constitutes an unsupervised bot detector based on multivariate time series. Specifically, the proposed approach consists of the following steps: multidimensionaltemporal features extracted from user timelines, dimensionality reduction using an autoencoder, extraction of statistical global features (optional), concatenation of global features with vectorial features in output from the encoder (optional), and clustering algoritm. Results showed that the proposed approach yielded satisfactory results in both the binary task and the multiclass classification one.

The bot detection task was considered as an anomaly detection task in [38]. The authors utilized 95 one-gram features from tweet text along with user features and modified two stream clustering algorithms, namely StreamKM++ and DenStream.

II-E Reinforcement Learning

A reinforcement learning approach was proposed by [39] for searching the GNN architecture. In this way, the most suitable multi-hop neighborhood and the number oflayers in the GNN architecture are found. An Heterogeneous Information Network was also exploited for modelling the entities and relationships in the social networks. Finally, the authors exploited self-supervised learning approaches. Also, the authors in [40] introduced a deep Q-network architecture by incorporating a Deep Q-Learning (DQL) model. The authors extracted tweet-based, user profile-based, and social graph-based features. In terms of the tweet-based, the authors utilized syntax, semantic, and temporal behaviour features. Regarding the user profile features, the authors employed features pertaining to user behaviour (Posting tweets in several languages, URL/hashtag/mention ratio) and user interactions (Number of active days, Number of retweeted tweets). Finally, they defined social graph-based attributes, such as clustering coefficient, closeness centrality, betweenness centrality and pagerank centrality for each user.

The authors in [41] extracted URL-based features, including URL redirection, frequency of shared URLs, and spam content in URL, and designed a learning automata based model. Specifically, the authors designed a trust computation model, which contains two parameters, namely direct trust and indirect trust. Findings suggested that the proposed approach improved the existing ones.

II-F Related Work Review Findings

In spite of the rise of deep learning algorithms, existing research initiatives extract still a large number of features and train traditional machine learning algorithms. However, feature extraction constitutes a tedious procedure demanding domain expertise. Thus, the optimal set of features may not be found. Additionally, existing research initiatives still exploit GloVe embeddings and train LSTMs instead of employing language models based on transformers, which achieve state-of-the-art results across a large number of domains. In addition, methods employing BERT cannot be employed for languages other than the english one. Recently, there have been introduced graph-based approaches. However, these methods still extract a large number of features per user for creating a feature vector per node-user. At the same time, large graphs are created, which require access to computational resources and need a lot of time for training.

Therefore, our proposed work differs significantly from the aforementioned research works, since we (1) exploit only the user description field and create images according to the type and content of the tweets posted by each user, (2) create images consisting of three channels and train several pretrained models comparing their performances, (3) employ TwHIN-BERT which is a new multi-lingual Tweet language model that is trained on 7 billion Tweets from over 100 distinct languages, (4) introduce multimodal models, which take as input the user description field and the 3d images.

III Datasets

III-A Cresci’17 Dataset

We use the dataset introduced in [42] for conducting our experiments. Specifically, this dataset consists of genuine accounts and some sets of social and traditional spambots. In the present study, we exploit the set of genuine accounts and the set of social spambots #1. For collecting genuine users, the authors adopted the methodology proposed in [43]. Specifically, the authors contacted with some accounts and asked them a question. Then, the authors examined the users, who replied to the question and verified 3,474 accounts. In terms of the social spambots users, the authors used a group of social bots that was discovered on Twitter during the last Mayoral election in Rome, in 2014. Specifically, this set of social spambots corresponds to retweeters of an Italian political candidate. As mentioned in SectionII-D, the authors presented a figure illustrating the LCS curves of humans and bots. Findings showed that the LCS of bots were long, even when the number of account increases. A sudden dropin LCS length of bots is observed when the number of accounts gets close to the group size. On the contrary, the LCS curve of humans indicates little to no similarity. Also, in [42], the authors illustrate the cumulative distribution function (CDF) of join date and number of followers. Results indicate that social spambots have anomalous distributions. In our experiments, we keep users having available the user description field in their profile. To address the issue of the imbalanced dataset, we create a dataset consisting of 943 real users and 943 social spambots. This technique, which addresses the issue of imbalanced dataset by downsampling the set of real users, has been adopted by previous studies [15, 37].

III-B TwiBot-20 Dataset

We use the TwiBot-20 dataset to conduct our experiments [44]. Contrary to other datasets which include a specific type of users, this dataset consists of diversified bots that co-exist on the real-world Twittersphere. Users of this dataset have an interest in four domains, including politics, business, entertainment, and sports. Thus, this dataset constitutes a challenging task, since it includes multiple types of bots. For proving the high quality of annotations, the authors illustrate the CDF of account reputation (number of friends and followers), user tweet counts, and screen name likelihood. CDF plots show that genuine accounts present higher reputation scores than bots in the TwiBot-20 dataset. In terms of user tweet counts, CDF plots indicate that bots generate fewer tweets than humans, in order to avoid the traditional detection methods. Finally, findings of CDF plot regarding the screen name likelihood show that bot users in TwiBot-20 do have slightly lower screen name likelihood. Also, the authors in [44] conduct a user diversity analysis and study the distribution of profile locations and user interests. Findings of a geographic analysis study demonstrate that India, United States, Europe, and Africa are represented in the TwiBot-20 dataset. Finally, the authors examine the most frequent hashtags, present their analysis through a bar plot, and show that twitter users in TwiBot-20 are proved to be diversified in interest domains.TwiBot-20 dataset includes 229,573 users. However, the authors have labelled 11,826 users. This dataset includes the 200 recent tweets per user.The authors in [44] have divided TwiBot-20 dataset into a train, validation, and test set. We remove users who have not posted any tweets and have not added a description field in their profile. TableI presents the distribution of genuine accounts and bots into each set.

genuine usersbots
train set3,2623,911
development set9591,101
test set494533

IV Unimodal Models Utilizing only Images for Detecting Bots in Twitter

We adopt the methodology introduced by [13, 14] for constructing a DNA sequence. Inspired by the biological DNA sequence consisting of A (adenine), C (cytosine), G (guanine) and T (thymine), the authors in [13, 14] introduce the digital DNA for creating a DNA sequence based either on type of the tweet or the content of the tweet. In this study, we create two digital DNA sequences based on both the type and content of the tweet. First, we set all tweets into a chronological order. Regarding the first approach, we create a sequence consisting of A, T, and C. A "T" denotes a retweet, a "C" indicates a reply, while an "A" denotes a tweet. In terms of the second approach, which is based on the content of the tweet, we create a digital DNA sequence consisting of N, U, H, M, and X. Specifically, a "N" indicates that the tweet contains no entities (plain text), a "U" denotes that the tweet contains one or more ULRs, a "H" means that the tweet contains one or more hashtags, a "M" indicates that the tweet contains one or more mentions, and a "X" indicates that the tweet contains entities of mixed types.

After having created a digital DNA sequence per user, we adopt the method introduced in [15] for transforming the DNA sequence into an image consisting of three channels. First, we check whether the length of the longest DNA sequence is a perfect square. If it is a perfect square, then we define the image size as the square root of the length of the longest DNA sequence. Otherwise, we consider the perfect square closest to and larger than the maximum length. In this way, all strings can be converted to images of equal sizes. Next, we assign to each symbol of the DNA sequence, i.e., A, C, and T, a "color" for creating the image. The image is colored pixel by pixel based on the coors assigned to thecorrespondent symbol. The same methodology is adopted in terms of the digital DNA sequence based on the content of the tweets, which consists of N, U, H, M, and X. The resulting image is a grayscale image in a (1, H, W) format. Similar to [15], we convert the images into a (3, H, W) format utilizing the grayscale transformation111https://pytorch.org/vision/main/generated/torchvision.transforms.Grayscale. html.

Each image is resized to 256×256256256256\times 256256 × 256 pixels. Next, we fine-tune the following pretrained models: GoogLeNet (Inception v1) [45], ResNet50 [46], WideResNet-50-2 [47], AlexNet [48], SqueezeNet1_0 [49], DenseNet-201 [50], MobileNetV2[51], ResNeXt-50 32×\times×4d [52], VGG16 [53], and EfficientNet-B4222We experimented with EfficientNet-B0 to B7, but EfficientNet-B4 was the best performing model. [54].

IV-A Experiments

IV-A1 Experimental Setup

In terms of the Cresci’17 dataset, we divide the dataset into a train, validation, and test set (80%-10%-10%). Regarding the TwiBot-20 dataset, the train, validation, and test sets are provided by the authors [44]. We use EarlyStopping and stop training if the validation loss has stopped decreasing for 6 consecutive epochs. We use Adam optimizer with a learning rate of 1e-5 and exploit ReduceLROnPlateau, where we reduce the learning rate by a factor of 0.1 if the validation loss has not presented an improvement after 3 consecutive epochs. We minimize the cross-entropy loss function. In terms of the TwiBot-20 dataset, we apply class-weights to the loss function to deal with the data imbalance. We train the proposed approaches for a maximum of 30 epochs. All models have been created using the PyTorch library [55]. All experiments are conducted on a single Tesla T4 GPU.

IV-A2 Evaluation Metrics

We use Precision, Recall, F1-score, Accuracy, and Specificity for evaluating the results of the proposed approaches. These metrics have been computed by considering the class of bots as the positive one (label=1). Results are obtained over five runs using different random seedsreporting the average and the standard deviation.

IV-A3 Results

The results of our proposed approaches are reported in Tables II-V. Specifically, TablesII and III report the results on the Cresci’17 dataset, while TablesIV and V report the results on the TwiBot-20 dataset.

Cresci’17 Dataset

The results of our proposed approaches are reported in Tables II and III.

As one can easily observe in TableII, VGG16 constitutes the best performing model outperforming the other pretrained models in Recall, F1-score, and Accuracy. It is worth noting that there are models surpassing VGG16 or obtaining equal performance in Precision and Specificity. However, VGG16 outperforms these models in F1-score, which constitutes the weighted average of Precision and Recall. Additionally, F1-score is a more important metric than Specificity, since high Specificity and low F1-score indicates that some social spambots are falsely detected as real accounts. VGG16 outperforms the other models in Recall by 0.44-3.48%, in F1-score by 0.22-2.06%, and in Accuracy by 0.19-2.12%. WideResNet-50-2 constitutes the second best performing model attaining an Accuracy of 99.58% and a F1-score of 99.57%. Specifically, it surpasses the other models, except for VGG16, in Accuracy by 0.64-1.91% and in F1-score by 0.59-1.84%. F1-score ranging from 98.17% to 98.98% is obtained by the rest of the models, except for EfficientNet-B4. Similarly, Accuracy ranging from 98.20% to 98.94% is attained by the rest of the models, except for EfficientNet-B4. The highest Precision and Specificity scores are obtained by GoogLeNet and ResNet50 and are equal to 1.00. EfficientNet-B4 obtains the worst evaluation results reaching Accuracy and F1-score up to 97.67% and 97.73% respectively.

As one can easily observe in TableIII, VGG16 constitutes our best performing model obtaining an Accuracy of 99.89%, a Precision of 1.00%, a Recall of 99.78%, a F1-score of 99.89%, and a Specificity of 1.00%. Specifically, it outperforms the other models in Recall by 0.44-2.14%, in F1-score by 0.52-2.87%, and in Accuracy by 0.52-2.75%, while it achieves equal Precision and Specificity scores with GoogLeNet, ResNet50, and AlexNet. Additionally, GoogLeNet, WideResNet-50-2, AlexNet, and MobileNetV2 achieve Accuracy and F1-scores over 99.00%. AlexNet constitutes the second best performing model obtaining an Accuracy of 99.37% and a F1-score of 99.37%. EfficientNet-B4 achieves the worst evaluation results reaching Accuracy and F1-score up to 97.14% and 97.02% respectively.

Evaluation metrics
ArchitecturePrecisionRecallF1-scoreAccuracySpecificity
digital DNA (type of tweets)
GoogLeNet (Inception v1)100.0097.6398.7998.73100.00
±plus-or-minus\pm±0.00±plus-or-minus\pm±1.39±plus-or-minus\pm±0.72±plus-or-minus\pm±0.79±plus-or-minus\pm±0.00
ResNet50100.0097.8398.9098.94100.00
±plus-or-minus\pm±0.00±plus-or-minus\pm±1.00±plus-or-minus\pm±0.52±plus-or-minus\pm±0.47±plus-or-minus\pm±0.00
WideResNet-50-299.7899.3599.5799.5899.79
±plus-or-minus\pm±0.44±plus-or-minus\pm±1.29±plus-or-minus\pm±0.87±plus-or-minus\pm±0.85±plus-or-minus\pm±0.42
AlexNet99.5997.1298.3498.3199.55
±plus-or-minus\pm±0.81±plus-or-minus\pm±1.23±plus-or-minus\pm±0.88±plus-or-minus\pm±0.91±plus-or-minus\pm±0.91
SqueezeNet1_099.1797.2398.1798.2099.14
±plus-or-minus\pm±1.02±plus-or-minus\pm±1.41±plus-or-minus\pm±0.55±plus-or-minus\pm±0.54±plus-or-minus\pm±1.05
DenseNet-20199.3798.5998.9898.9499.35
±plus-or-minus\pm±0.51±plus-or-minus\pm±0.98±plus-or-minus\pm±0.55±plus-or-minus\pm±0.58±plus-or-minus\pm±0.53
MobileNetV299.5897.7098.6398.6299.57
±plus-or-minus\pm±0.52±plus-or-minus\pm±1.28±plus-or-minus\pm±0.76±plus-or-minus\pm±0.72±plus-or-minus\pm±0.53
ResNeXt-50 32 ×\times× 4d99.7997.5998.6798.6299.78
±plus-or-minus\pm±0.43±plus-or-minus\pm±1.64±plus-or-minus\pm±1.01±plus-or-minus\pm±1.04±plus-or-minus\pm±0.44
VGG1699.7899.5599.6799.6899.80
±plus-or-minus\pm±0.44±plus-or-minus\pm±0.54±plus-or-minus\pm±0.44±plus-or-minus\pm±0.42±plus-or-minus\pm±0.41
EfficientNet-B499.2096.3197.7397.6799.08
±plus-or-minus\pm±1.14±plus-or-minus\pm±0.60±plus-or-minus\pm±0.47±plus-or-minus\pm±0.54±plus-or-minus\pm±1.35
Evaluation metrics
ArchitecturePrecisionRecallF1-scoreAccuracySpecificity
digital DNA (content of tweets)
GoogLeNet (Inception v1)100.0098.0999.0399.05100.00
±plus-or-minus\pm±0.00±plus-or-minus\pm±1.27±plus-or-minus\pm±0.65±plus-or-minus\pm±0.62±plus-or-minus\pm±0.00
ResNet50100.0097.8198.8998.84100.00
±plus-or-minus\pm±0.00±plus-or-minus\pm±1.13±plus-or-minus\pm±0.58±plus-or-minus\pm±0.62±plus-or-minus\pm±0.00
WideResNet-50-298.9399.3499.1499.1598.95
±plus-or-minus\pm±0.71±plus-or-minus\pm±0.89±plus-or-minus\pm±0.69±plus-or-minus\pm±0.63±plus-or-minus\pm±0.64
AlexNet100.0098.7599.3799.37100.00
±plus-or-minus\pm±0.00±plus-or-minus\pm±1.20±plus-or-minus\pm±0.61±plus-or-minus\pm±0.62±plus-or-minus\pm±0.00
SqueezeNet1_099.3697.6498.4998.5299.36
±plus-or-minus\pm±0.53±plus-or-minus\pm±1.06±plus-or-minus\pm±0.64±plus-or-minus\pm±0.62±plus-or-minus\pm±0.52
DenseNet-20199.5997.9098.7398.7399.56
±plus-or-minus\pm±0.49±plus-or-minus\pm±1.37±plus-or-minus\pm±0.58±plus-or-minus\pm±0.54±plus-or-minus\pm±0.54
MobileNetV299.5998.9899.2899.2699.56
±plus-or-minus\pm±0.50±plus-or-minus\pm±1.27±plus-or-minus\pm±0.74±plus-or-minus\pm±0.79±plus-or-minus\pm±0.54
ResNeXt-50 32 ×\times× 4d99.3598.4398.8898.9499.38
±plus-or-minus\pm±0.53±plus-or-minus\pm±1.31±plus-or-minus\pm±0.52±plus-or-minus\pm±0.47±plus-or-minus\pm±0.51
VGG16100.0099.7899.8999.89100.00
±plus-or-minus\pm±0.00±plus-or-minus\pm±0.43±plus-or-minus\pm±0.22±plus-or-minus\pm±0.21±plus-or-minus\pm±0.00
EfficientNet-B496.1397.9897.0297.1496.36
±plus-or-minus\pm±2.26±plus-or-minus\pm±2.01±plus-or-minus\pm±1.38±plus-or-minus\pm±1.28±plus-or-minus\pm±2.14
TwiBot-20 Dataset

The results of our proposed approaches are reported in Tables IV and V.

As one can easily observe in TableIV, AlexNet constitutes the best performing model in terms of F1-score and Accuracy. Specifically, it outperforms the other models in F1-score by 0.48-1.34% and in Accuracy by 0.06-0.78%. Although other models outperform AlexNet in Precision and Recall, F1-score is a more important metric, since it is a weighted average of precision and recall. DenseNet-201 and MobileNetV2 achieve almost equal Accuracy scores accounting for 66.35% and 66.34% respectively. The second highest F1-score is achieved by SqueezeNet1_0 and is equal to 66.63%. GoogLeNet obtains the worst performance reaching Accuracy and F1-score up to 65.63% and 65.71% respectively.

As one can easily observe in TableV, AlexNet is the best performing model outperforming the rest of the models in Precision by 2.11-3.16%, in F1-score by 0.78-3.06%, in Accuracy by 1.60-2.50%, and in Specificity by 1.13-4.58%. GoogLeNet, SqueezeNet1_0, MobileNetV2, VGG16, and EfficientNet-B4 obtain almost equal Accuracy scores ranging from 65.45% to 65.59%, with MobileNetV2 outperforming these models in terms of both Accuracy and F1-score. ResNet50 obtains the worst performance reaching Accuracy and F1-score up to 64.69% and 63.74% respectively.

Evaluation metrics
ArchitecturePrecisionRecallF1-scoreAccuracySpecificity
digital DNA (type of tweets)
GoogLeNet (Inception v1)68.1263.5365.7165.6367.89
±plus-or-minus\pm±0.61±plus-or-minus\pm±2.58±plus-or-minus\pm±1.29±plus-or-minus\pm±0.67±plus-or-minus\pm±1.84
ResNet5067.9065.2966.4665.9066.56
±plus-or-minus\pm±1.44±plus-or-minus\pm±4.30±plus-or-minus\pm±1.90±plus-or-minus\pm±0.94±plus-or-minus\pm±3.98
WideResNet-50-267.5365.6366.5265.7665.91
±plus-or-minus\pm±0.73±plus-or-minus\pm±2.85±plus-or-minus\pm±1.30±plus-or-minus\pm±0.63±plus-or-minus\pm±2.31
AlexNet68.3365.8967.0566.4166.96
±plus-or-minus\pm±1.26±plus-or-minus\pm±2.05±plus-or-minus\pm±0.72±plus-or-minus\pm±0.62±plus-or-minus\pm±2.79
SqueezeNet1_067.5266.1966.6365.7465.26
±plus-or-minus\pm±2.13±plus-or-minus\pm±5.54±plus-or-minus\pm±1.72±plus-or-minus\pm±0.26±plus-or-minus\pm±6.27
DenseNet-20168.8664.3266.4566.3568.54
±plus-or-minus\pm±1.03±plus-or-minus\pm±3.10±plus-or-minus\pm±1.31±plus-or-minus\pm±0.52±plus-or-minus\pm±2.83
MobileNetV269.0464.0966.4466.3468.95
±plus-or-minus\pm±0.79±plus-or-minus\pm±2.26±plus-or-minus\pm±0.91±plus-or-minus\pm±0.37±plus-or-minus\pm±2.20
ResNeXt-50 32 ×\times× 4d68.2464.9566.4966.0867.29
±plus-or-minus\pm±1.15±plus-or-minus\pm±3.13±plus-or-minus\pm±1.22±plus-or-minus\pm±0.43±plus-or-minus\pm±3.14
VGG1667.8565.7866.5765.8665.95
±plus-or-minus\pm±2.35±plus-or-minus\pm±5.69±plus-or-minus\pm±1.81±plus-or-minus\pm±0.67±plus-or-minus\pm±6.68
EfficientNet-B467.8764.6566.1165.6966.80
±plus-or-minus\pm±1.44±plus-or-minus\pm±4.17±plus-or-minus\pm±1.54±plus-or-minus\pm±0.36±plus-or-minus\pm±4.25
Evaluation metrics
ArchitecturePrecisionRecallF1-scoreAccuracySpecificity
digital DNA (content of tweets)
GoogLeNet (Inception v1)67.5364.5465.9765.4766.48
±plus-or-minus\pm±0.78±plus-or-minus\pm±2.12±plus-or-minus\pm±0.94±plus-or-minus\pm±0.50±plus-or-minus\pm±2.07
ResNet5068.2559.8963.7464.6969.88
±plus-or-minus\pm±0.91±plus-or-minus\pm±2.86±plus-or-minus\pm±1.39±plus-or-minus\pm±0.52±plus-or-minus\pm±2.54
WideResNet-50-267.5762.5564.7864.8367.29
±plus-or-minus\pm±2.19±plus-or-minus\pm±4.80±plus-or-minus\pm±1.85±plus-or-minus\pm±0.81±plus-or-minus\pm±5.59
AlexNet70.3663.6466.8067.1971.01
±plus-or-minus\pm±1.11±plus-or-minus\pm±1.53±plus-or-minus\pm±0.37±plus-or-minus\pm±0.28±plus-or-minus\pm±2.20
SqueezeNet1_068.0863.3865.5265.4967.77
±plus-or-minus\pm±1.36±plus-or-minus\pm±4.16±plus-or-minus\pm±1.66±plus-or-minus\pm±0.31±plus-or-minus\pm±4.01
DenseNet-20167.2063.6065.3164.9766.43
±plus-or-minus\pm±1.02±plus-or-minus\pm±2.39±plus-or-minus\pm±0.96±plus-or-minus\pm±0.50±plus-or-minus\pm±2.70
MobileNetV267.6064.4766.0265.5966.80
±plus-or-minus\pm±0.46±plus-or-minus\pm±2.38±plus-or-minus\pm±1.14±plus-or-minus\pm±0.52±plus-or-minus\pm±1.71
ResNeXt-50 32 ×\times× 4d67.8461.6564.4064.7768.14
±plus-or-minus\pm±2.00±plus-or-minus\pm±5.19±plus-or-minus\pm±1.77±plus-or-minus\pm±0.40±plus-or-minus\pm±5.86
VGG1668.1363.1965.4165.4567.89
±plus-or-minus\pm±1.61±plus-or-minus\pm±4.95±plus-or-minus\pm±2.06±plus-or-minus\pm±0.64±plus-or-minus\pm±4.74
EfficientNet-B467.5664.5065.9665.4766.52
±plus-or-minus\pm±0.90±plus-or-minus\pm±2.09±plus-or-minus\pm±0.66±plus-or-minus\pm±0.13±plus-or-minus\pm±2.46

V Our Proposed Multimodal Models for Detecting Bots in Twitter

In this section, we describe our proposed models employing both textual and vision modalities. Our introduced models are illustrated in Fig.1.

Multimodal Detection of Bots on X (Twitter) using Transformers (1)
Multimodal Detection of Bots on X (Twitter) using Transformers (2)
Multimodal Detection of Bots on X (Twitter) using Transformers (3)

V-A Concatenation

In terms of the textual modality, we pass the user description field through a TwHIN-BERT model [16] and extract the [CLS] token. Let ftdtsuperscript𝑓𝑡superscriptsubscript𝑑𝑡f^{t}\in\mathbb{R}^{d_{t}}italic_f start_POSTSUPERSCRIPT italic_t end_POSTSUPERSCRIPT ∈ blackboard_R start_POSTSUPERSCRIPT italic_d start_POSTSUBSCRIPT italic_t end_POSTSUBSCRIPT end_POSTSUPERSCRIPT denote the representation vector of the textual modality. dtsubscript𝑑𝑡d_{t}italic_d start_POSTSUBSCRIPT italic_t end_POSTSUBSCRIPT denotes the dimensionality and is equal to 768. In terms of the visual modality, we create a 3d image as described in SectionIV. Then, we exploit our best performing model, namely VGG16 for Cresci’17 dataset and AlexNet for TwiBot-20 dataset, as mentioned in SectionIV-A3. Specifically, we remove the last layer of VGG16 (and AlexNet) and replace the next-to-last dense layer consisting of 4096 units with one dense layer consisting of 768 units. Let fvdvsuperscript𝑓𝑣superscriptsubscript𝑑𝑣f^{v}\in\mathbb{R}^{d_{v}}italic_f start_POSTSUPERSCRIPT italic_v end_POSTSUPERSCRIPT ∈ blackboard_R start_POSTSUPERSCRIPT italic_d start_POSTSUBSCRIPT italic_v end_POSTSUBSCRIPT end_POSTSUPERSCRIPT represent the visual representation vector. dvsubscript𝑑𝑣d_{v}italic_d start_POSTSUBSCRIPT italic_v end_POSTSUBSCRIPT is equal to 768.

After having calculated ftsuperscript𝑓𝑡f^{t}italic_f start_POSTSUPERSCRIPT italic_t end_POSTSUPERSCRIPT and fvsuperscript𝑓𝑣f^{v}italic_f start_POSTSUPERSCRIPT italic_v end_POSTSUPERSCRIPT, we concatenate these two representation vectors into a single vector as following:

z=[ft,fv]𝑧superscript𝑓𝑡superscript𝑓𝑣z=[f^{t},f^{v}]italic_z = [ italic_f start_POSTSUPERSCRIPT italic_t end_POSTSUPERSCRIPT , italic_f start_POSTSUPERSCRIPT italic_v end_POSTSUPERSCRIPT ](1)

, where zd𝑧superscript𝑑z\in\mathbb{R}^{d}italic_z ∈ blackboard_R start_POSTSUPERSCRIPT italic_d end_POSTSUPERSCRIPT. d=dt+dv𝑑subscript𝑑𝑡subscript𝑑𝑣d=d_{t}+d_{v}italic_d = italic_d start_POSTSUBSCRIPT italic_t end_POSTSUBSCRIPT + italic_d start_POSTSUBSCRIPT italic_v end_POSTSUBSCRIPT and is equal to 1536.

After this, we pass z𝑧zitalic_z through a dense layer consisting of 128 units with a ReLU activation function. Finally, we use a dense layer consisting of two units, which gives the final prediction.

Our introduced model is illustrated in Fig.1(a).

V-B Gated Multimodal Unit

In terms of the textual modality, we pass the user description field through a TwHIN-BERT model and extract the [CLS] token. Let ftdtsuperscript𝑓𝑡superscriptsubscript𝑑𝑡f^{t}\in\mathbb{R}^{d_{t}}italic_f start_POSTSUPERSCRIPT italic_t end_POSTSUPERSCRIPT ∈ blackboard_R start_POSTSUPERSCRIPT italic_d start_POSTSUBSCRIPT italic_t end_POSTSUBSCRIPT end_POSTSUPERSCRIPT denote the representation vector of the textual modality. dtsubscript𝑑𝑡d_{t}italic_d start_POSTSUBSCRIPT italic_t end_POSTSUBSCRIPT denotes the dimensionality and is equal to 768. In terms of the visual modality, we create a 3d image as described in SectionIV. Then, we exploit our best performing model, namely VGG16 for Cresci’17 dataset and AlexNet for TwiBot-20 dataset, as mentioned in SectionIV-A3. Specifically, we remove the last layer of VGG16 (and AlexNet) and replace the next-to-last dense layer consisting of 4096 units with one dense layer consisting of 768 units. Let fvdvsuperscript𝑓𝑣superscriptsubscript𝑑𝑣f^{v}\in\mathbb{R}^{d_{v}}italic_f start_POSTSUPERSCRIPT italic_v end_POSTSUPERSCRIPT ∈ blackboard_R start_POSTSUPERSCRIPT italic_d start_POSTSUBSCRIPT italic_v end_POSTSUBSCRIPT end_POSTSUPERSCRIPT represent the visual representation vector. dvsubscript𝑑𝑣d_{v}italic_d start_POSTSUBSCRIPT italic_v end_POSTSUBSCRIPT is equal to 768.

Next, we use a gated multimodal unit introduced in [56] for controlling the information flow of the textual and visual modalities. The equations governing the gated multimodal unit are presented below:

ht=tanh(Wtft+bt)superscript𝑡superscript𝑊𝑡superscript𝑓𝑡superscript𝑏𝑡h^{t}=\tanh{(W^{t}f^{t}+b^{t})}italic_h start_POSTSUPERSCRIPT italic_t end_POSTSUPERSCRIPT = roman_tanh ( italic_W start_POSTSUPERSCRIPT italic_t end_POSTSUPERSCRIPT italic_f start_POSTSUPERSCRIPT italic_t end_POSTSUPERSCRIPT + italic_b start_POSTSUPERSCRIPT italic_t end_POSTSUPERSCRIPT )(2)
hv=tanh(Wvfv+bv)superscript𝑣superscript𝑊𝑣superscript𝑓𝑣superscript𝑏𝑣h^{v}=\tanh{(W^{v}f^{v}+b^{v})}italic_h start_POSTSUPERSCRIPT italic_v end_POSTSUPERSCRIPT = roman_tanh ( italic_W start_POSTSUPERSCRIPT italic_v end_POSTSUPERSCRIPT italic_f start_POSTSUPERSCRIPT italic_v end_POSTSUPERSCRIPT + italic_b start_POSTSUPERSCRIPT italic_v end_POSTSUPERSCRIPT )(3)
z=σ(Wz[ft;fv]+bz)𝑧𝜎superscript𝑊𝑧superscript𝑓𝑡superscript𝑓𝑣superscript𝑏𝑧z=\sigma(W^{z}[f^{t};f^{v}]+b^{z})italic_z = italic_σ ( italic_W start_POSTSUPERSCRIPT italic_z end_POSTSUPERSCRIPT [ italic_f start_POSTSUPERSCRIPT italic_t end_POSTSUPERSCRIPT ; italic_f start_POSTSUPERSCRIPT italic_v end_POSTSUPERSCRIPT ] + italic_b start_POSTSUPERSCRIPT italic_z end_POSTSUPERSCRIPT )(4)
h=zht+(1z)hv𝑧superscript𝑡1𝑧superscript𝑣h=z*h^{t}+(1-z)*h^{v}italic_h = italic_z ∗ italic_h start_POSTSUPERSCRIPT italic_t end_POSTSUPERSCRIPT + ( 1 - italic_z ) ∗ italic_h start_POSTSUPERSCRIPT italic_v end_POSTSUPERSCRIPT(5)
Θ={Wt,Wv,Wz}Θsuperscript𝑊𝑡superscript𝑊𝑣superscript𝑊𝑧\Theta=\{W^{t},W^{v},W^{z}\}roman_Θ = { italic_W start_POSTSUPERSCRIPT italic_t end_POSTSUPERSCRIPT , italic_W start_POSTSUPERSCRIPT italic_v end_POSTSUPERSCRIPT , italic_W start_POSTSUPERSCRIPT italic_z end_POSTSUPERSCRIPT }(6)

, where hhitalic_h is a weighted combination of the textual and visual information htsuperscript𝑡h^{t}italic_h start_POSTSUPERSCRIPT italic_t end_POSTSUPERSCRIPT and hvsuperscript𝑣h^{v}italic_h start_POSTSUPERSCRIPT italic_v end_POSTSUPERSCRIPT respectively. ΘΘ\Thetaroman_Θ denotes the parameters to be learned, while [;][;][ ; ] indicates the concatenation operation.

Finally, we pass hhitalic_h through a dense layer consisting of two units for getting the final prediction.

Our introduced model is illustrated in Fig.1(b).

V-C Crossmodal Attention

In terms of the textual modality, we pass the user description field through a TwHIN-BERT model. Let ftN×dtsuperscript𝑓𝑡superscript𝑁subscript𝑑𝑡f^{t}\in\mathbb{R}^{N\times d_{t}}italic_f start_POSTSUPERSCRIPT italic_t end_POSTSUPERSCRIPT ∈ blackboard_R start_POSTSUPERSCRIPT italic_N × italic_d start_POSTSUBSCRIPT italic_t end_POSTSUBSCRIPT end_POSTSUPERSCRIPT denote the textual representation. N𝑁Nitalic_N indicates the sequence length, while dtsubscript𝑑𝑡d_{t}italic_d start_POSTSUBSCRIPT italic_t end_POSTSUBSCRIPT denotes the dimensionality and is equal to 768. In terms of the visual modality, we create a 3d image as described in SectionIV. Next:

  • In terms of the Cresci’17 dataset, we exploit our best performing model, namely VGG16, as mentioned in SectionIV-A3. Specifically, we get the output of the last CNN layer (after max pooling) as the output of the pretrained VGG16 model. Let fvT×dvsuperscript𝑓𝑣superscript𝑇subscript𝑑𝑣f^{v}\in\mathbb{R}^{T\times d_{v}}italic_f start_POSTSUPERSCRIPT italic_v end_POSTSUPERSCRIPT ∈ blackboard_R start_POSTSUPERSCRIPT italic_T × italic_d start_POSTSUBSCRIPT italic_v end_POSTSUBSCRIPT end_POSTSUPERSCRIPT represent the visual representation vector. T𝑇Titalic_T and dvsubscript𝑑𝑣d_{v}italic_d start_POSTSUBSCRIPT italic_v end_POSTSUBSCRIPT are equal to 64 and 512 respectively. Next, we pass fvsuperscript𝑓𝑣f^{v}italic_f start_POSTSUPERSCRIPT italic_v end_POSTSUPERSCRIPT through a dense layer consisting of 768 units.

  • In terms of the TwiBot-20 dataset, we exploit our best performing model, namely AlexNet, as mentioned in SectionIV-A3. Specifically, we get the output of the last CNN layer (after max pooling) as the output of the pretrained AlexNet model. Let fvT×dvsuperscript𝑓𝑣superscript𝑇subscript𝑑𝑣f^{v}\in\mathbb{R}^{T\times d_{v}}italic_f start_POSTSUPERSCRIPT italic_v end_POSTSUPERSCRIPT ∈ blackboard_R start_POSTSUPERSCRIPT italic_T × italic_d start_POSTSUBSCRIPT italic_v end_POSTSUBSCRIPT end_POSTSUPERSCRIPT represent the visual representation vector. T𝑇Titalic_T and dvsubscript𝑑𝑣d_{v}italic_d start_POSTSUBSCRIPT italic_v end_POSTSUBSCRIPT are equal to 49 and 256 respectively. Next, we pass fvsuperscript𝑓𝑣f^{v}italic_f start_POSTSUPERSCRIPT italic_v end_POSTSUPERSCRIPT through a dense layer consisting of 768 units.

Motivated by [57, 58, 59], we exploit two crossmodal attention layers, i.e., one from textual ftsuperscript𝑓𝑡f^{t}italic_f start_POSTSUPERSCRIPT italic_t end_POSTSUPERSCRIPT to visual features fvsuperscript𝑓𝑣f^{v}italic_f start_POSTSUPERSCRIPT italic_v end_POSTSUPERSCRIPT and one from visual to textual features.

Specifically, we calculate the scaled dot attention [60] as follows:

z=softmax(QKTd)V𝑧𝑠𝑜𝑓𝑡𝑚𝑎𝑥𝑄superscript𝐾𝑇𝑑𝑉z=softmax\left(\frac{QK^{T}}{\sqrt{d}}\right)Vitalic_z = italic_s italic_o italic_f italic_t italic_m italic_a italic_x ( divide start_ARG italic_Q italic_K start_POSTSUPERSCRIPT italic_T end_POSTSUPERSCRIPT end_ARG start_ARG square-root start_ARG italic_d end_ARG end_ARG ) italic_V(7)

, where the textual modality corresponds to the query (Q), while the visual modality corresponds to the key (K) and value (V).

Similarly, we design another scaled dot attention layer, where the visual modality corresponds to the query (Q), while the textual modality corresponds to the key (K) and value (V). Let y𝑦yitalic_y be the output of the scaled dot attention layer.

Next, we concatenate z𝑧zitalic_z and y𝑦yitalic_y and pass the resulting matrix into a global average pooling layer.

Finally, we exploit a dense layer consisting of two units for getting the final output.

Our introduced model is illustrated in Fig.1(c).

VI Experiments

VI-A Baselines

  • Comparison with state-of-the-art approaches

    • Cresci’17 Dataset

      • *

        DeepSBD [27]: This method extracts profile, temporal, activity, and content information. Finally, a deep neural network is trained consisting of LSTMs, CNNs, and an attention layer.

      • *

        DNNBD [10]: This method uses only profile information along with SMOTE. We report the results reported in [27].

      • *

        DBDM [28]: This method models social behavior and content information. A deep neural network consisting of CNNs and BiLSTMs is trained. We report the results reported in [27].

      • *

        DeBD [61]: This method passes the tweet content and the relationship between them into a CNN. Secondly, it uses LSTM to extract the potential temporal features of the tweet metadata. Finally, the authors concatenate the temporal features with the joint content features for detecting social bots. We report the results reported in [27].

      • *

        MulBot-Glob_Hier [37]: This method extracts multidimensional temporal features from user’s timeline, employs dimensionality reduction algorithms, extracts global features, and performs the Agglomerative Hierarchical clustering algorithm.

      • *

        DNA - sequence (supervised) [14]: This method extracts the digital DNA sequence per user by using either the type or content of tweets. Then, the authors leverage the longest common substring (LCS) curves for detecting the social spambots.

      • *

        Ahmed_DBSCAN [37, 62]: This method adopts the approach proposed by [62] and uses DBSCAN to increase the performance. We report the results reported in [37].

      • *

        Ahmed and Abulaish [62]: This approach exploits the Euclidean distance between feature vectors to create a similarity graph of the accounts. Next, graph clustering and community detection algorithms are used for identifying groups of similar accounts in the graph. We report the results reported in [14].

      • *

        Botometer [18]: Botometer is a publicly available bot detection tool via the website, Python, or REST APIs. This approach is trained with more than 1,000 features using a Random Forest classifier.

    • TwiBot-20 dataset

      • *

        Kudugunta and Ferrara [10]: This method utilizes user metadata features and tweet content features for identifying bots.

      • *

        Wei and Nguyen [31]: This method uses GloVe embeddings and trains a deep neural network consisting of three layers of BiLSTMs.

      • *

        Miller et al. [38]: Bot detection task is considered as an anomaly detection problem. Specifically, this method extracts 107 features and modifies two stream clustering algorithms, namely StreamKM++ and DenStream.

      • *

        Cresci et al. [13]: This method extracts the digital DNA sequence per user by using either the type or content of tweets. Then, the authors leverage the longest common substring (LCS) curves for detecting bots in groups.

      • *

        Botometer [18].

      • *

        Alhosseini et al. [36]: This method extracts a set of features per user, including age, favourites_count, statuses_count, friends_count, followers_count, and account length name. Then, a Graph Convolutional Neural Network is trained.

  • TwHIN-BERT using only the user description field

  • Our best performing model described in SectionIV

VI-B Experimental Setup

All the details are provided in SectionIV-A1. We use the TwHIN-BERT-base version from the Transformers library in Python [63].

VI-C Evaluation Metrics

We use the evaluation metrics, which are described in SectionIV-A2.

VII Results

The results of our proposed approaches are reported in Tables VI and VII. Specifically, TableVI reports the results on the Cresci’17 dataset, while TableVII reports the results on the TwiBot-20 dataset.

VII-A Cresci’17 Dataset

The results of our proposed approaches described in SectionV are reported in TableVI.

Evaluation metrics
ArchitecturePrecisionRecallF1-scoreAccuracySpecificity
Comparison with state-of-the-art approaches
DeepSBD [27]100.00-99.8199.83-
DNNBD [10]77.66-75.6378.20-
DBDM [28]100.00-98.8299.32-
DeBD [61]97.73-97.5997.74-
MulBot-Glob_Hier [37]99.5099.5099.0099.30-
DNA - sequence (supervised) [14]98.2097.7097.7097.7098.10
Ahmed_DBSCAN [37, 62]93.0093.0093.0092.80-
Ahmed and Abulaish [62]94.5094.4094.4094.3094.50
Botometer [18]--97.3195.97-
Unimodal approaches (only user description)
TwHIN-BERT99.5999.1899.3899.3799.56
±plus-or-minus\pm±0.50±plus-or-minus\pm±0.75±plus-or-minus\pm±0.39±plus-or-minus\pm±0.40±plus-or-minus\pm±0.54
Unimodal approaches (only images)
VGG16 (type of tweets)99.7899.5599.6799.6899.80
±plus-or-minus\pm±0.44±plus-or-minus\pm±0.54±plus-or-minus\pm±0.44±plus-or-minus\pm±0.42±plus-or-minus\pm±0.41
VGG16 (content of tweets)100.0099.7899.8999.89100.00
±plus-or-minus\pm±0.00±plus-or-minus\pm±0.43±plus-or-minus\pm±0.22±plus-or-minus\pm±0.21±plus-or-minus\pm±0.00
Proposed Transformer-based models (images based on the type of tweets)
TwHIN-BERT + VGG1699.7899.3499.5699.5899.80
(Concatenation)±plus-or-minus\pm±0.44±plus-or-minus\pm±0.88±plus-or-minus\pm±0.54±plus-or-minus\pm±0.52±plus-or-minus\pm±0.41
TwHIN-BERT + VGG1699.7999.7999.7999.7999.78
(GMU)±plus-or-minus\pm±0.42±plus-or-minus\pm±0.42±plus-or-minus\pm±0.42±plus-or-minus\pm±0.42±plus-or-minus\pm±0.43
TwHIN-BERT + VGG16100.0099.7999.9099.89100.00
(Cross-Modal Attention)±plus-or-minus\pm±0.00±plus-or-minus\pm±0.41±plus-or-minus\pm±0.21±plus-or-minus\pm±0.21±plus-or-minus\pm±0.00
Proposed Transformer-based models (images based on the content of the tweets)
TwHIN-BERT + VGG1699.7799.7799.7799.7999.80
(Concatenation)±plus-or-minus\pm±0.46±plus-or-minus\pm±0.46±plus-or-minus\pm±0.46±plus-or-minus\pm±0.42±plus-or-minus\pm±0.39
TwHIN-BERT + VGG16100.0099.5899.7999.79100.00
(GMU)±plus-or-minus\pm±0.00±plus-or-minus\pm±0.52±plus-or-minus\pm±0.26±plus-or-minus\pm±0.26±plus-or-minus\pm±0.00
TwHIN-BERT + VGG16100.0099.9699.9899.98100.00
(Cross-Modal Attention)±plus-or-minus\pm±0.00±plus-or-minus\pm±0.08±plus-or-minus\pm±0.04±plus-or-minus\pm±0.04±plus-or-minus\pm±0.00

Regarding our proposed approaches (transformer-based models based on the type of tweets), one can observe that TwHIN-BERT + VGG16 (Cross-Modal Attention) constitutes the best performing model achieving an Accuracy of 99.89%, a Precision of 100.00%, a Recall of 99.79%, a F1-score of 99.90%, and a Specificity of 100.00%. It outperforms the other introduced models in Accuracy by 0.10-0.31%, in Precision by 0.21-0.22%, in F1-score by 0.11-0.34%, and in Specificity by 0.20-0.22%. Also, it outperforms both TwHIN-BERT and VGG16 (type of tweet). Specifically, it surpasses TwHIN-BERT in Precision by 0.41%, in Recall by 0.61%, in F1-score by 0.52%, in Accuracy by 0.52%, and in Specificity by 0.44%. It outperforms also VGG16 (type of tweet) in Precision by 0.22%, in Recall by 0.24%, in F1-score by 0.23%, in Accuracy by 0.21%%, and in Specificity by 0.20%. Additionally, TwHIN-BERT + VGG16 (GMU) outperforms TwHIN-BERT + VGG16 (Concatenation) in Precision by 0.01%, in Recall by 0.45%, in F1-score by 0.23%, and in Accuracy by 0.22%. Also, it outperforms both TwHIN-BERT and VGG16 (type of tweets) in Accuracy by 0.42% and 0.11% respectively. In terms of TwHIN-BERT + VGG16 (Concatenation), one can observe that this model obtains worse performance than VGG16 (type of tweets). Specifically, VGG16 (type of tweets) outperforms TwHIN-BERT (Concatenation) in Recall by 0.21%, in F1-score by 0.11%, and in Accuracy by 0.10%. We speculate that this decrease in performance is attributable to the concatenation operation, which assigns equal importance to each modality ignoring the inherent correlations between the two modalities. Overall, we believe that TwHIN-BERT + VGG16 (Cross-Modal Attention) obtains better performance than the other two multimodal models, since it captures the crossmodal interactions. On the other hand, the Gated Multimodal Unit controls the information flow from each modality, while the concatenation operation neglects the inherent correlations.

In terms of our proposed transformer-based models (images based on the content of tweets), one can observe that TwHIN-BERT + VGG16 (Cross-Modal Attention) constitutes the best performing model outperforming both TwHIN-BERT + VGG16 (GMU) and TwHIN-BERT + VGG16 (Concatenation). Specifically, TwHIN-BERT + VGG16 (Cross-Modal Attention) outperforms TwHIN-BERT + VGG16 (GMU) in Recall by 0.42%, in F1-score by 0.19%, and in Accuracy by 0.19%. Although equal Precision scores are achieved, TwHIN-BERT + VGG16 (Cross-Modal Attention) yields a better F1-score, which constitutes the weighted average of Precision and Recall. Additionally, TwHIN-BERT + VGG16 (Cross-Modal Attention) outperforms TwHIN-BERT + VGG16 (Concatenation) in Precision by 0.23%, in Recall by 0.19%, in F1-score by 0.21%, in Accuracy by 0.19%, and in Specificity by 0.20%. In comparison with unimodal approaches employing either text or images, we observe that TwHIN-BERT + VGG16 (Crossmodal Attention) outperforms both TwHIN-BERT and VGG16 (content of tweets). Finally, TwHIN-BERT + VGG16 (Cross-Modal Attention) with images based on the content of tweets outperforms all the introduced models exploiting images based on the type of tweets. Therefore, TwHIN-BERT + VGG16 (Cross-Modal Attention) with images based on the content of tweets constitutes our best performing model.

In comparison with the state-of-the-art approaches, one can observe that our best performing model, namely TwHIN-BERT + VGG16 (Cross-Modal Attention) with images based on the content of the tweets, outperforms these approaches in Precision by 0.50-22.34% (except for DeepSBD and DBDM), in Recall by 0.46-6.96%, in F1-score by 0.17-24.35%, in Accuracy by 0.15-21.78%, and in Specificity by 1.90-5.50%. Similarly, TwHIN-BERT + VGG16 (Cross-Modal Attention) with images based on the type of tweets, surpasses the existing research initiatives in Precision by 0.50-22.34% (except for DeepSBD and DBDM), in Recall by 0.29-6.79%, in F1-score by 0.09-24.27%, in Accuracy by 0.06-21.69%, and in Specificity by 1.90-5.50%. Although our best performing model outperforms DeepSBD by a small margin of 0.17% and 0.15% in F1-score and Accuracy respectively, our best performing model has multiple advantages over DeepSBD. Firstly, DeepSBD processes the tweets of each user, extracts GloVe embeddings, and creates a 3d matrix which is given as input to a 6-layer CNN increasing the computational demands. Additionally, this method extracts a set of features per user. On the contrary, our method seems to be simpler and more effective, since it does not rely on a feature extraction strategy; it relies only on the user description and the sequence of actions performed by the user. Additionally, the authors use GloVe embeddings instead of using a language model based on transformers. Thus, this approach inherits the limitations of the GloVe embeddings. In terms of the training time per epoch, 120 to 180 seconds are required for training by our model, while the training time of DeepSBD ranges from 248 to 634 seconds.

VII-B TwiBot-20 Dataset

The results of our proposed approaches described in SectionV are reported in TableVII.

In terms of the proposed transformer-based models using images based on the type of tweets posted by the users, one can observe that TwHIN - BERT + AlexNet (Cross-Modal Attention) constitutes the best performing model in terms of Recall, F1-score, and Accuracy. Specifically, it outperforms the other introduced multimodal models in Recall by 2.55-2.85%, in F1-score by 0.83-1.26%, and in Accuracy by 0.28-0.76%. Additionally, TwHIN-BERT + AlexNet (Cross-Modal Attention) with images based on the type of tweets, surpasses the performance achieved by unimodal models, i.e., TwHIN-BERT (using user description) and AlexNet (using images). Similar differences in models’ performance are observed, when the images based on the content of tweets are used as input to proposed approaches. To be more precise, the usage of the crossmodal attention as a fusion method yields the best results surpassing the performance achieved by concatenation and gated multimodal unit as fusion approaches. TwHIN-BERT + AlexNet (Cross-modal Attention) outperforms the other multimodal models in Recall, F1-score, and Accuracy by 3.16-4.43%, 1.26-2.03%, and 0.70-1.36% respectively. TwHIN-BERT + AlexNet (Cross-Modal Attention) outperforms also the unimodal models employing either text or images. Specifically, it surpasses TwHIN-BERT in F1-score and Accuracy by 3.18% and 3.25% respectively, while it also outperforms AlexNet in F1-score and Accuracy by 9.50% and 7.47% respectively. Overall, we observe that the usage of multiple modalities improves bots’ detection performance. Cross-modal Attention boosts performance, since it models the cross-modal interactions. On the other hand, the gated multimodal unit refers to a weighting strategy, which controls the contribution of each modality to the bots’ prediction task. The concatenation operation is not capable of capturing the interactions of the two modalities or controlling the contribution of each modality, since equal importance to text and image information is assigned.

In comparison with state-of-the-art approaches, our best performing model outperforms these approaches in Accuracy and F1-score by 3.40-26.73% and 0.97-65.58% respectively. We observe that although Botometer was capable of identifying the social spambots in Cresci’17 dataset, the performance drops significantly in TwiBot-20 dataset, which verifies the fact that bots in Twitter have changed their behavior and have managed to evade previous detection methods. Similarly, we observe that the other approaches employing feature extraction techniques fail to detect bots. On the contrary, our approach relying only on the user description and the sequence of actions performed by the user based on the content of tweets seems to be an effective method.

Evaluation metrics
ArchitecturePrecisionRecallF1-scoreAccuracySpecificity
Comparison with state-of-the-art approaches
Kudugunta and Ferrara [10]--47.2659.59-
Wei and Nguyen [31]--75.3371.26-
Miller et al. [38]--62.6648.01-
Cresci et al. [13]--10.7247.93-
Botometer [18]--48.9255.84-
Alhosseini et al. [36]--73.1868.13-
Unimodal approaches (only user description)
TwHIN-BERT71.4175.1273.1271.4167.41
±plus-or-minus\pm±1.64±plus-or-minus\pm±4.15±plus-or-minus\pm±1.61±plus-or-minus\pm±1.07±plus-or-minus\pm±3.76
Unimodal approaches (only images)
AlexNet (type of tweets)68.3365.8967.0566.4166.96
±plus-or-minus\pm±1.26±plus-or-minus\pm±2.05±plus-or-minus\pm±0.72±plus-or-minus\pm±0.62±plus-or-minus\pm±2.79
AlexNet (content of tweets)70.3663.6466.8067.1971.01
±plus-or-minus\pm±1.11±plus-or-minus\pm±1.53±plus-or-minus\pm±0.37±plus-or-minus\pm±0.28±plus-or-minus\pm±2.20
Proposed Transformer-based models (images based on the type of tweets)
TwHIN-BERT + AlexNet74.2275.0174.5273.4271.70
(Concatenation)±plus-or-minus\pm±1.93±plus-or-minus\pm±3.35±plus-or-minus\pm±0.88±plus-or-minus\pm±0.53±plus-or-minus\pm±3.98
TwHIN-BERT + AlexNet74.7075.3174.9573.9072.39
(GMU)±plus-or-minus\pm±1.56±plus-or-minus\pm±3.02±plus-or-minus\pm±1.26±plus-or-minus\pm±1.00±plus-or-minus\pm±2.91
TwHIN-BERT + AlexNet73.9877.8675.7874.1870.20
(Cross-Modal Attention)±plus-or-minus\pm±2.40±plus-or-minus\pm±3.02±plus-or-minus\pm±0.48±plus-or-minus\pm±0.78±plus-or-minus\pm±4.59
Proposed Transformer-based models (images based on the content of the tweets)
TwHIN-BERT + AlexNet74.3174.2674.2773.3072.27
(Concatenation)±plus-or-minus\pm±0.90±plus-or-minus\pm±1.39±plus-or-minus\pm±0.35±plus-or-minus\pm±0.29±plus-or-minus\pm±1.78
TwHIN-BERT + AlexNet74.7475.5375.0473.9672.27
(GMU)±plus-or-minus\pm±1.88±plus-or-minus\pm±3.58±plus-or-minus\pm±0.93±plus-or-minus\pm±0.31±plus-or-minus\pm±3.93
TwHIN-BERT + AlexNet74.1678.6976.3074.6671.61
(Cross-Modal Attention)±plus-or-minus\pm±1.14±plus-or-minus\pm±2.97±plus-or-minus\pm±0.83±plus-or-minus\pm±0.31±plus-or-minus\pm±3.16

VIII Qualitative and Error Analysis

In this section, we perform a qualitative and error analysis of the best performing model on the TwiBot-20 dataset, namely TwHIN-BERT + AlexNet (Cross-Modal Attention) using images based on the content of tweets. TableVIII reports the predictions on some sample instances. Specifically, rows 1 and 2 refer to correct classifications as real users. Similarly, rows 3-7 refer to correct predictions as bots. Rows 8-10 refer to incorrect predictions made by our best performing model. Specifically. our model predicts these instances as belonging to a bot, while the real label corresponds to a genuine account. Similarly, rows 11 and 12 indicate incorrect classifications, where the actual label is a bot, while the predicted label is a real user.

Regarding correct classifications corresponding to genuine accounts, we observe that the Digital DNA of a real user includes a great number of consecutive mentions. Thus, our model learns that consecutive mentions indicate a real user, since mentions indicate often interactions with other users. Additionally, we observe that the user description field refers to a simple description of the biography of the user.

In terms of the correct classifications referring to bots, we observe in row 3 that the digital DNA includes a lot of consecutive hashtags. Although Digital DNAs of rows 4-7 include a lot of mentions, we observe that the user description fields have common characteristics. Specifically, we observe that the users’ descriptions of rows 4-6 include many words separated by commas (,). The user description of row 7 includes many hashtags (8 in number) and emojis. Therefore, both user description and the activity of users are important towards the final prediction.

Next, we examine reasons of misclassifications, i.e., rows 8-10. Specifically, we hypothesize that our model misclassifies row 8 as bot, since the Digital DNA consists of a great number of consecutive URLs. Also, we believe that row 9 is predicted as bot, since the user description includes many mentions. Finally, we observe that row 10 is similar to rows 4-6, and thus this row is predicted falsely as bot. Specifically, we observe that the user description of row 10 includes many words separated by commas.

Finally, we investigate the reasons of misclassifications in rows 11 and 12. We observe that Digital DNAs contain a lot of consecutive mentions, which refer to a genuine account (see rows 1 and 2). This verifies the fact that bots have found ways to evade detection approaches and mimic human behavior.

IDUser DescriptionDigital DNARealPred.
1Sports Broadcaster & Event Host @skysportsMXMMMNUUNUMXUMUMMMUUMXXXXUMMMMMMMMMXXUXMMMMXUNUXXMMUMUMXUUXXXMMMUUXXUXXMUXUMMMUXMXXXUXUMMMUMXUUXUMMUXXXMMMXMUUUMMXUUMMMMMXXUUMMXUMMMUMMUXXMMMMUMUXXUUMMUMMUUUMMMUXXXUUNUUMXMMMXMXXUMXMUUUUXMMMMMUUMXMUMXReal UserReal User
2Former NASA and Apple engineer. Current YouTuber and friend of science. <URL>XMUMMMMXMXMMXXMXXMMXMXUMMXMMUXUUMMMMMMXMMMMMMMMMMXMMMMMUUUMMXMMMMXMMMXXUNXMMUMMMMMMMMUMUXMMNXXUUMXUMXMUMMMMMMMXMMUUUMMUMMMMMMMUUMMMUMMMMMMMMMNMNMUUXMMXMNMMXMMMMMXMMXUMUMMMMMMUUNXUMMMMMUMMMMUMMUUXXMXXReal UserReal User
3Fun, outgoing guy that loves to meet new people. Casual gamer with a competitive nature.HXXXXXMMXXMXMXMMXXHXXHHHHMXXXXXXXXXXXXXMXXXXXXXXUMHHXXUXMUXXXUXXXXXUHMHMXMMMMXXXHHHMHHHHHXHXHHHHXHHHHHHHHHHXHHHHHHMHXMMMMXHXHHHMHHHMXHNXHHHMMXXXXHMMMHHHHHHHHMHHHMXHMHHHNHHHHHMHHHHHHHHHHHHHUHHHXHHHXXHMBotBot
4progressive, vegan, animal rights, social justice, anti racism, resistance, anti trump, lock him up, the SDNY will get the whole crime family <raised fist emoji> #FBRMXMMMMXMMUXMMMMXUMMMMMMMMUMUMMMMUXMXM XMMMMMXMMMMMXMXMMMMMMMMMMMMMMMMUMMMMMM MUXMMXXUMUMMMMMXMMMMMMMMXXMMMMMXXUMMMM MMXMMXUMUMMMMMXMMMMMMMMMXMXMXUMMMMMUXM MMXMMMXMMMXMXMMUXMXMMMUMMMXMMMUMMXUXXM XMMMXMXMMMMBotBot
5wife, mom, gma, Dem, liberal, truth, justice, equality, free press, free speech, LGBTQ, unions, proChoice, singlePayer, science#Resistance, #VoteBluetoSaveAmericaMXMMMMXMMMMMMMUMMMMMMXMMMMMMXXMMHXMXXXMMMXMMMMMXXMXMMMMMXMMMXMMXMXXXXMMMMMMMMMMXMMMMMMXMMXMXMMUUMMMMUUXXMXMMXMXMMMMXMMMMMMMMMUMXMMXMMMMUUMUMMXMMUXMXMUM MMMMXMMMMUMXMMMMMXMMMMMMMMXMMMXMMMMMMMMMMXMXMMXXXBotBot
6Christian, Conservative, Army,Life member NRA,Believe in God my Savior and Lord.Believe in America and believe in our great State of TEXAS !!!Senator Cruz III%XMMMMXMXMXMXMMMMXMMNMMXMMMXMMXMMXXMXXM MXMMMMMMMMMMMMMMMMXMMMMMMMMMMMXMNMMMUMM MXMMMMMMMXMMMMMXMMMMMMUXMMMUMXMMMUMMMM XMXMMMMMMMXMXMMMMMMMMMMMMMMMXMMMMMMMMMM MXMXMMXMMMMMXMMMMMMMMXMMMBotBot
7I AM HERE 4 POTUS #GGOAT #MomentsMatter <glowing star emoji> #BELIEVING=PEACE <folded hands emoji> <paw prints emoji> #BeachBabeBum <desert island emoji> <eyes emoji> <fox emoji> #TrumpTrain <locomotive emoji> #MAGA #KAG #RedWomenRealWomen ArkTrumper <anchor emoji> SADLY_DisabledMMXXXMMMXXMMMMMMMXXMXMMMMMMMMMMMMMMMMMMXMXMMMMMNUXMMXMMMXMXMMMMMMMMXMXXXXXXMXNNNMUMXMMXXXXXXMMMXMMMMMXMMXXMMMMMMXXMMMMXXMMMMMMXMMXMMMXXMXXMMXXMMXMMMMMMMMMXXMMMMMMMMMMMMMMXXMMMXMMMMXMMMXMXXMMMXMMMMNMMXBotBot
8Pasta lover. I don’t tweet much. My new Netflix series Master of None is now streaming on Netflix. I wrote a book called Modern Romance.XXXXXXXXXUXXXUUUUUUUUUUUUUXUUUUUUUUUUUUUUUUUUUUUXUUUUUUUUUUUUUUUUUUNUXUXUUUUUUUUXXNNUXXHUUXXXNXXXXXUUUUMNXUUXMUUUUUUXUNXXUUXUUXXUUMXXUUNXXXUUUUXMXNNUUUUUMUUUUHUUUUUXUUUUU UUUUUUUUUUUUUUUUUUUUUMUUUUUUUUReal UserBot
9@mu_foundation @veteransgarage & @SSAFA ambassador. @helpforheroes @rmchcharity and @manchesterpride patron.XMUUXXMXUMXUUUUUMUMMNMUMUUMUMUU MUUUMMMNXUXMUMMNUM UMUXXUXXUUMMUUXXNUMMUXUXMXMXXMMXXUNUMMUUUUUMUUUUUM MXMNUMXUUUMXUUUUXXMMXMXMMUUNMMMMXMUUMMMMXNXMMMMMM MXUNMNNNMXXXUXNUNXMUMXXUUMMHUUUUNNXXMMMMMMXUMXUUNMNMReal UserBot
10news/media, music/guitar, ECB, EU, Frankfurt, Greece, politics. Opinions my own, not necessarily my employer’s (=ECB). RT’s often mean I agree, but not always.MXMXXXMXUMMMMUMMUMXMXXXUXXMNUMUUXXMUMXMMMMUXXXMMXMUXMNMMMUXMMXMMXMMMMMMMMMMXMMUMMMXUNNNNNMMXXXUMMMMXXMMUXMMXMXMMMXMMMMXMUMMMXMMXMMMXXXMMXMMMMXUXMXXXXMMXXMMXMXUXXMMMXXXMMMMMMMMMXNXMXXXXMMMXMMMMXM UXXXUUReal UserBot
11Political Booking Producer at @nbcnews | @todayshow @nbcnightlynewsMMMXNXNXMMMMXMMXMMMMMXMMMMXMMMMMMMMXMXXXXXMXMMMXMMXMMMXMMMMMMXMXXMMXMMMMMMXMMXXMMMMMXXMMXUMXXMMMMMMMXXMMM MMXMXHMMMMMMMMMMMMXMMMMXMMXXXXNXMMXXMMMXMMMXMMMMMMMMXMMMXNXMXXXMMXMXMM MXMMMXMXMMMXMMMMMMMXNMXMXBotReal User
12Lets Hope and pray for The BestMMMMMMXMMMMMMMMMMMMMMMMMMMMMMMMXMMMXMMUMMMMBotReal User

IX Discussion

IX-A Limitations

This study has some limitations. Firstly, we did not apply explainability approaches to explain the predictions made by our proposed approaches. Additionally, this study requires labelled data. On the contrary, self-supervised learning approaches address the issue of labels’ scarcity.

IX-B Scalability and Real-time detection

An often neglected concern among researchers introducing methods for identifying bots in Twitter is the reliance on the platform APIs [64]. Specifically, data must be retrieved from social media platforms, so as to ensure that the researchers perform the necessary experiments.

However, in 2023, Twitter decided to end free access to their APIs. As a result of this modification, several research projects333https://www.reuters.com/technology/elon-musks-x-restructuring-curtails-disinformation-research-spurs-legal-fears-2023-11-06/ have been cancelled or suspended for a period of time. After that, legal obligations in the bloc’s Digital Services Act (DSA)444https://techcrunch.com/2023/11/17/change-in-xs-terms-indicate-eu-researchers-will-get-api-access/ require larger platforms, including Meta, Twitter, and Google, to provide data access to external researchers doing public interest research on systemic risks. However, the access to Twitter APIs still remains uncertain.

Modifications also to the API affect the bot detection algorithms. For example, the field geo_enabled was removed in 2019 due to privacy reasons. Therefore, algorithms trained with this features, should be modified.

Data access is inextricably linked with scalability. Scalability refers to the analysis of streaming data with limited computing resources [65]. Specifically, the speed at which a method processes a group of accounts depends highly on the possibility for access to the Twitter API. For instance, although the construction of large graphs leads to detection of bots with increasing evaluation performance, fetching such information is not feasible. The study in [66] claims that the following methods are required for improving the scalability of bot detection algorithms:

  • Model compression and distillation

  • Incremental Learning and online algorithms

  • Parallel and distributed processing

  • Stream-based processing and data reduction

IX-C Generalization of our approach on other social media

The introduced approaches in this paper can be generalized on other social media, including Facebook and Reddit. Similar to Twitter, users on the Facebook platform make posts, reposts, and reply to other users. At the same time, their posts may include mentions to other users, hashtags, and URLs. Additionally, each user may include a description field in his/her profile.

However, platforms’ policy about data access is critical. For instance, investigating bot detection approaches on the Facebook platform is a very difficult task due to the unwillingness of Facebook to share individual account data [64].

X Conclusion and Future Work

In this paper, we present the first study introducing multimodal and cross-modal models for detecting bots in Twitter by exploiting only the user description and 3d images, which represent the actions of each user. Firstly, we create two digital DNA sequences based on both the type and content of the tweets each user posts. Next, we apply a DNA-to-image conversion algorithm and create two 3d images per user based on the two digital DNA sequences. Finally, we fine-tune several pretrained models of the vision domain and show that VGG16 achieves the highest evaluation results. Next, we introduce multimodal and crossmodal models. First, we pass each user description field into a TwHIN-BERT and obtain a textual representation. We pass each 3d image through a VGG16 model and obtain a visual representation. Finally, we compare three methods for fusing the textual and visual representations, including the concatenation operation, the gated multimodal unit, and the cross-modal attention. Findings show that the crossmodal attention outperforms the other introduced approaches, be it either unimodal or multimodal approaches, obtaining also comparable performance with the state-of-the-art approaches.

In the future, we aim to apply our models in a federated learning framework. Applying also continual learning approaches is one of our future plans. Finally, we plan to exploit explainability techniques for rendering the introduced approaches explainable.

References

  • [1]L.Ilias, S.Mouzakitis, and D.Askounis, “Calibration of transformer-based models for identifying stress and depression in social media,” IEEE Transactions on Computational Social Systems, vol.11, no.2, pp. 1979–1990, 2024.
  • [2]S.Cresci, “Detecting malicious social bots: story of a never-ending clash,” in Disinformation in Open Online Media: First Multidisciplinary International Symposium, MISDOOM 2019, Hamburg, Germany, February 27–March 1, 2019, Revised Selected Papers 1.Springer, 2020, pp. 77–88.
  • [3]A.Bessi and E.Ferrara, “Social bots distort the 2016 us presidential election online discussion,” First monday, vol.21, no. 11-7, 2016.
  • [4]F.Brachten, S.Stieglitz, L.Hofeditz, K.Kloppenborg, and A.Reimann, “Strategies and influence of social bots in a 2017 german state election-a case study on twitter,” arXiv preprint arXiv:1710.07562, 2017.
  • [5]M.Shafahi, L.Kempers, and H.Afsarmanesh, “Phishing through social bots on twitter,” in 2016 IEEE International Conference on Big Data (Big Data), 2016, pp. 3703–3712.
  • [6]C.Shao, G.L. Ciampaglia, O.Varol, K.-C. Yang, A.Flammini, and F.Menczer, “The spread of low-credibility content by social bots,” Nature communications, vol.9, no.1, pp. 1–9, 2018.
  • [7]J.Uyheng, D.Bellutta, and K.M. Carley, “Bots amplify and redirect hate speech in online discourse about racism during the covid-19 pandemic,” Social Media + Society, vol.8, no.3, p. 20563051221104749, 2022. [Online]. Available: https://doi.org/10.1177/20563051221104749
  • [8]L.Ilias and I.Roussaki, “Detecting malicious activity in twitter using deep learning techniques,” Applied Soft Computing, vol. 107, p. 107360, 2021. [Online]. Available: https://www.sciencedirect.com/science/article/pii/S1568494621002830
  • [9]S.Ouni, F.Fkih, and M.N. Omri, “Bots and gender detection on twitter using stylistic features,” in International Conference on Computational Collective Intelligence.Springer, 2022, pp. 650–660.
  • [10]S.Kudugunta and E.Ferrara, “Deep neural networks for bot detection,” Information Sciences, vol. 467, pp. 312–322, 2018. [Online]. Available: https://www.sciencedirect.com/science/article/pii/S0020025518306248
  • [11]K.Hayawi, S.Mathew, N.Venugopal, M.M. Masud, and P.-H. Ho, “Deeprobot: a hybrid deep neural network model for social bot detection based on user profile data,” Social Network Analysis and Mining, vol.12, no.1, p.43, 2022.
  • [12]S.Feng, H.Wan, N.Wang, and M.Luo, “Botrgcn: Twitter bot detection with relational graph convolutional networks,” in Proceedings of the 2021 IEEE/ACM International Conference on Advances in Social Networks Analysis and Mining, ser. ASONAM ’21.New York, NY, USA: Association for Computing Machinery, 2022, p. 236–239. [Online]. Available: https://doi.org/10.1145/3487351.3488336
  • [13]S.Cresci, R.DiPietro, M.Petrocchi, A.Spognardi, and M.Tesconi, “Dna-inspired online behavioral modeling and its application to spambot detection,” IEEE Intelligent Systems, vol.31, no.5, pp. 58–64, 2016.
  • [14]S.Cresci, R.D. Pietro, M.Petrocchi, A.Spognardi, and M.Tesconi, “Social fingerprinting: Detection of spambot groups through dna-inspired behavioral modeling,” IEEE Transactions on Dependable and Secure Computing, vol.15, no.4, pp. 561–576, 2018.
  • [15]E.DiPaolo, M.Petrocchi, and A.Spognardi, “From online behaviours toimages: A novel approach tosocial bot detection,” in Computational Science – ICCS 2023, J.Mikyška, C.deMulatier, M.Paszynski, V.V. Krzhizhanovskaya, J.J. Dongarra, and P.M. Sloot, Eds.Cham: Springer Nature Switzerland, 2023, pp. 593–607.
  • [16]X.Zhang, Y.Malkov, O.Florez, S.Park, B.McWilliams, J.Han, and A.El-Kishky, “Twhin-bert: A socially-enriched pre-trained language model for multilingual tweet representations,” arXiv preprint arXiv:2209.07562, 2022.
  • [17]S.B. Jr, G.F.C. Campos, G.M. Tavares, R.A. Igawa, M.L.P. Jr, and R.C. Guido, “Detection of human, legitimate bot, and malicious bot in online social networks based on wavelets,” ACM Trans. Multimedia Comput. Commun. Appl., vol.14, no.1s, mar 2018. [Online]. Available: https://doi.org/10.1145/3183506
  • [18]C.A. Davis, O.Varol, E.Ferrara, A.Flammini, and F.Menczer, “Botornot: A system to evaluate social bots,” in Proceedings of the 25th International Conference Companion on World Wide Web, ser. WWW ’16 Companion.Republic and Canton of Geneva, CHE: International World Wide Web Conferences Steering Committee, 2016, p. 273–274. [Online]. Available: https://doi.org/10.1145/2872518.2889302
  • [19]E.Arin and M.Kutlu, “Deep learning based social bot detection on twitter,” IEEE Transactions on Information Forensics and Security, vol.18, pp. 1763–1772, 2023.
  • [20]S.Najari, M.Salehi, and R.Farahbakhsh, “Ganbot: A gan-based framework for social bot detection,” Social Network Analysis and Mining, vol.12, pp. 1–11, 2022.
  • [21]L.Yu, W.Zhang, J.Wang, and Y.Yu, “Seqgan: Sequence generative adversarial nets with policy gradient,” in Proceedings of the Thirty-First AAAI Conference on Artificial Intelligence, ser. AAAI’17.AAAI Press, 2017, p. 2852–2858.
  • [22]B.Wu, L.Liu, Y.Yang, K.Zheng, and X.Wang, “Using improved conditional generative adversarial networks to detect social bots on twitter,” IEEE Access, vol.8, pp. 36 664–36 680, 2020.
  • [23]D.Martín-Gutiérrez, G.Hernández-Peñaloza, A.B. Hernández, A.Lozano-Diez, and F.Álvarez, “A deep learning approach for robust detection of bots in twitter using transformers,” IEEE Access, vol.9, pp. 54 591–54 601, 2021.
  • [24]V.Chawla and Y.Kapoor, “A hybrid framework for bot detection on twitter: Fusing digital dna with bert,” Multimedia Tools and Applications, pp. 1–24, 2023.
  • [25]A.S. Alhassun and M.A. Rassam, “A combined text-based and metadata-based deep-learning framework for the detection of spam accounts on the social media platform twitter,” Processes, vol.10, no.3, 2022. [Online]. Available: https://www.mdpi.com/2227-9717/10/3/439
  • [26]Y.Wu, Y.Fang, S.Shang, J.Jin, L.Wei, and H.Wang, “A novel framework for detecting social bots with deep neural networks and active learning,” Knowledge-Based Systems, vol. 211, p. 106525, 2021. [Online]. Available: https://www.sciencedirect.com/science/article/pii/S0950705120306547
  • [27]M.Fazil, A.K. Sah, and M.Abulaish, “Deepsbd: A deep neural network model with attention mechanism for socialbot detection,” IEEE Transactions on Information Forensics and Security, vol.16, pp. 4211–4223, 2021.
  • [28]C.Cai, L.Li, and D.Zeng, “Detecting social bots by jointly modeling deep behavior and content information,” in Proceedings of the 2017 ACM on Conference on Information and Knowledge Management, ser. CIKM ’17.New York, NY, USA: Association for Computing Machinery, 2017, p. 1995–1998. [Online]. Available: https://doi.org/10.1145/3132847.3133050
  • [29]F.Wei and U.T. Nguyen, “Twitter bot detection using neural networks and linguistic embeddings,” IEEE Open Journal of the Computer Society, pp. 1–12, 2023.
  • [30]F.Wei and T.Nguyen, “A lightweight deep neural model for sms spam detection,” in 2020 International Symposium on Networks, Computers and Communications (ISNCC), 2020, pp. 1–6.
  • [31]F.Wei and U.T. Nguyen, “Twitter bot detection using bidirectional long short-term memory neural networks and word embeddings,” in 2019 First IEEE International Conference on Trust, Privacy and Security in Intelligent Systems and Applications (TPS-ISA), 2019, pp. 101–109.
  • [32]S.Feng, Z.Tan, R.Li, and M.Luo, “Heterogeneity-aware twitter bot detection with relational graph transformers,” Proceedings of the AAAI Conference on Artificial Intelligence, vol.36, no.4, pp. 3977–3985, Jun. 2022. [Online]. Available: https://ojs.aaai.org/index.php/AAAI/article/view/20314
  • [33]Z.Lei, H.Wan, W.Zhang, S.Feng, Z.Chen, J.Li, Q.Zheng, and M.Luo, “Bic: Twitter bot detection with text-graph interaction and semantic consistency,” 2023.
  • [34]E.Alothali, M.Salih, K.Hayawi, and H.Alashwal, “Bot-mgat: A transfer learning model based on a multi-view graph attention network to detect social bots,” Applied Sciences, vol.12, no.16, 2022. [Online]. Available: https://www.mdpi.com/2076-3417/12/16/8117
  • [35]H.Peng, Y.Zhang, H.Sun, X.Bai, Y.Li, and S.Wang, “Domain-aware federated social bot detection with multi-relational graph neural networks,” in 2022 International Joint Conference on Neural Networks (IJCNN), 2022, pp. 1–8.
  • [36]S.AliAlhosseini, R.BinTareaf, P.Najafi, and C.Meinel, “Detect me if you can: Spam bot detection using inductive representation learning,” in Companion Proceedings of The 2019 World Wide Web Conference, ser. WWW ’19.New York, NY, USA: Association for Computing Machinery, 2019, p. 148–153. [Online]. Available: https://doi.org/10.1145/3308560.3316504
  • [37]L.Mannocci, S.Cresci, A.Monreale, A.Vakali, and M.Tesconi, “Mulbot: Unsupervised bot detection based on multivariate time series,” in 2022 IEEE International Conference on Big Data (Big Data).Los Alamitos, CA, USA: IEEE Computer Society, dec 2022, pp. 1485–1494. [Online]. Available: https://doi.ieeecomputersociety.org/10.1109/BigData55660.2022.10020363
  • [38]Z.Miller, B.Dickinson, W.Deitrick, W.Hu, and A.H. Wang, “Twitter spammer detection using data stream clustering,” Information Sciences, vol. 260, pp. 64–73, 2014. [Online]. Available: https://www.sciencedirect.com/science/article/pii/S0020025513008037
  • [39]Y.Yang, R.Yang, Y.Li, K.Cui, Z.Yang, Y.Wang, J.Xu, and H.Xie, “Rosgas: Adaptive social bot detection with reinforced self-supervised gnn architecture search,” ACM Trans. Web, vol.17, no.3, may 2023. [Online]. Available: https://doi.org/10.1145/3572403
  • [40]G.Lingam, R.R. Rout, and D.V. Somayajulu, “Adaptive deep q-learning model for detecting social bots and influential users in online social networks,” Applied Intelligence, vol.49, pp. 3947–3964, 2019.
  • [41]R.R. Rout, G.Lingam, and D.V. L.N. Somayajulu, “Detection of malicious social bots using learning automata with url features in twitter network,” IEEE Transactions on Computational Social Systems, vol.7, no.4, pp. 1004–1018, 2020.
  • [42]S.Cresci, R.DiPietro, M.Petrocchi, A.Spognardi, and M.Tesconi, “The paradigm-shift of social spambots: Evidence, theories, and tools for the arms race,” in Proceedings of the 26th International Conference on World Wide Web Companion, ser. WWW ’17 Companion.Republic and Canton of Geneva, CHE: International World Wide Web Conferences Steering Committee, 2017, p. 963–972. [Online]. Available: https://doi.org/10.1145/3041021.3055135
  • [43]M.Avvenuti, S.Bellomo, S.Cresci, M.N. LaPolla, and M.Tesconi, “Hybrid crowdsensing: A novel paradigm to combine the strengths of opportunistic and participatory crowdsensing,” in Proceedings of the 26th International Conference on World Wide Web Companion, ser. WWW ’17 Companion.Republic and Canton of Geneva, CHE: International World Wide Web Conferences Steering Committee, 2017, p. 1413–1421. [Online]. Available: https://doi.org/10.1145/3041021.3051155
  • [44]S.Feng, H.Wan, N.Wang, J.Li, and M.Luo, “Twibot-20: A comprehensive twitter bot detection benchmark,” in Proceedings of the 30th ACM International Conference on Information & Knowledge Management, ser. CIKM ’21.New York, NY, USA: Association for Computing Machinery, 2021, p. 4485–4494. [Online]. Available: https://doi.org/10.1145/3459637.3482019
  • [45]C.Szegedy, W.Liu, Y.Jia, P.Sermanet, S.Reed, D.Anguelov, D.Erhan, V.Vanhoucke, and A.Rabinovich, “Going deeper with convolutions,” in 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2015, pp. 1–9.
  • [46]K.He, X.Zhang, S.Ren, and J.Sun, “Deep residual learning for image recognition,” in 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2016, pp. 770–778.
  • [47]S.Zagoruyko and N.Komodakis, “Wide residual networks,” arXiv preprint arXiv:1605.07146, 2016.
  • [48]A.Krizhevsky, “One weird trick for parallelizing convolutional neural networks,” arXiv preprint arXiv:1404.5997, 2014.
  • [49]F.N. Iandola, S.Han, M.W. Moskewicz, K.Ashraf, W.J. Dally, and K.Keutzer, “Squeezenet: Alexnet-level accuracy with 50x fewer parameters and< 0.5 mb model size,” arXiv preprint arXiv:1602.07360, 2016.
  • [50]G.Huang, Z.Liu, L.Van DerMaaten, and K.Q. Weinberger, “Densely connected convolutional networks,” in 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2017, pp. 2261–2269.
  • [51]M.Sandler, A.Howard, M.Zhu, A.Zhmoginov, and L.-C. Chen, “Mobilenetv2: Inverted residuals and linear bottlenecks,” in 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2018, pp. 4510–4520.
  • [52]S.Xie, R.Girshick, P.Dollár, Z.Tu, and K.He, “Aggregated residual transformations for deep neural networks,” in 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2017, pp. 5987–5995.
  • [53]K.Simonyan and A.Zisserman, “Very deep convolutional networks for large-scale image recognition,” arXiv preprint arXiv:1409.1556, 2014.
  • [54]M.Tan and Q.Le, “EfficientNet: Rethinking model scaling for convolutional neural networks,” in Proceedings of the 36th International Conference on Machine Learning, ser. Proceedings of Machine Learning Research, K.Chaudhuri and R.Salakhutdinov, Eds., vol.97.PMLR, 09–15 Jun 2019, pp. 6105–6114. [Online]. Available: https://proceedings.mlr.press/v97/tan19a.html
  • [55]A.Paszke, S.Gross, F.Massa, A.Lerer, J.Bradbury, G.Chanan, T.Killeen, Z.Lin, N.Gimelshein, L.Antiga, A.Desmaison, A.Kopf, E.Yang, Z.DeVito, M.Raison, A.Tejani, S.Chilamkurthy, B.Steiner, L.Fang, J.Bai, and S.Chintala, “Pytorch: An imperative style, high-performance deep learning library,” in Advances in Neural Information Processing Systems 32, H.Wallach, H.Larochelle, A.Beygelzimer, F.d'Alché-Buc, E.Fox, and R.Garnett, Eds.Curran Associates, Inc., 2019, pp. 8024–8035. [Online]. Available: http://papers.neurips.cc/paper/9015-pytorch-an-imperative-style-high-performance-deep-learning-library.pdf
  • [56]J.Arevalo, T.Solorio, M.Montes-y Gomez, and F.A. González, “Gated multimodal networks,” Neural Computing and Applications, pp. 1–20, 2020.
  • [57]Y.-H.H. Tsai, S.Bai, P.P. Liang, J.Z. Kolter, L.-P. Morency, and R.Salakhutdinov, “Multimodal transformer for unaligned multimodal language sequences,” in Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics.Florence, Italy: Association for Computational Linguistics, Jul. 2019, pp. 6558–6569. [Online]. Available: https://aclanthology.org/P19-1656
  • [58]L.Ilias, D.Askounis, and J.Psarras, “Detecting dementia from speech and transcripts using transformers,” Computer Speech & Language, vol.79, p. 101485, 2023. [Online]. Available: https://www.sciencedirect.com/science/article/pii/S0885230823000049
  • [59]D.SánchezVillegas and N.Aletras, “Point-of-interest type prediction using text and images,” in Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing.Online and Punta Cana, Dominican Republic: Association for Computational Linguistics, Nov. 2021, pp. 7785–7797. [Online]. Available: https://aclanthology.org/2021.emnlp-main.614
  • [60]A.Vaswani, N.Shazeer, N.Parmar, J.Uszkoreit, L.Jones, A.N. Gomez, L.Kaiser, and I.Polosukhin, “Attention is all you need,” in Proceedings of the 31st International Conference on Neural Information Processing Systems, ser. NIPS’17.Red Hook, NY, USA: Curran Associates Inc., 2017, p. 6000–6010.
  • [61]H.Ping and S.Qin, “A social bots detection model based on deep learning algorithm,” in 2018 IEEE 18th International Conference on Communication Technology (ICCT), 2018, pp. 1435–1439.
  • [62]F.Ahmed and M.Abulaish, “A generic statistical approach for spam detection in online social networks,” Computer Communications, vol.36, no.10, pp. 1120–1129, 2013. [Online]. Available: https://www.sciencedirect.com/science/article/pii/S0140366413001047
  • [63]T.Wolf, L.Debut, V.Sanh, J.Chaumond, C.Delangue, A.Moi, P.Cistac, T.Rault, R.Louf, M.Funtowicz, J.Davison, S.Shleifer, P.von Platen, C.Ma, Y.Jernite, J.Plu, C.Xu, T.L. Scao, S.Gugger, M.Drame, Q.Lhoest, and A.M. Rush, “Transformers: State-of-the-art natural language processing,” in Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing: System Demonstrations.Online: Association for Computational Linguistics, Oct. 2020, pp. 38–45. [Online]. Available: https://www.aclweb.org/anthology/2020.emnlp-demos.6
  • [64]K.-C. Yang, O.Varol, A.C. Nwala, M.Sayyadiharikandeh, E.Ferrara, A.Flammini, and F.Menczer, “Social bots: Detection and challenges. in: Yasseri, t. (ed.),” in Handbook of Computational Social Science.Edward Elgar Publishing Ltd, 2024.
  • [65]K.-C. Yang, O.Varol, P.-M. Hui, and F.Menczer, “Scalable and generalizable social bot detection through data selection,” Proceedings of the AAAI Conference on Artificial Intelligence, vol.34, no.01, pp. 1096–1103, Apr. 2020. [Online]. Available: https://ojs.aaai.org/index.php/AAAI/article/view/5460
  • [66]E.Ferrara, “Social bot detection in the age of chatgpt: Challenges and opportunities,” First Monday, vol.28, no.6, Jun. 2023. [Online]. Available: https://firstmonday.org/ojs/index.php/fm/article/view/13185

XI Biography Section

Multimodal Detection of Bots on X (Twitter) using Transformers (4)Loukas Ilias received the integrated master’s degreefrom the School of Electrical and Computer Engineering (SECE), National Technical University ofAthens (NTUA), Athens, Greece, in June 2020,where he is currently pursuing the Ph.D. degree withthe Decision Support Systems (DSS) Laboratory,SECE. He has completed a Research Internshipwith University College London (UCL), London,U.K.He is a Researcher with the DSS Laboratory,NTUA, where he is involved in EU-funded researchprojects. He has published in numerous journals, including IEEE Journal of Biomedical and Health Informatics, IEEE Transactions on Computational Social Systems, Knowledge-Based Systems (Elsevier), Expert Systems With Applications (Elsevier), Applied Soft Computing (Elsevier), Online Social Networks and Media (Elsevier), Computer Speech andLanguage (Elsevier), IEEE Access, Frontiers in Aging Neuroscience, and Frontiers in Big Data. Hisresearch has also been accepted for presentation at international conferences,including the IEEE-EMBS International Conference on Biomedical and Health Informatics (BHI’22)and the IEEE International Conference on Acoustics, Speech, and SignalProcessing (ICASSP) 2023. His research interests include speech processing,natural language processing, social media analysis, and the detection ofcomplex brain disorders.
Multimodal Detection of Bots on X (Twitter) using Transformers (5)Ioannis Michail Kazelidis received the integrated master’s degreefrom the School of Electrical and Computer Engineering (SECE), National Technical University ofAthens (NTUA), Athens, Greece, in October 2023. His research interests include deep learning and social media analysis.
Multimodal Detection of Bots on X (Twitter) using Transformers (6)Dimitris Askounis was the Scientific Director ofover 50 European research projects in the aboveareas (FP7, Horizon2020, and so on). For a numberof years, he was an Advisor to the Minister ofJustice and the Special Secretary for Digital Convergence for the introduction of information andcommunication technologies in public administration. Since June 2019, he has been the Presidentof the Information Society SA, Kallithea, Greece.He is currently a Professor at the School of Electrical and Computer Engineering, National TechnicalUniversity of Athens (NTUA), Athens, Greece, and the Deputy Directorof the Decision Support Systems Laboratory. He has over 25 years ofexperience in decision support systems, intelligent information systems andmanufacturing, e-business, e-government, open and linked data, big dataanalytics, Artificial Intelligence (AI) algorithms, and the application of modernInformation Technology (IT) techniques in the management of companies andorganizations.
Multimodal Detection of Bots on X (Twitter) using Transformers (2024)
Top Articles
Peaco*ck streaming: Pricing, plans, Olympics coverage, and how to sign up
Peaco*ck TV shows and movies: Everything you can watch on NBC’s streaming service
Pau.blaz
The Shoppes At Zion Directory
Ucsf Ilab
19 Awesome Things to Do in Redmond, Oregon
Ogre From Halloweentown
Haul auf deutsch: Was ist das? Übersetzung, Bedeutung, Erklärung - Bedeutung Online
Craigslist Carpet Installers
0.0Gomovies
Nala Ahegao
What You Need to Know About County Jails
Job Shop Hearthside Schedule
Barefoot Rentals Key Largo
Lucio Surf Code
Lake Charles, LA Houses and Single Family Homes For Rent | realtor.com®
Cooktopcove Com
Slmd Skincare Appointment
PNC Bank Review 2024
Lyons Prismhr
Does Publix Pharmacy Accept Sunshine Health
Staples Ups Drop Off
Cara In Creekmaw Code
Kirksey's Mortuary Obituaries
Tbom Genesis Retail Phone Number
Drys Pharmacy
Sam's Club Gas Price Spring Hill Fl
Susan Bowers Facebook
AC Filters | All About Air Filters for AC | HVAC Filters
Savannah Riverboat Cruise Anniversary Package
Tnt Tony Superfantastic
Core Relief Texas
2022 Jeep Grand Cherokee Lug Nut Torque
Rolling-Embers Reviews
Subway And Gas Station Near Me
Craigslist Used Motorhomes For Sale By Owner
Amazon Ups Drop Off Locations Near Me
M3Gan Showtimes Near Cinemark North Hills And Xd
The Nun 2 Showtimes Near Cinemark Towson And Xd
Oakly Rae Leaks
Why Larry the cat of 10 Downing Street wishes Starmer hadn’t won the election
Tamilrockers 2023 Tamil Movies Download Kuttymovies
Mensenlinq: Overlijdensberichten zoeken in 2024
ARK Fjordur: Ultimate Resource Guide | Where to Find All Materials - Games Fuze
Sun Massage Tucson Reviews
Telegram No Limit Groups - Telegram Rocket
Stock Hill Restaurant Week Menu
Grayson County Craigslist
Texture Ids For Custom Glove In Slap Battles
CareLink™ Personal Software | Medtronic
Neuer Extraction-Shooter auf Steam will Escape from Tarkov Konkurrenz machen, wird von echten Militär-Veteranen entwickelt
Fitgirl Starfield
Latest Posts
Article information

Author: Annamae Dooley

Last Updated:

Views: 6272

Rating: 4.4 / 5 (65 voted)

Reviews: 80% of readers found this page helpful

Author information

Name: Annamae Dooley

Birthday: 2001-07-26

Address: 9687 Tambra Meadow, Bradleyhaven, TN 53219

Phone: +9316045904039

Job: Future Coordinator

Hobby: Archery, Couponing, Poi, Kite flying, Knitting, Rappelling, Baseball

Introduction: My name is Annamae Dooley, I am a witty, quaint, lovely, clever, rich, sparkling, powerful person who loves writing and wants to share my knowledge and understanding with you.