metricas
covid
Buscar en
Journal of Innovation & Knowledge
Toda la web
Inicio Journal of Innovation & Knowledge Online cheaters: Profiles and motivations of internet users who falsify their da...
Información de la revista
Vol. 8. Núm. 2.
(abril - junio 2023)
Compartir
Compartir
Descargar PDF
Más opciones de artículo
Visitas
823
Vol. 8. Núm. 2.
(abril - junio 2023)
Acceso a texto completo
Online cheaters: Profiles and motivations of internet users who falsify their data online
Visitas
823
Laura Sáez-Ortuñoa,
Autor para correspondencia
laurasaez@ub.edu

Corresponding author.
, Santiago Forgas-Colla, Ruben Huertas-Garciaa, Javier Sánchez-Garcíab
a Universitat de Barcelona, Business Department, Avda. Diagonal, 690, 08034 Barcelona, Spain
b Universitat Jaume I, Department of Business Administration and Marketing, Avda. Vicent Sos Baynat, s/n, 12071 Castelló de la Plana, Spain
Este artículo ha recibido
Información del artículo
Resumen
Texto completo
Bibliografía
Descargar PDF
Estadísticas
Figuras (1)
Tablas (6)
Table 1. Distribution of unintentional and intentional errors for name and surname compared to sample total.
Table 2. Frequency distribution, expected frequency and relative frequency of inattentive and intentional errors for names and surnames by sex and generations.
Table 3. Frequency distribution, expected frequency and relative frequency of the number of hard bounces and soft bounces of email by reported sex and generation.
Table 4. Frequency distribution, expected frequency and relative frequency of the number of telephone number errors by reported sex and generation.
Table 5. Interview results and categorisation from the open coding.
Table 6. Relevance of the factors that motivate the introduction of false information obtained through statistical regression inference.
Mostrar másMostrar menos
Abstract

The digital environment, which includes the Internet and social networks, is propitious for digital marketing. However, the collection, filtering and analysis of the enormous, constant flow of information on social networks is a major challenge for both academics and practitioners. The aim of this research is to assist the process of filtering the personal information provided by users when registering online, and to determine which user profiles lie the most, and why. This entailed conducting three different studies. Study 1 estimates the percentage of Spanish users by stated sex and generation who lie the most when registering their personal data by analysing a database of 5,534,702 participants in online sweepstakes and quizzes using a combination of error detection algorithms, and a test of differences in proportions to measure the profiles of the most fraudulent users. Estimates show that some user profiles are more inclined to make mistakes and others to forge data intentionally, the latter being the majority. The groups that are most likely to supply incorrect data are older men and younger women. Study 2 explores the main motivations for intentionally providing false information, and finds that the most common reasons are related to amusement, such as playing pranks, and lack of faith in the company's data privacy and security measures. These results will enable academics and companies to improve mechanisms to filter out cheaters and avoid including them in their databases.

Keywords:
Social networks
Online cheaters
False data
Sweepstakes
Data collection
JEL Codes:
M0 General, M31 Marketing, M37 Advertising, Y1 Data: Tables and Charts, O3 Innovation
Research and Development
Technological Change
Intellectual Property Rights, M15 IT Management
Texto completo
Introduction

As social media spreads, more and more people are using it to seek, consume and exchange information (Shu et al., 2017), resulting in the generation of a massive amount of data (Kapoor et al., 2018). The reason behind this trend lies in the very nature of social media, as it allows for more timely, easier and less costly consumption and dissemination of information than traditional news media (Shu et al., 2017). This environment is propitious for digital marketing to understand new forms of online consumer behaviour and to promote and sell its products (Kumar et al., 2016). Consumers can be analysed and segmented by referring to information about their demographic characteristics, consumption habits, etc., which can be captured from social networks (e.g. Instagram, Facebook, Twitter, TikTok, Pinterest, Snapchat, etc.) in order to generate leads. This can be done in a variety of ways, ranging from the publication of advertisements, participation in social networks, joining conversations, and creating online contests and sweepstakes (Desai, 2019). However, the information that users provide is not always correct. Many people take advantage of the anonymity offered by social networks to falsify their information, and to act in a dishonest manner (Allcott & Gentzkow, 2017; Bonald et al., 2009; Vosoughi et al., 2018; Wu et al., 2022).

A cheater is someone who participates in a game but breaks the rules in order to gain an advantage. In other words, s/he wants to join in, but is not willing to play fair (Cosmides & Tooby, 2016). Since ancient times, cheating has been a perplexing problem for society and has been an especially huge obstacle for businesses (Cosmides & Tooby, 2016). One form of cheating is to provide false information about oneself by misrepresenting or impersonating another person (Lwin et al., 2016). Although the problems that this causes are recognised, little is known about the profiles of users who are most inclined to do so (Axelrod & Hamilton, 1981; Di Domenico et al., 2020). In the digital marketing environment, where strategies depend on the provision of truthful, accurate information about consumers, it is essential to detect users who enter false information in order to remove them from databases (Blackburn et al., 2014; Cosmides & Tooby, 2016; Pascual-Ezama at al., 2020). Knowledge about the user profiles that are more likely to misrepresent their data could help to refine detection methods and eliminate/reduce fraudulent practices (Ahmed, 2009). The incorporation of this information into artificial intelligence and machine learning algorithms that sift information could help improve their performance (Saura, 2021; Zhang et al., 2020).

The global market for collection, storage and distribution of digital marketing-related data was worth nearly $17.7 billion in 2021, with the US being the largest market, accounting for 47% of the global value, around $24.7 billion. In Spain, digital advertising amounted to €3.03 billion in 2020 and the country is among the top ten in Europe with the highest spending in this area, with a figure that surpasses that for traditional advertising.

Digital marketers need accurate sources of data on potential consumers to target and optimise their marketing investments (Lee et al., 2012). Therefore, it is essential for organizations that build databases from social media to eliminate as much pollution by fraudulent users as possible. The literature on social media marketing has focused more on data collection related to web traffic, on user engagement with each other and with the company or brand, than on the motivations or profiling of digital fraudsters (Chambers et al., 2010). Certain studies have analysed the social and psychological motivations that lead consumers to provide their information online (Balint et al., 2011; Fritsch et al.,2006), but fewer have focused on detecting the profiles of those who lie. One exception is Nazir et al. (2010), which analysed behaviour with false accounts used to play Facebook games. It found that users’ main motivation for providing inaccurate personal data was to gain what they believed to be an advantage in the game, but the study did not profile these users.

One study that analysed a larger number of cases of cheating was carried out by Blackburn et al. (2014). These authors examined the cheaters flagged in an online game, finding that their number is not correlated with population density or the size of the game community. However, they did not provide information on cheaters' profiles, such as their age or sex, or the fields in which they were most likely to lie, although they do suggest that the costs of cheating are extremely significant, especially those to the industry as it seeks to detect and reduce the practice.

This paper aims to contribute by identifying the motivations and profiles of users who provide false data on the Internet. To this end, we requested the collaboration of one of the top lead generation companies in Europe, which has been operating in Spain since 2009. This company allowed us to study certain pieces of information from its database, which we used to estimate the amount of fraudulent data and to characterise cheaters by stated sex and age. We also sought to learn about their motivations for supplying false information. The data generation industry is particularly sensitive and exposed to cheaters, so early detection is critical to prevent them from contaminating the databases that will subsequently be used by companies to offer their products and services. In addition, good quality databases help to target commercial activity better, and generate greater acceptance, engagement and brand loyalty (Menon et al., 2019).

This study analyses the characteristics of users who have provided false information on the lead generation company's web pages, in order to fill the following gaps in the literature:

  • detection of false information using AI algorithms and, in turn, the users who entered false information

  • determination of whether users who entered false information did so intentionally or negligently

  • examination of the fields where cheaters have entered the most false information

  • characterisation by stated sex and generation of users who falsify their data the most, in order to incorporate these profiles into prediction algorithms, and

  • understanding of the main motivations for intentionally providing false information.

To our knowledge, no previous work in the literature has analysed users’ profiles and their motivations to enter false information when participating in online sweepstakes and quizzes with the aim of facilitating automatic detection of cheaters on social media. This research adopts a mixed-method approach that combines descriptive and exploratory research. For the descriptive research, we benefited from collaboration with the lead generation company CoRegistros, S.L.U., which provided us with several fields of a database of more than 5 million users.

The rest of the paper is organised as follows. First, a conceptual framework is presented focusing on the profiles of users who enter inaccurate personal data online and their motivations for doing so. Second, the methods and results of the two studies on which this research is based are presented. Following a discussion of the results, the implications for academia and management are addressed. The study concludes by proposing the key themes that emerged from the results, discussing its limitations, and suggesting certain avenues for future research.

Theoretical frameworkDefinition and types of fake information created by cheaters

Although the concept of fake news originated in the 15th century (Shu et al., 2017), that of online misinformation was coined six centuries later, in the early 21th century, to refer to a series of untruthful news stories and announcements generated and disseminated by websites (Mintz, 2002; Wendling, 2018) that affect most social domains (Allcott & Gentzkow, 2017; Fonseca & Borges-Tiago, 2021). When information hits the web, cheaters, under protection of the anonymity afforded by the online environment, manipulate that information and re-distribute it, generating false content (Allcott & Gentzkow, 2017). Here, it is important to distinguish between misinformation and disinformation, as some studies have used them indiscriminately (Zubiaga et al. 2018). The terms have different meanings, for while misinformation refers to communications whose veracity is not yet confirmed and may or may not contain false information, disinformation involves deliberate manipulation to give the impression that the content is true (Tandoc et al., 2018). That is, while the former concerns the authenticity of the information, the latter implies intentionality (Shu et al., 2017).

This generation of misinformation, its ontology, detection methods and the motivations behind it have aroused much interest in the scientific community, which has carried out several studies to improve our understanding of the phenomenon. There have been studies such as the one by Habib et al. (2019), which endeavoured to classify misinformation into rumours, fake news, disinformation and hoaxes, and also described their characteristics to facilitate their detection and prevent cheating. Meanwhile, Tandoc et al. (2018) sought to categorise the purposes of false information that is disseminated online into satire, parody, political propaganda, advertising and manipulation.

Automated processes of online information dissemination are changing and increasingly attractive headlines and very limited and short-lived content are becoming more and more common, making manual monitoring impossible and thus favouring proliferation of fake news and the detection of cheaters (Conroy et al., 2015). Although online misinformation is a recent phenomenon, some authors propose the adaptation of methods described in earlier literature to the detection of cheaters in different fields of application. Examples include Conroy et al. (2015), Parikh and Atrey (2018) and Shu et al. (2017), who focus on the automatic detection of false information once it has been generated, while others such as Zubiaga et al. (2018) address the problem more holistically, rather than merely detecting it once it has been produced. Along similar lines, Bondielli and Marcelloni (2019) approach false information from its origin, i.e. in terms of data sources and the way in which information is captured. Regardless of how cheaters are detected, all these methodologies recognise that they entail certain limitations and that they need to recurrently train their algorithms by means of behavioural and socio-demographic data.

Profiling types of errors: accidental or intentional

Since misinformation can be generated by accident, it is important to detect whether or not there is any malicious intent behind its creation (Pennycook et al., 2021). To reduce the likelihood of error, it is proposed that robust protocols should be used to control the way that users complete registration forms (Karlova & Fisher, 2013). For example, they might be asked to enter the same data more than once, without being able to see what they typed previously, and the submission is only accepted if both entries match (Fallis, 2014). But these procedures are unable to prevent users from intentionally entering false data (Karlova & Fisher, 2013). For example, if a user gives his/her name as "Fool", and the system asks him/her to repeat it, s/he will do the same thing again. But if the algorithm detects that the word "Fool" is incorrect and lets him/her know, s/he is likely to use a fake, but apparently real, name on the second try, which is much harder to detect. It is therefore important to distinguish whether false data is provided due to error, misinformation, or where this is done intentionally, disinformation, and also to know which kinds of users are more likely to do so. In the former case, to improve the robustness of online forms, and, in the latter, in order to control and isolate such practices (Karlova & Fisher, 2013). Thus, the following research question is proposed:

  • RQ1. Are errors mainly produced accidentally, generating misinformation, or intentionally, generating disinformation, when filling in personal data online?

Fake information created by cheaters: detection of cheating through leads and user attitudes on registration

Digital marketing often uses databases that gather information from potential consumers (leads) in order to target commercial offers better, and one way to create these databases is through lead generation (Desai, 2019). In the past, leads were acquired by making phone calls, usually without the respondent's authorisation or consent, but nowadays such processes are largely carried out through digital channels (Rothman, 2014) where the user gives their consent under a regulated framework (Spanish Data Protection Agency, 2022).

There are several ways to generate online leads, such as offering interesting content on blogs or websites (Bondarenko et al., 2019), electronic requests made by social activists (Huang et al., 2015), offering financial incentives such as prize draws or direct product discounts, or revealing the answers to a quiz in exchange for the user's data. Another technique is snowballing, which consists of users winning rewards in exchange for recruiting friends and acquaintances, whose information thus becomes available to the company too (Baltar & Brunet, 2012). These all follow the principle of the social contract (Cosmides & Tooby, 2016), whereby participants are given the chance to win a prize (e.g. an iPhone, a gift voucher, a trip, etc.) or some kind of emotional reward, for getting the answers right to a quiz, test or challenge, in exchange for providing personal information.

However, the information provided by participants often contains errors, and checks need to be performed to safeguard the quality of the database. This essentially involves input control and screening of the provided information. Robust data entry procedures are often used for the former, while the latter uses algorithms to detect patterns in transcription errors (incorrect names, missing phone numbers, etc.) regardless of whether they are accidental or intentionally malicious (Thakur et al., 2017).

However, despite the importance of detecting the profiles and patterns of false registrations and, consequently, cheaters, no previous studies have been found that have considered the declared sex and/or age of participants that create false profiles (Perez-Escoda et al., 2021). Some, such as Sharif and Zhang (2014), did identify the main ways in which consumers could mislead and deceive on social media and how such deception can be detected. Others such as Viviani and Pasi (2017) identified and quantified a user's credibility when entering information on social media, while Conroy et al. (2015) demonstrated that some techniques are more effective than others in detecting online deception and identifying fraudsters. Although previous studies have addressed different aspects of the problem (Conroy et al., 2015; Habib et al., 2019; Parikh & Atrey, 2018; Shu et al., 2017; Viviani & Pasi, 2017; Zubiaga et al., 2018), they all highlight the need to create control mechanisms to ensure the quality of databases, and to use the knowledge extracted from them to compare approaches and profile cheaters better.

Profiling cheaters based on their generation and declared sex

Generational cohort marketing, first defined in the US at the turn of the last century, is still being used in marketing around the world (Meredith & Schewe, 1994). Cohorts are groups of individuals who are born around the same time and experience external events in a similar manner in their late teenage/early adult years. These "defining moments" influence their values, references, attitudes and purchasing behaviour in ways that persist throughout their lives (Meredith & Schewe, 1994). The experiences shared during the highly impressionable "coming of age" period [approximately 17-23 years of age] embody these values or "cohort effects" and remain relatively unchanged throughout life. Each generation is defined by its birth years and typically lasts 20 to 25 years, or about as long as it takes to grow up and have children. But a cohort can be as long or short as the external events that define it. Thus, the cohort defined by World War II might only be 6 years long (Meredith & Schewe, 1994). Schuman and Scott (1989) demonstrated that individuals of similar age have similar memories, related mainly to adolescence and young adulthood, and common experiences of major events, which they refer to throughout their lives. These characteristics mean that each cohort is a separate market segment and it can be particularly useful for marketing campaigns to target them in specific ways. In the US, seven distinct cohorts have been delineated as internally homogeneous in values yet heterogeneous across cohort groups. The most widespread classification of generational cohorts is usually: Silent Generation (also known as Mature, born between 1925 and 1942), Baby Boomers (born between 1943 and 1960), Generation X (born between 1961 and 1981), Millennial Generation (often referred to as Generation Y or Millennials, born between 1982 and 2000) (Brosdahl & Carpenter, 2011) and Generation Z (born between 2001 and 2009) (Yadav & Rai, 2017).

While there is growing interest in understanding the use of social media by different generations (Bolton et al., 2013), little is known about which generations cheat the most. Thus, the following research question is proposed:

  • RQ2. Are there generational differences when entering incorrect personal data?

Another of the most common variables in marketing segmentation is the declared sex of users (Nickel et al., 2020). Consumers have often been classified according to stated sex in order to optimise product design, as well as to create targeted communication and advertising campaigns (Meyers-Levy et al., 2015). The selectivity hypothesis is based on using declared sex as a basic criterion to segment the market between male and female products (Moss, 2009). This theory suggests that most people who claim to be of certain sexes report different preferences and tastes and react differently to commercial stimuli (Nickel et al., 2020). Although there is abundant literature on the different attitudes of men and women towards new technologies and internet use (Alalwan et al., 2017) and even on their attitudes towards sweepstakes in the face of different types of stimuli and incentives (Schulten & Rauch, 2015), there is no evidence of studies that analyse sex differences among cheaters when entering personal data online. Therefore, the following research question is proposed:

  • RQ3. Are there any differences with regard to declared sex when entering incorrect personal data?

Motivations of users with fake identities

Although the instruments for collecting leads follow the logic of the social contract, in which financial and emotional incentives are offered in exchange for information, not everyone is willing to comply (Cosmides & Tooby, 2016). Previous studies have attempted to find the motivations for the generation of disinformation, and lack of understanding about the need to provide personal data, as well as privacy and security concerns, are cited as the main reasons (Sannon et al., 2018). One of the ways to conceal information is the use of pseudonyms, whereby the participant can not only hide his/her identity, but might also impersonate someone else either for fun or as a joke, or for criminal reasons (harassment). Find that amusement is the main motivation for supplying false information. Also, when companies request a large amount of information this can generate distrust, and Keusch et al. (2019) showed that users feel more confident if data collection is limited to the minimum necessary and, moreover, if data protection rules are clearly explained. Although financial incentives also help, they are not completely decisive (Keusch et al., 2019). Other authors, such as Sullivan et al. (2019), find that clearly describing the purpose of requesting information helps to prevent users from worrying that their privacy might be in jeopardy. Based on the evidence gathered, this study poses the following research question:

  • RQ4. What are the main motivations for intentionally entering incorrect online data, i.e. to generate disinformation?

Overview of the studies

To address the research questions, this study has used triangulation, which is the combination of different methods to study the same phenomenon (Denzin, 1978). We used three methodologies in our two studies: Study 1 is quantitative and descriptive, and is used to estimate the volume of cheaters by stated sex and cohort of the database made up of the information provided by volunteer participants in online sweepstakes and tests. Meanwhile, Study 2 is mixed, combining qualitative exploratory research (2a) and quantitative descriptive research (2b), to determine the weight of the main motivators declared in Study 2a. Specifically, Study 1 used AI algorithms to estimate the amount of erroneous and falsified data in a sample provided by the lead generation company of 5,534,702 participants in online sweepstakes and quizzes between 2010 and 2021. Study 2 aimed to explore and estimate the main motivators for intentionally falsifying data provided to sweepstake sponsors. To this end, in the first stage, the exploratory Study 2a consisted of 33 in-depth interviews with participants to enquire about the main motivators for falsifying data and, in the second stage, the descriptive Study 2b used a choice-based conjoint analysis methodology with a sample of 269 participants to estimate the weight of the factors revealed in the first stage.

Study 1

In order to build a profile of cheaters, a descriptive analysis was proposed of certain fields of the database provided by the lead generation company after signing a confidentiality agreement. This database contains information provided by participants in sweepstakes (96%) and self-assessment quizzes (intelligence, geography, cooking, etc.) (4%) over a period of eleven years, from 2010 to 2021, and was collected through the use of landing pages (Fig. 1 shows an example) that offer the possibility of winning an iPhone in exchange for the participant providing personal information. Other examples of landing pages can be found on the company's own websites (https://www.sorteopremios.com, https://www.mitest.de).

Fig. 1.

Example of a sample data collection form from www.sorteopremios.com.

(0,38MB).

On average, each user takes 3 minutes and 47 seconds to enter his/her data. To comply with the European Data Protection Regulation and the respective Spanish legislation, LOPD-RGPD, "Ley Orgánica 3/2018, de 5 de diciembre, de Protección de Datos Personales y Garantía de Los Derechos Digitales" (2018), all users accessing the landing page were previously informed. They also had to opt-in by checking the consent box to agree to the different purposes for which their data was collected, declare that they were over 18 years old, confirm that they had read and accepted the entry conditions and data protection policy, and agree to receive commercial information from the sponsors.

Participation in the company's sweepstakes or quizzes does not entail any entry barrier, as the only requirement is to be over 18 years old and have an Internet connection, this latter requirement being met by more than 90% of the Spanish population (INE, 2020). The company has 5,534,702 registered users and it provided us with information on their names and surnames, emails, telephone numbers, and declared sex and ages. Data from users who did not declare their sex was not included in this study as there were only 55 such cases, amounting to just 0.001%. All users said they were of legal age and Spanish.

Based on their stated sex and age, the sample was divided into women and men, and into five generational cohorts: Silent Generation, Baby Boomers, Generation X, Millennials and Generation Z. The sample is somewhat asymmetrical, as there are more women, 65%, than men, 35%. The distribution of the cohorts is also heterogeneous, although all subsamples are representative (Rao et al., 2021). Generation X (52%) is the most highly represented, followed by Millennials (40%), Baby Boomers (7%), Silent Generation (0.8%) and, finally, Generation Z (0.2%). Furthermore, in all cohorts, except for the Silent Generation, the relative frequency of women was almost double that of men.

Measures

Different procedures were followed to estimate the number of errors made unintentionally, misinformation, and intentionally, disinformation, in each of the available fields of the database. These ranged from developing a debugging algorithm for name and surname, to comparisons with official databases, verifications by means of chatbots, and automatic forwarding to registered email addresses. With all these tracking and control mechanisms, estimates of the amount of fraudulently entered data could be calculated.

Starting with the name and surname records, a debugging algorithm was developed from Node.js®, with which JavaScript is used to write command line tools (Escobar-Jeria et al., 2007). This algorithm, which we cannot publish for copyright reasons, compares all the names and surnames registered by users who played sweepstakes and self-assessment quizzes with those in the databases of the repository of the National Statistics Institute and the IDA-Padrón, and detects all those words that do not match Spanish names and/or surnames or which appear fewer than 20 times in the country (or 5 per province). Once all records of suspicious names and surnames had been detected, an algorithm was applied to them to detect typographical errors, which are considered unintentional, while all other unusual names and surnames were considered fraudulent. In addition, to estimate the goodness of fit of the distribution of fraudulent names (Chi-square test), this was compared to the distribution of names in the company's standardised tables and blacklists. Subsequently, this frequency distribution of unusual names was incorporated into the algorithm as a contribution to machine learning. This means that the name registration software will not allow users to register names that have previously been identified as fraudulent.

Regarding the telephone number and e-mail address fields, the registration system is double-entry, which prevents typographical errors. Hence all errors made when entering this information were considered fraudulent. In order to detect bogus telephone numbers, even when they present a valid format according to the Spanish National Commission for Markets and Competition (2021), interactive voice response (IVR) systems were used (Dillman et al., 2009). Call control samples were carried out automatically, by means of chatbots, and manually (control calls) to confirm that the supplied data exists and is valid.

Finally, for email address registration, verification simply consisted of sending automated messages and checking whether they were opened, click rates and other metrics. The bounce rate was measured to estimate fraudulent email addresses, aggregating soft bounces and hard bounces (Poulos et al., 2020). While hard bounces occur when the e-mail indicator is incorrect and/or the user's name before the @ is false, soft bounces occur when, for example, a user cannot receive emails because their inbox is full, the sender's address has been blocked as spam, or the mail server is temporarily down (Maaß et al., 2021).

Analysis and results

Having estimated the fraudulent data entered by users in the different fields of the database following the different procedures outlined above, an analysis was performed in different stages to determine the most fraudulent profiles. Following Saunders et al. (2009), we compared the differences between generations and declared sex for name and surname, telephone number and e-mail address information. We used the χ2 test (Chi-square) and then paired t-tests on the distributions and observed significant differences in terms of the results.

We estimated that 325,096 users included one or more errors in their registrations, representing 5.87% of the over 5 million unique users. Regarding the registration of names and surnames, the results show that 268,980 users (4.86% of the total) intentionally supplied fraudulent information, disinformation, a much higher figure than that for those who made errors due to inattention, misinformation, 55,903 (1.01% of the total). Hence, in response to RQ1, we conclude that intentionality and, therefore, the generation of disinformation is the main reason for errors.

In response to RQ2 and RQ3, the results also suggest different inclinations to provide incorrect information among the five generational and sex cohorts. Generation Z and Silent users are found to proportionally make the most unintentional errors and Millennials make the fewest. These are also the cohorts that make the most (Generation Z and Silent) and fewest (Millennials) intentional errors (see Table 1).

Table 1.

Distribution of unintentional and intentional errors for name and surname compared to sample total.

Type of error  Silent Generation  Baby Boomers  Generation X  Millennials  Generation Z  Total 
Unintentional errors  1,038 (421)59.4%  6,996 (3,871)44.6%  30,456 (28,887)5.1%  16,955 (22,630)-33.4%  458 (91)79.9%  55,903 
Intentional errors  5,743 (2,027)64.7%  34,880 (18,627)46.5%  131,426 (138,994)-5.7%  88,404 (108,889)-23.1%  8,527 (441)94.8%  268,980 
Total sample  41,710  383,282  2,860,029  2,240,587  9,094  5,534,702 

Notes: In each cell: top figure, absolute frequencies; in brackets, expected values; and percentage deviation from the expected value. Unintentional errors χ2 (4) = 6393.05, p = .000; Intentional errors χ2 (4) = 173165, p = .000.

Regarding the comparative analysis by stated sex and generation, the male Silent Generation (72%), followed at a considerable distance by the male Baby Boomers (57%), are far more likely to make mistakes due to inattention. Similar figures were observed for disinformation among males: Silent Generation (69%) and Baby Boomers (58%). However, among females, it is the younger cohorts, Millennials (51%) and Generation X (50%), who have a slightly higher tendency to make errors due to inattention and these generations also make slightly more intentional errors, although Generation Z (58%) does so the most. To summarise, as shown in Table 2, men in the older cohorts and women in the younger cohorts are most likely to provide incorrect data both by accident and intentionally.

Table 2.

Frequency distribution, expected frequency and relative frequency of inattentive and intentional errors for names and surnames by sex and generations.

Errors due to lack of attention
Sex  Silent Generation  Baby BoomersGeneration XMillennials  Generation Z  Total 
Male  745  4,01415,2318,346  226  28,562 
  (530.3)  (3,574.4)(15,560.6)(8,662.7)  (234.0)  (28,562) 
  72%  57%50%49%  49%  51% 
Female  293  2,98215,2258,609  232  27,341 
  (507.7)  (3,421.6)(14,895.4)(8,292.3)  (224.0)  (27,341) 
  28%  43%50%51%  51%  49% 
Total  1,038  6,99630,45616,955  458  55,903 
  100%  100%100%100%  100%  100% 
Notes: In each cell: top figure, absolute frequencies; in brackets, expected values; and, below, relative frequency percentage. Unintentional errors χ2 (4) = 326.70, p = .000. In bold, significant differences compared to the total.
Intentional errors
Sex  Silent Generation  Baby Boomers  Generation XMillennialsGeneration Z  Total 
Male  3,969.0  20,332.0  64,556.044,156.03,593.0  136,606.0 
  (2,916.7)  (17,714.4)  (66,746.9)(44,897.5)(4,330.6)  (136,606) 
  69%  58%  49%50%42%  51% 
Female  1,774.0  14,548.0  66,870.044,248.04,934.0  132,374.0 
  (2,826.3)  (17,165.6)  (64,679.1)(43,506.5)(4,196.4)  (132,374) 
  31%  42%  51%50%58%  49% 
Total  5,743.0  34,880.0  131,426.088,404.08,527.0  268,980.0 
  100%  100%  100%100%100%  100% 

Notes: In each cell: top figure, absolute frequencies; in brackets, expected values; and, below, relative frequency percentage. Intentional errors χ2 (4) = 2000.0, p = .000. In bold, significant differences compared to the total.

With regard to email address bounce rates, the results of the frequency analysis of hard and soft bounces (see Table 3) reveal that both the sex and generational cohorts affect the inclination to defraud. In the case of hard bounces, generated by entering invalid addresses, a higher frequency was observed among men of the Silent generation (50%) and Millennials (34%). As for soft bounces, which can be caused by the recipient's mailbox being full, the highest frequencies are also found among the Silent generation (70%) and also among Baby Boomers (24%). This reveals that older generations of men have a higher propensity to cheat when entering their email address. Among women, Generation X (71% for hard bounces and 82% for soft bounces) presents similar figures for both hard and soft bounces. The other generation of females that is most prone to soft bounces is Millennials (80%), while for hard bounces, it is Baby Boomers (70%). In the case of Generation Z, in both sexes the values were too low to be considered.

Table 3.

Frequency distribution, expected frequency and relative frequency of the number of hard bounces and soft bounces of email by reported sex and generation.

Hard bounces
Sex  Silent GenerationBaby Boomers  Generation X  Millennials  Generation Z  Total 
Male  52497  1.195  737  2.481 
  (32)(507)  (1.270)  (672)  (0.30)  (2.481) 
  50%30%  29%  34%  0%  31% 
Female  531.149  2.929  1.444  1,00  5.576 
  (73)(1.139)  (2.854)  (1.509)  (0.70)  (5.576) 
  50%70%  71%  66%  100%  69% 
Total  1051.646  4.124  2.181  8.057 
  100%100%  100%  100%  100%  100% 
Notes: In each cell: top figure, absolute frequencies; in brackets, expected values; and, below, relative frequency percentage. Hard bounces of email χ2 (4) = 33.5953, p = .000. In bold, significant differences compared to the total.
Soft bounces
SexSilent Generation  Baby BoomersGeneration X  Millennials  Generation Z  Total 
Male14  3080  45  172 
(4)  (26)(94)  (47)  (1)  (172) 
70%  24%18%  20%  60%  21% 
Female96373  182  659 
(16)  (100)(359)  (180)  (4)  (659) 
30%  76%82%  80%  40%  79% 
Total20  126453  227  831 
100%  100%100%  100%  100%  100% 

Notes: In each cell: top figure, absolute frequencies; in brackets, expected values; and, below, relative frequency percentage. Soft bounces of email χ2 (4) = 37.7187, p = .000. In bold, significant differences compared to the total.

With regard to the number of fraudulent telephone numbers entered, there are also differences between sex and generation of users (see Table 4). Again, males from older cohorts (Silent Generation, 66%, and Baby Boomers, 51%) and, in this case, younger cohorts (Generation Z, 48%) are more likely to make mistakes when filling in their phone number. Among women, it is the middle-aged cohorts (Generation X, 55%, and Millennials, 54%) who have a higher propensity to disinform in this field. The greatest divergence between men and women is between the Silent Generation (with significant male participation) and Generation X (with high female participation).

Table 4.

Frequency distribution, expected frequency and relative frequency of the number of telephone number errors by reported sex and generation.

Sex  Silent Generation  Baby Boomers  Generation X  Millennials  Generation Z  Total 
Male  1,406  5,974  35,898  27,502  200  70,980 
  (990)  (5,501)  (36,770)  (27,523)  (196)  (70,980) 
  66%  51%  45%  46%  48%  47% 
Female  723  5,852  43,147  31,666    81,609 
  (1,139)  (6,325)  (42,276)  (31,645)  (225)  (81,609) 
  34%  49%  55%  54%  52%  53% 
Total  2,129  11,826  79,045  59,168  421  152,589 
  100%  100%  100%  100%  100%  100% 

Notes: In each cell: top figure, absolute frequencies; in brackets, expected values; and, below, relative frequency percentage. Errors in phone χ2 (4) = 440.9970, p = .000. In bold, significant differences by sex compared to the total.

In general, in the fields studied, there continues to be a tendency for older male generations (Silent Generation and Baby Boomers) and middle-aged female generations (Generation X and Millennials) to cheat when registering their personal data.

Study 2a

In order to discover users' main motivations for entering false information, disinformation, a sample of regular users of these online sweepstakes residing in Barcelona was invited to attend an in-depth interview at a central location, with a €50 cheque being offered as an incentive. This qualitative in-depth interview technique is used when a researcher wants to get a clearer idea about a phenomenon or when prior information is insufficient. To recruit participants, 650 telephone calls were made, of which 293 were answered, 163 expressed an intention to participate, and 33 were finally selected (16 stated male (M) and 17 female (F)). The saturation criterion was employed to select the sample size, i.e. the sample recruitment process ended when no new information was received from new sampled units (Lincoln & Guba, 1985).

Data collection process

Once participants arrived at the venue, they were welcomed, given a brief introduction to the study and told about the economic and social consequences of introducing false information in online communication. In order not to condition responses, we followed Sannon et al's (2018) procedure of downplaying the importance of socially reprehensible behaviour. The respondents were told that we did not view the use of lies or falsification of data as good or bad, but that we were simply interested in analysing an important part of human communication. In order to contextualise the participants in the topic of the study, the interviewer presented a series of examples of false information supplied by participants in online sweepstakes and which had been collected by the lead generation company. The interviewees were then asked to try to explain the motives that might have led the entrants to provide incorrect data. After completing the consent form, the interviews were recorded. The interviews lasted an average of about 32 minutes.

Data analysis and results

The recorded data was transcribed and analysed sequentially following the principles of thematic analysis. Coding was performed in three stages: First, the transcripts were read, the interesting sections were highlighted, and annotations were added in the margins. Second, the interesting parts were openly coded, and 24 codes were identified. Third, the codes were grouped hierarchically into a three-order structure. This axial condensation process (Tuomi et al., 2021) ultimately resulted in three main themes: privacy concerns as a consequence of asking for too much information; trust in the company or website providing such quizzes and sweepstakes, and amusement.

To check the analytical consistency of the coding process, the codebook, the descriptions of each main theme, and selected paragraphs from the interviews were emailed to an independent reviewer for recoding. Following the instructions of Tuomi et al (2021), this reviewer was not connected in any way to the research and also came from a different university background (Coder: 37 years old, Computer Engineer). The reliability between the two proposed codifications was determined by Cohen's Kappa, indicating very good inter-coder agreement (>0.80) (Landis & Koch, 1977).

Regarding the results, the participants stated that they use online sweepstakes and quizzes as a source of entertainment, and that transcription errors (‘typos’) are indeed common, as this is the least entertaining or interesting part of the activity. They also comment that intentional mistakes (giving a different name to their own) are made in order to preserve their anonymity. In addition, issues such as the topics of greatest interest (history, geography, celebrities, music, etc.) were raised, including whether the prize was more or less attractive. Regarding the motives for providing false information, the different topics were grouped into three categories that were labelled Privacy, Trust and Amusement.

  • (1)

    Privacy. Respondents express concern about the loss of anonymity and that websites ask for too much information, which conveys a sense of risk.

  • (2)

    Trust. Participants expressed some doubts as to who is sponsoring the online sweepstakes and tests. It was commented that advertising should offer guarantees that it is safe and should also engender trust. There was consensus that the site from which data is requested is important. If it belongs to a public body, so much the better.

  • (3)

    Amusement. Some users impersonate the names of acquaintances for fun. There is also talk of minors, whose participation is not allowed by the system, so they do so by entering false information.

The results of the categorisation from the open coding are shown in Table 5.

Table 5.

Interview results and categorisation from the open coding.

Themes  Description  Sample QuotationsListing Key Phrases 
Privacy  On the Digital Marketing side, the privacy component that is considered by users is feeling that their privacy might be jeopardised (Sannon et al., 2018"At first you get really excited when you see the prize, but then you think, why do they need all this data to give me the prize? I don't mind giving my email address but my phone number!" A, 32 "Why do you ask me for so much information, and what use will you make of it?" R, 43 "Could it be a scam?" O, 51Too much information Excessive amount of information requested Risk 
Trust  Trust refers to the data that users must provide to sponsors and raises questions about who is sponsoring online sweepstakes (Lwin et al., 2016"I can understand why Social Security asks for your data, such as your national ID number, but why does a private entity need it? Either you are very clear that it is a necessary requirement to obtain the prize, and they guarantee me security, or I will only offer my valid email address, the rest of the data will be invented" J, 22 "Who is behind the sweepstake, can I trust them, will it be a scam?" D, 27 "Is it worth giving all this information for the prize I might possibly get?" R, 37PhishingHackersSecurityPossibleBenefit vs. risk 
Joke  Impersonating other people is form of amusement (Małgorzata et alt.,2018)  "You put the name of an acquaintance for a laugh, you hope they call and that it will be a surprise" R,19"Minors, who cannot enter because the system does not allow it, can impersonate adults" X, 18"Surprise a friend" A, 24"I often get bored and don't know what to do with my time, so I enjoy playing jokes" N, 18Playing pranksKill boredomWaiting times 
Study 2b

Based on the results obtained from the qualitative study, a quantitative study was used to measure the importance of the three factors revealed by the thematic content analysis. Since the aim was to measure the weight of factors of socially reprehensible behaviour (Sannon et al., 2018), instead of asking direct questions, a decomposition methodology (conjoint analysis) was used to estimate the users' preference structure. This consists of forming scenarios by combining the motivations that arose from the exploratory research (privacy, trust and amusement) and asking participants to choose the scenario that best identifies them.

Process of data collection, measurement and analysis

Discrete choice-based decomposition methods require five steps:

  • (1)

    Determine the number of factors and levels. In this study, three factors at two levels were considered: F1 (Information: +1 excessive information, -1 not too much information), F2 (Distrust: +1 high distrust, -1 trust), F3 (Amusement: +1 for fun, -1 not for fun).

  • (2)

    Create the experimental design. Considering three factors at two levels, the number of possible scenarios is 2 3 = 8. However, instead of asking participants to compare 8 scenarios and choose the one that best identified with them, as is usual in classical conjoint analysis, we used an adaptive conjoint analysis (ACA) design consisting of twelve blocks of two profiles (Huertas-García et al., 2016). Each respondent was randomly assigned a scenario consisting of three blocks of two profiles each (e.g. block 1 consisting of profiles 1 and 7, block 2 consisting of 2 and 5, and block 3 consisting of 5 and 8). In total, four 3-block scenarios with two profiles each were assigned (24 profiles). From each block, the respondent had to choose one of the two, so three pieces of information were collected from each respondent to allow estimation not only of the weight of the main factors but also between two-factor interactions. This ACA experimental design was proposed by Huertas-García et al. (2016) and a practical application can be found in Perdiger et al. (2019).

  • (3)

    Develop the appropriate question to elicit the choice in each choice set. The proposal was: "Imagine that you are participating in an online quiz and you have to fill in the data shown below (Figure 1) in order to win the prize. Which of the following sentences best describes your opinion regarding the supply of false information?" An example choice set is: "Please choose only one of the following options":

Option 1 (+ 1 excessive information; -1 confidence; -1 not for fun). "Because an excessive amount of information is requested, although I trust the site, and I do not create false names for fun".

Option 2 (- 1 not too much information, + 1 great mistrust, - 1 not for fun). "Because, although the amount of information requested is not excessive, I am very suspicious of the site, and I am not in favour of creating false names for fun".

Option 3. None of the options identifies me.

  • (1)

    Implement the choice sets following the experimental design with a sample of consumers. A purposive sampling strategy was used by sending emails and using Google Forms to create and share online forms and analyse responses in real time. 13,500 emails were sent to regular users of the online sweepstakes inviting them to complete the questionnaire and encouraging them to enter the IPhone 13 sweepstake, of which 2,929 were opened, 336 questionnaires were completed, and 269 were valid.

  • (2)

    Analyse the data with an appropriate analytical model. The results were estimated using the Multinomial Logit Model (Rao, 2014). Data was collected in May 2022.

Results

Table 6 summarises the results and shows the weight of main factors and two-factor interactions that motivate the supply of false information.

Table 6.

Relevance of the factors that motivate the introduction of false information obtained through statistical regression inference.

Interception  Coefficients  Standard error 
F1  2,60359717857956⁎⁎⁎  0,827215655 
F2  1,94519835354427⁎⁎  0,827215655 
F3  2,90268521016663⁎⁎⁎  0,827215655 
F12  -1,960548827  1,169859598 
F13  -1,439019901  1,169859598 
F23  -1,44745917  1,169859598 
Coefficient of determination R^2  0,808261437   
Standard error  1,547578782   

F1 means "Not trusting enough", F2 = "Safeguarding one's privacy" and F3= "Amusement”; F12 means the interaction between F1 and F2, and so on subsequently.

⁎⁎

p<.05.

⁎⁎⁎

p< .01.

The results show that the main motivation for users to enter false information was amusement, F3 (Amusement: +1 for fun), followed by not having enough trust in the company's website, F2 (Distrust: +1 high distrust), and, finally, the desire to maintain their privacy and considering that too much information was being requested, F1 (Information: +1 excessive information). Furthermore, the results indicate that the three factors act independently, as no interaction between two factors reached significant values. Therefore, the response to RQ4 on the main motivations for entering false information are: amusement, lack of trust in the company's website, and the desire to maintain one's privacy and not reveal an excessive amount of information.

Recently, the problems generated by the proliferation of misinformation and disinformation on social networks, and the need to detect it, have attracted a great deal of attention (Di Domenico et al., 2020; Pascual-Ezama at al., 2020). Existing approaches to cheater detection are mainly based on the use of certain user characteristics, such as unusual names, offensive words, and non-existent phone numbers or email addresses, and the configuration of blacklists of users, which artificial intelligence algorithms detect quickly and accurately (Saura, 2021; Zhang et al., 2020). However, knowing which user profiles are more inclined to misinform can boost the performance of these bots. In this study, it has been detected that men of older generations and women of younger generations are more likely to falsify their data. In addition, the main reasons for this socially reprehensible behaviour are fun, lack of trust in the website requesting the data and safeguarding privacy. Identification of cheaters and their motivations can help academics and practitioners to try to improve methods for capturing information, and also ways of detecting cheaters on social networks.

General discussion and conclusions

The emergence of social networks and the information flows generated between them have created an enabling environment for digital marketing. However, it is not easy to synthesise the enormous volume of information that circulates on networks in a way that can help academics and digital marketers to make decisions. One of the ways to analyse relevant information is to use databases of potential consumers collected by lead-generating companies (Desai, 2019). However, in order for these databases to fulfil their function, they must be as reliable as possible, i.e., they must contain real data that is as clean as possible of misinformation.

This study describes the profiles of users who enter false information when registering for online sweepstakes and quizzes, based on estimates of negligent, misinformation, and intentional inaccuracy, disinformation. The results suggest that most errors are made intentionally, at a ratio of almost 5:1 with regard to unintentional actions. Furthermore, men of older generations and women of younger generations are more likely to falsify their data. However, and in line with the findings of Dabija et al. (2018), small differences are also observed regarding the disclosure of names and surnames, emails and telephone numbers. We found that the most repeated motivations for producing disinformation were, in the following order, amusement, lack of trust in the site requesting the data, and safeguarding privacy. These results are in line with previous research, which has shown that trust is a key aspect and can be considered a predictor of whether or not the users of social networks will provide false information (Gefen et al., 2003).

This study furthers our knowledge about the process of capturing data from internet users, in this case by means of online sweepstakes and quizzes, and the problems arising from the volume of fraud committed by users. Indeed, we have not found any previous study that estimates and analyses such practices when users register their information on websites with such a large sample (more than five million) and over such a long period of time (eleven years). Although each of the fields requires a different method for estimating errors, there are common trends among some generational and sex profiles.

In the analysis of names and surnames, the cohorts with a higher propensity to enter incorrect data are Generation Z, Silent and Baby Boomers. However, when crossing the data with declared sex, we find that it is men from the older generations and younger women who are most inclined to misrepresent. However, actions when filling in the email address and phone number fields do not follow the exact same pattern as the previous ones, results that are in line with those obtained by Dabija et al. (2018). The estimation using hard bounces highlights male Silents and Millennials and female Baby Boomers and Generation X as the most fraudulent. Finally, using call-backs, male Silents and Baby Boomers and female Generation X and Millennials were found to be the most fraudulent generations.

Theoretical implications

While this research provides evidence of and support for the tendency of users to enter fraudulent information on social networks (Islam et al., 2020; Pennycook & Rand, 2019), we also find that this occurs in less than 6% of cases (and less than 5% for intentional errors). However, while there is room for improvement in mechanisms to reduce unintentional errors, mechanisms to control for intentional errors should be directed towards cheater profiling (Cosmides & Tooby, 2016).

This study presents evidence that some user profiles are more inclined than others to enter false information when registering on the Internet, so their identification can help predict such behaviour and target measures to control these practices better (Song et al., 2021). For example, it would be logical to assume that older people (Silent and Baby Boomers) are more inclined to make unintentional errors than younger people, as they are more affected by deterioration in physical condition and cognitive abilities (problems with sight, memory loss, difficulty typing letters correctly on a keyboard, etc.). However, this is only true when they are compared with middle-aged generations, but not with younger people (Generation Z) who are the most inclined towards such practices. Maybe, although young people are more accustomed to the Internet, they write in a hurried manner without checking that the information is correct (Valentine & Powers, 2013). Regarding intentional errors, one might assume, on the one hand, that more mature people, with more life experience and who have adopted these technologies much later, use them for a clear purpose and to obtain a specific outcome (Dabija & Grant, 2016). However, the results only partially confirm this assumption, as older, self-trained male cohorts tend to be more likely to enter false information. On the other hand, nor does the assumption hold that younger generations, who were born in the age of the Internet and social networks, behave differently to other generations, for the results of this study do not point in that direction (Lenhart et al., 2010). In fact, in the analysis of the name and surname fields, younger users behave similarly to older generations.

Implications for management

Although the introduction of false personal data does not occur in alarming proportions, it does affect both individual users and businesses (Shu et al., 2020). Given that the proliferation of cheaters is inversely correlated with good practices in tacit or explicit negotiations (Axelrod & Hamilton, 1981), it is important to detect them in order to prevent and eliminate fraud. Moreover, online environments with a large number of users facilitate such practices (Allcott & Gentzkow, 2017). Although psychological mechanisms have been developed in offline environments to dissuade cheaters (Mealey et al., 1996), these mechanisms are not directly transferable to online environments, so AI and technology play a key role in developing devices to mitigate the consequences of fraudulent information (Zhang & Ghorbani, 2020). Tackling these problems creates opportunities in the innovation and development of tools for detecting, preventing and monitoring potential fraud, with significant economic benefits through value creation and capture. Therefore, having a clear profile of cheaters as well as knowledge of their motivations for cheating can be very valuable (Nambisan et al., 2019), as it helps to filter the information that is fed into the databases used by companies and decision-makers, and directly affects the outcome of their decisions (Zhang et al., 2016; Ogilvie et al., 2017; Bondarenko et al., 2019; Lin et al., 2021). The determination of the most common cheater profiles (in terms of generation and stated sex) can help to filter databases so that companies can offer better personalised services to their customers (Zhang et al., 2016), preventing them from receiving information that is of no interest to them (Agrawal et al., 2011), and instead increasing the likelihood of making attractive offers and maximising returns (Zhang et al., 2016). For companies, more reliable databases will improve productivity (Lin et al., 2021), ensure they do not miss out on business opportunities (Bondarenko et al., 2019) and, ultimately, raise their profits (Tripathy et al., 2013). In turn, this will increase employee satisfaction, as they will achieve better sales, loyalty and more personalised customer services (Ogilvie et al., 2017). In short, our research helps companies to develop more targeted and effective communication strategies, which will have a positive impact on customer value and loyalty, as well as on the company's profits.

Limitations and future studies

The results of this study were based exclusively on the data contained in the database provided by the lead generation company. However, they would need to be validated against data supplied by other such companies (Jung et al., 2020), as well as other – even unstructured – data on user behaviour (Choudrie et al., 2021). It would also be interesting to contrast the results with other web data collection formats (Cruz-Benito et al., 2018). Also, the data was analysed globally without taking into account recruitment sources or methodologies, or different origins and social networks (Parekh et al., 2018). This additional information could enrich studies in this field. As indicated by authors such as Borges-Tiago et al., (2020) attitudes differ depending on the country that users come from. Our data was collected in Spain, and it remains to be seen whether its conclusions can be extrapolated to other countries and cultures and whether future generations will continue to behave in the same way (Altman and Bland, 1998). There is also no evidence of exploratory research into the sectoral clustering of profiles that enter their data online and whether there are differences in behaviour by generation or declared gender. It would be especially interesting for future research to examine how different profiles behave in terms of decision-purchase-post-purchase behaviour. It would also be useful to study the clustering of consumer profiles by sector and thus analyse how the resulting algorithm is affected by the false information entered, which would to help to devise mechanisms to correct or eliminate such practices. The particular casuistry of cohorts that are more prone to unintentional errors, such as older people who are more affected by health conditions and accessibility issues, leads to an ethical debate that could be explored further, namely that on mechanisms to avoid penalising the participation of these older users just because they might find it harder to read, write or remember information. In other words, it would be very interesting to look in depth at the ethical implications of systematically excluding or limiting the participation of certain users in prize draws and tests, simply because they may make mistakes due to health conditions, and to investigate why female members of the same cohorts do not seem to be affected by such difficulties to the same extent.

References
[Agrawal et al., 2011]
D. Agrawal, B. Bamieh, C. Budak, A. El Abbadi, A. Flanagin, S. Patterson.
Data-Driven Modeling and Analysis of Online Social Networks.
Web-age information management (Vol. 6897, pp. 3-+, Issues 12th international conference on web-age information management),
[Ahmed and Shirmohammadi, 2009]
D.T. Ahmed, S. Shirmohammadi.
An algorithm for measurement and detection of path cheating in virtual environments.
2009 IEEE international conference on virtual environments, human-computer interfaces and measurements systems, pp. 138-142
[Alalwan et al., 2017]
A.A. Alalwan, N.P. Rana, Y.K. Dwivedi, R. Algharabat.
Social media in marketing: A review and analysis of the existing literature.
Telematics and Informatics, 34 (2017), pp. 1177-1190
[Allcott and Gentzkow, 2017]
H. Allcott, M. Gentzkow.
Social media and fake news in the 2016 election.
Journal of Economic Perspectives, 31 (2017), pp. 211-236
[Altman and Bland, 1998]
D.G. Altman, J.M. Bland.
Generalisation and extrapolation.
British Medical Journal, 317 (1998), pp. 409-410
[Axelrod and Hamilton, 1981]
R. Axelrod, W.D. Hamilton.
The evolution of cooperation.
Science, 211 (1981), pp. 1390-1396
[Balint et al., 2011]
M. Balint, V. Posea, A. Dimitriu, A. Iosup.
An analysis of social gaming networks in online and face-to-face bridge communities.
Proceedings of the 3rd international workshop on large-scale system and application performance, ACM, pp. 35-42
[Baltar and Brunet, 2012]
F. Baltar, I. Brunet.
Social research 2.0: Virtual snowball sampling method using Facebook.
Internet research, 22 (2012), pp. 57-74
[Blackburn et al., 2014]
J. Blackburn, N. Kourtellis, J. Skvoretz, M. Ripeanu, A. Iamnitchi.
Cheating in online games: A social network perspective.
ACM Transactions on Internet Technology, 13 (2014), pp. 9
[Bolton et al., 2013]
R.N. Bolton, A. Parasuraman, A. Hoefnagels, N. Migchels, S. Kabadayi, T. Gruber, Y. Komarova Loureiro, D. Solnet.
Understanding Generation Y and their use of social media: A review and research agenda.
Journal of Service Management, 24 (2013), pp. 245-267
[Bonald et al., 2009]
T. Bonald, M. Feuillet, A. Proutiere.
Is the ''Law of the Jungle'' Sustainable for the internet?”.
IEEE INFOCOM 2009, pp. 28-36
[Bondarenko et al., 2019]
S. Bondarenko, O. Laburtseva, О. Sadchenko, V. Lebedieva, O. Haidukova, T. Kharchenko.
Modern lead generation in internet marketing for the development of enterprise potential.
International Journal of Innovative Technology and Exploring Engineering, 8 (2019), pp. 3066-3071
[Bondielli and Marcelloni, 2019]
A. Bondielli, F. Marcelloni.
A survey on fake news and rumour detection techniques.
Information Sciences, 497 (2019), pp. 38-55
[Borges-Tiago et al., 2020]
T. Borges-Tiago, F. Tiago, O. Silva, J.M. Guaita Martinez, D. Botella-Carrubi.
Online users' attitudes toward fake news: Implications for brand management.
Psychology & Marketing, 37 (2020), pp. 1171-1184
[Brosdahl and Carpenter, 2011]
D.J.C. Brosdahl, J.M. Carpenter.
Shopping orientations of US males: A generational cohort comparison.
Journal of Retailing and Consumer Services, 18 (2011), pp. 548-554
[Chambers et al., 2010]
C. Chambers, W.-C. Feng, S. Sahu, D. Saha, D. Brandt.
Characterizing online games.
IEEE/ACM Transactions on Networking, 18 (2010), pp. 899-910
[Choudrie et al., 2021]
J. Choudrie, S. Banerjee, K. Kotecha, R. Walambe, H. Karende, J. Ameta.
Machine learning techniques and older adults processing of online information and misinformation: A Covid 19 study.
Computers In Human Behavior, 119 (2021),
[Conroy et al., 2015]
N.J. Conroy, V.L. Rubin, Y. Chen.
Automatic deception detection: Methods for finding fake news.
Proceedings of the Association for Information Science and Technology, 52 (2015), pp. 1-4
[Cosmides and Tooby, 2016]
L. Cosmides, J. Tooby.
Adaptations for reasoning about social exchange.
The handbook of evolutionary psychology, pp. 625-668
[Cruz-Benito et al., 2018]
J. Cruz-Benito, J.C. Sánchez-Prieto, A. Vázquez-Ingelmo, R. Therón, F.J. García-Peñalvo, M. Martín-González.
How different versions of layout and complexity of web forms affect users after they start it? A pilot experience.
Trends and advances in information systemsand technologies, pp. 971-979
[Da Fonseca and Borges-Tiago, 2021]
J.M.R.R. Da Fonseca, M.T. Borges-Tiago.
Cyberbullying from a research viewpoint: A bibliometric approach.
Handbook of research on cyber crime and information privacy, pp. 182-200
[Dabija et al., 2018]
D.-C. Dabija, B.M. Bejan, N. Tipi.
Generation X versus Millennials communication behaviour on social media when purchasing food versus tourist services.
E+M Ekonomie a Management, 21 (2018), pp. 191-205
[Dabija and Grant, 2016]
D.-C. Dabija, D.B. Grant.
Investigating shopping experience and fulfillment in omnichannel retailing: A proposed comparative study in Romanian and UK of generation Y consumers.
The 21st LRN annual logistic network conference,
[Desai, 2019]
V. Desai.
Digital marketing: A review.
International Journal of Trend in Scientific Research and Development, 5 (2019), pp. 196-200
[Di Domenico and Visentin, 2020]
G. Di Domenico, M. Visentin.
Fake news or true lies? Reflections about problematic contents in marketing.
International Journal of Market Research, 62 (2020), pp. 409-417
[Dillman et al., 2009]
D.A. Dillman, G. Phelps, R. Tortora, K. Swift, J. Kohrell, J. Berck, B.L. Messer.
Response rate and measurement differences in mixed-mode surveys using mail, telephone, interactive voice response (IVR) and the Internet.
Social Science Research, 38 (2009), pp. 1-18
[Denzin, 1978]
N.K. Denzin.
The research act: A theoretical introduction to sociological methods.
2nd ed., McGraw-Hill, (1978),
[Escobar-Jeria et al., 2007]
V.H. Escobar-Jeria, M.J. Martín-Bautista, D. Sánchez, M.-A. Vila.
Analysis of log files applying mining techniques and fuzzy logic.
BT - New trends in applied artificial intelligence, pp. 483-492
[Fallis, 2014]
D. Fallis.
The varieties of disinformation.
The philosophy of information quality, pp. 135-161
[Fritsch et al., 2006]
T. Fritsch, B. Voigt, J. Schiller.
Distribution of online hardcore player behavior: (how hardcore are you?).
Proceedings of 5th ACM SIGCOMM workshop on network and system support for games,
[Gefen et al., 2003]
D. Gefen, E. Karahanna, D.W. Straub.
Trust and TAM in online shopping: An integrated model.
MIS Quarterly, 27 (2003), pp. 51-90
[Habib et al., 2019]
A. Habib, M.Z. Asghar, A. Khan.
False information detection in online content and its role in decision making: A systematic literature review.
Social Network Analysis and Mining, 9 (2019), pp. 1-20
[Huang et al., 2015]
S.-W. Huang, M. Suh, B.M. Hill, G. Hsieh.
How activists are both born and made: An analysis of users on Change.org.
Proceedings of the 33rd annual ACM conference on human factors in computing systems, pp. 211-220
[Huertas-Garcia et al., 2016]
R. Huertas-Garcia, J.C. Gázquez-Abad, S. Forgas-Coll.
A design strategy for improving adaptive conjoint analysis.
Journal of Business & Industrial Marketing, 31 (2016), pp. 328-338
[INE, 2020]
INE. (2020). Survey on equipment and use of information and communication technologies in Households 2019, INEbase, National Institute of Statistics. Retrieved 6/7/2022 from https://www.ine.es/dynt3/inebase/en/index.htm?padre=6898∩sel=6933
[Islam et al., 2020]
M.R. Islam, S. Liu, X. Wang, G. Xu.
Deep learning for misinformation detection on online social networks: A survey and new perspectives.
Social Network Analysis and Mining, 10 (2020), pp. 1-20
[Jung et al., 2020]
W.-J. Jung, S. Yang, H.-W. Kim.
Design of sweepstakes-based social media marketing for online customer engagement.
Electronic Commerce Research, 20 (2020), pp. 119-146
[Kapoor et al., 2018]
K.K. Kapoor, K. Tamilmani, N.P. Rana, P. Patil, Y.K. Dwivedi, S. Nerur.
Advances in social media research: Past, present and future.
Information Systems Frontiers, 20 (2018), pp. 531-558
[Karlova and Fisher, 2013]
N.A. Karlova, K.E. Fisher.
A social diffusion model of misinformation and disinformation for understanding human information behaviour.
Information Research, 18 (2013), pp. 573
[Keusch et al., 2019]
F. Keusch, B. Struminskaya, C. Antoun, M. Couper, F. Kreuter.
Willingness to participate in passive mobile data collection.
Public Opinion Quarterly, 83 (2019), pp. 210-235
[Kumar et al., 2016]
A. Kumar, R. Bezawada, R. Rishika, R. Janakiraman, P.K. Kannan.
From social to sales: The effects of firm-generated content in social media on customer behavior.
Journal of Marketing, 80 (2016), pp. 7-25
[Landis and Koch, 1977]
J.R. Landis, G.G. Koch.
The measurement of observer agreement for categorical data.
Biometrics, 33 (1977), pp. 159-174
[Lee et al., 2012]
K.-C. Lee, B. Orten, A. Dasdan, W. Li.
Estimating conversion rate in display advertising from past performance data.
Proceedings of the 18th ACM SIGKDD international conference on Knowledge discovery and data mining,
[Lenhart et al., 2010]
A. Lenhart, K. Purcell, A. Smith, K. Zickuhr.
Social media & mobile internet use among teens and young adults.
[Lin et al., 2021]
S.-Y. Lin, Y.-W. Chen, H.-C. Kang, Y.-J. Wu, P.-Z. Chen, C.-W. Wu, C.-S. Lin, F.-L.L. Wu, L.-J. Shen, Y.-M. Huang, C.-F. Huang.
Effects of a pharmacist-managed anticoagulation outpatient clinic in Taiwan: Evaluation of patient knowledge, satisfaction, and clinical outcomes.
Postgraduate Medicine, 133 (2021), pp. 964-973
[Lincoln and Guba, 1985]
Y.S. Lincoln, E.G. Guba.
Naturalistic inquiry.
Sage Publications, (1985),
[Lwin et al., 2016]
M.O. Lwin, J. Wirtz, A.J.S. Stanaland.
The privacy dyad: Antecedents of promotion- and prevention-focused online privacy behaviors and the mediating role of trust and privacy concern.
Internet Research, 26 (2016), pp. 919-941
[Maaß et al., 2021]
M. Maaß, M.-P. Clement, M. Hollick.
Snail mail beats email any day: On effective operator security notifications in the internet.
The 16th international conference on availability, reliability and security,
[Mealey et al., 1996]
L. Mealey, C. Daood, M. Krage.
Enhanced memory for faces of cheaters.
Ethology and Sociobiology, 17 (1996), pp. 119-128
[Menon et al., 2019]
R.V. Menon, V. Sigurdsson, N.M. Larsen, A. Fagerstrøm, H. Sørensen, H.G. Marteinsdottir, G.R. Foxall.
How to grow brand post engagement on Facebook and Twitter for airlines? An empirical investigation of design and content factors.
Journal of Air Transport Management, 79 (2019),
[Meredith and Schewe, 1994]
G. Meredith, C. Schewe.
The power of cohorts.
American Demographics, 16 (1994), pp. 22-31
[Meyers-Levy and Loken, 2015]
J. Meyers-Levy, B. Loken.
Revisiting gender differences: What we know and what lies ahead.
Journal of Consumer Psychology, 25 (2015), pp. 129-149
[Mintz, 2002]
A.P. Mintz.
Web of deception: Misinformation on the Internet.
Information Today, Inc., (2002),
[Moss, 2009]
G. Moss.
Gender, design, and marketing. How gender drives our perception of design and marketing.
Routledge, (2009),
[Nazir et al., 2010]
A. Nazir, S. Raza, C.-N. Chuah, B. Schipper.
Ghostbusting Facebook: Detecting and characterizing phantom profiles in online social gaming applications.
Proceedings of the 3rd workshop on online social networks (WOSN’10),
[Nambisan et al., 2019]
S. Nambisan, M. Wright, M. Feldman.
The digital transformation of innovation and entrepreneurship: Progress, challenges and key themes.
Research Policy, 48 (2019),
[Nickel et al., 2020]
K. Nickel, U.R. Orth, M. Kumar.
Designing for the genders: The role of visual harmony.
International Journal of Research in Marketing, 37 (2020), pp. 697-713
[Ogilvie et al., 2017]
J. Ogilvie, A. Rapp, D.G. Bachrach, R. Mullins, J. Harvey.
Do sales and service compete? The impact of multiple psychological climates on frontline employee performance.
Journal of Personal Selling & Sales Management, 37 (2017), pp. 11-26
[Pascual-Ezama 2020]
Pascual-Ezama.
Cheaters, Liars, or Both? A new classification of dishonesty profiles.
Psychological Science, 31 (2020), pp. 1097-1106
[Parekh et al., 2018]
D. Parekh, A. Amarasingam, L. Dawson, D. Ruths.
Studying jihadists on social media: A critique of data collection methodologies.
Perspectives on Terrorism, 12 (2018), pp. 5-23
[Parikh and Atrey, 2018]
S.B Parikh, P.K. Atrey.
Media-rich fake news detection: A survey.
2018 IEEE conference on multimedia information processing and retrieval (MIPR),
[Pennycook et al., 2021]
G. Pennycook, Z. Epstein, M. Mosleh, A.A. Arechar, D. Eckles, D.G. Rand.
Shifting attention to accuracy can reduce misinformation online.
Nature, 592 (2021), pp. 590-595
[Pennycook and Rand, 2019]
G. Pennycook, D.G. Rand.
Lazy, not biased: Susceptibility to partisan fake news is better explained by lack of reasoning than by motivated reasoning.
[Perez-Escoda et al., 2021]
A. Perez-Escoda, G. Baron-Dulce, J. Rubio-Romero.
Mapping media consumption among youngest: Social networks, fake news and trustworthy in pandemic times.
Index Comunicacion, 11 (2021), pp. 187-208
[Prediger et al., 2019]
M. Prediger, R. Huertas-Garcia, J.C. Gázquez-Abad.
How store flyers design affect perceived variety of retailers’ assortment.
Industrial Management & Data Systems, 119 (2019), pp. 382-396
[Poulos et al., 2020]
M. Poulos, N. Korfiatis, S. Papavlassopoulos.
Assessing stationarity in web analytics: A study of bounce rates.
Expert Systems, 37 (2020), pp. e12502
[Rao et al., 2021]
G. Rao, Y. Wang, W. Chen, D. Li, W. Wu.
Matching influence maximization in social networks.
Theoretical Computer Science, 857 (2021), pp. 71-86
[Rao, 2014]
V.R. Rao.
Applied conjoint analysis.
Springer, (2014),
[Rothman, 2014]
D. Rothman.
Lead generation for dummies.
Wiley, (2014),
[Sannon et al., 2018]
S. Sannon, N.N. Bazarova, D. Cosley.
Privacy lies: Understanding how, when, and why people lie to protect their privacy in multiple online contexts.
Proceedings of the 2018 CHI conference on human factors in computing systems, pp. 1-13
[Saunders et al., 2009]
M. Saunders, P. Lewis, A. Thornhill.
Research methods for business students.
Pearson Education, (2009),
[Saura, 2021]
J.R. Saura.
Using data sciences in digital marketing: Framework, methods, and performance metrics.
Journal of Innovation and Knowledge, 6 (2021), pp. 92-102
[Sharif and Zhang, 2014]
S.M. Sharif, X. Zhang.
A survey on deceptions in online social networks.
Computer and information sciences (ICCOINS), 2014 international conference on IEEE, pp. 1-6
[Shu et al., 2017]
K. Shu, A. Sliva, S. Wang, J. Tang, H. Liu.
Fake news detection on social media: A data mining perspective.
ACM SIGKDD Explorations Newsletter, 19 (2017), pp. 22-36
[Shu et al., 2020]
K. Shu, D. Mahudeswaran, S.H. Wang, D. Lee, H. Liu.
FakeNewsNet: A data repository with news content, social context, and spatiotemporal information for studying fake news on social media.
Big Data, 8 (2020), pp. 171-188
[Schulten and Rauch, 2015]
M.B. Schulten, M. Rauch.
Ready to Win? Generating high-quality leads through online sweepstakes and conquizzes.
Journal of Marketing Theory and Practice, 23 (2015), pp. 21-37
[Schuman and Scott, 1989]
H. Schuman, J. Scott.
Generations and collective memories.
American Sociological Review, 54 (1989), pp. 359-381
[Song et al., 2021]
C. Song, N. Ning, Y. Zhang, B. Wu.
A multimodal fake news detection model based on crossmodal attention residual and multichannel convolutional neural networks.
Information Processing and Management, 58 (2021),
[Sullivan et al., 2019]
E. Sullivan, D. Bountouridis, J. Harambam, S. Najafian, F. Loecherbach, M. Makhortykh, D. Kelen, D. Wilkinson, D. Graus, N. Tintarev.
Reading news with a purpose: Explaining user profiles for self-actualization.
Publication of the 27th conference on user modeling, adaptation and personalization, pp. 241-245
[Tandoc et al., 2018]
E.C. Tandoc Jr., Z.W. Lim, R. Ling.
Defning fake news: A typology of scholarly definitions.
Digital Journalism, 6 (2018), pp. 137-153
[Thakur et al., 2017]
S. Thakur, Er. Meenakshi, A. Priya.
Detection of malicious URLs in big data using ripper algorithm.
2017 2nd IEEE international conference on recent trends in electronics, information and communication technology (RTEICT), pp. 1296-1301
[Tripathy et al., 2013]
R.M. Tripathy, A. Bagchi, S. Mehta.
Towards combating rumors in social networks: Models and metrics.
Intelligent Data Analysis, 17 (2013), pp. 149-175
[Tuomi et al., 2021]
A. Tuomi, I.P. Tussyadiah, P. Hanna.
Spicing up hospitality service encounters: The case of Pepper™.
International Journal of Contemporary Hospitality Management, 33 (2021), pp. 3906-3925
[Valentine and Powers, 2013]
D.B. Valentine, T.L. Powers.
Generation Y values and lifestyle segments.
Journal of Consumer Marketing, 30 (2013), pp. 597-606
[Viviani and Pasi, 2017]
M. Viviani, G. Pasi.
Credibility in social media: Opinions, news, and health information-a survey.
Wires Data Mining and Knowledge Discovery, 7 (2017), pp. e1209
[Vosoughi et al., 2018]
S. Vosoughi, D. Roy, S. Aral.
The spread of true and false news online.
Science, 359 (2018), pp. 1146-1151
[Wendling, 2018]
Wendling, M. (2018). The (almost) complete history of “fake news''. Retrieved 30/11/2022 from https://www.bbc.com/news/blogs-trending-42724320
[Wu et al., 2022]
Y. Wu, E.W.T. Ngai, P. Wu, C. Wu.
Fake news on the internet: A literature review, synthesis and directions for future research.
Internet Research, 32 (2022), pp. 1662-1699
[Yadav and Rai, 2017]
G.P. Yadav, J. Rai.
The generation Z and their social media usage: A review and a research outline.
Global Journal of Enterprise Information System, 9 (2017), pp. 110-116
[Zhang et al., 2016]
H. Zhang, M.A. Alim, X. Li, M.T. Thai, H.T. Nguyen.
Misinformation in online social networks: Detect them all with a limited budget.
Acm Transactions On Information Systems, 34 (2016), pp. 1-24
[Zhang et al., 2020]
W. Zhang, W. Du, Y. Bian, C.H. Peng, Q. Jiang.
Seeing is not always believing: An exploratory study of clickbait in WeChat.
Internet Research, 30 (2020), pp. 1043-1058
[Zhang and Ghorbani, 2020]
X.C. Zhang, A.A. Ghorbani.
An overview of online fake news: Characterization, detection, and discussion.
Information Processing & Management, 57 (2020),
[Zubiaga et al., 2018]
A. Zubiaga, A. Aker, K. Bontcheva, M. Liakata, R. Procter.
Detection and resolution of rumours in social media: A survey.
ACM Computing Surveys, 51 (2018), pp. 1-36
Copyright © 2023. The Author(s)
Opciones de artículo
Quizás le interese:
10.1016/j.jik.2024.100510
No mostrar más