While fire season kicks in we talk of LinkedIn

In NSW and Southern Queensland, fire conditions are beginning, with yesterday seeing outbreaks. In the evening I had the privilege of speaking on ABC QLD radio, but it was not about fire. No. It was about etiquette on LinkedIn. Apparently a merciful break of fluff in the coverage, which people no doubt needed.

What I found fascinating about the topics we discussed was the push and shove that this platform faces between the professional profile and personal lives in how people use it. So I did a bit of research so that I didn’t sound like a complete duck and thought, well, I haven’t posted on my blog for a while, why not put it there.

I’ll cover the questions we discussed, the research insights I found into how LinkedIn is used, and the sparse information I found on the norms of interaction on the platform and how these may be changing. I’d observe that one of the reasons these norms change is connected to the changing dynamics of the broader ecosystem. Currently, this would include the backlash over Twitters rebrand to X and changing use (or disuse) patterns.

Social networking has transformed the process of building a personal brand, seeking challenging and fulfilling work, and advancing your career. The use of social media in the recruitment process has also grown exponentially. According to the literature, interaction on LinkedIn is focused around professional networking and career development.

The platform is designed to facilitate/afford how people make business connections, share their experiences and resumes, and find jobs. It is not the only platform where people do this, but it does seem to hold prevalence, at least in the circles I move in and the for its reported 950 million members sprinkled across 200 countries at the time of writing.

Research suggests that being active on LinkedIn does what it says on the box, supports career advancement. It also functions as a platform for B2B marketing, given that its functions are built around helping firms creating brands, build relationships and connect with existing and potential customers (relationship marketing).

As a feature, an active profile is probably the most important content project for any LinkedIn user, as it is the image that they project towards the wider LinkedIn community. Personal branding, including self marketing, reputation building and networking are key behaviours that social media platforms like LinkedIn afford. The platform also supports an individual user indicating availability for employment and to present their work experience in a way that can be ported across organisations and have endorsements from peers and co-workers of their work.

So when you look at someone’s LinkedIn profile, you are likely to know their education history, the probably that they are not a bot and a little about their socio-economic profile and reputational status. This point becomes relevant a little later, so keep it in mind.

So which question was most burning at the top of the show: Why are people sharing all their personal photos (babies, birthing, engagement, significant others… and making comments about everyone else’s) and where is the line on this platform? This question seems to be riding on the continuation of a trend of increased personal posts jumpstarted during the pandemic on the platform, and as mentioned above, changes in the platform ecosystem.

I’d suggest the continuance of this trend is probably a hodge podge of norms intersecting as people transfer what they do on more socially focused platforms in an attempt at authenticity and approachability, which is still an aspect of professional branding involving visibility and credibility signifiers. I found this piece bit of advice hanging prudishly in one article that attempts to put a normative compass around the overshare line. “Sharing your personal life on any social media platform puts you in a vulnerable position, but posting to LinkedIn as if it were Facebook can affect your professional reputation.” The observation I did make on air was that when posting we probably needed to remember that our job affiliation (and overlaps between personal and organisational branding) was being associated with the content we post. This is mostly different from how other social media platforms such as Instagram, TikTok and Facebook arrange and present information about individual users.

Next up, we talked about whether people were using LinkedIn like a dating app, to profile and hit up prospective candidates. The literature looking at the gender differences on impacts to men and women for being hit up, sexted or receiving sexually explicit content through LinkedIn suggests that women tend to decrease their active engagement with the app more than men in the face of these kinds of messages.

It appears that this is a real thing, because in August 2020 LinkedIn announced it was cracking down on unsolicited romantic advances and other forms of harassment on the website. They reportedly deploy machine learning designed to detect detect and hide potentially harassing messages from the recipient, who is then able to unhide, view or report the messages at their discretion. They also made their stance against the platform’s use as a dating site in their professional community policies. So it seems that the platform does not want to become the next Bumble.

The last couple of questions were about what functionalities and practices the site was best used for and possible shifts for it in the future. They really didn’t want to talk about the fires, hey.

In the literature I found research that maps along the classic strength of weak ties and job finding work by Mark Granovetter. But LinkedIn does this on gas using its contacts structure. Apparently, it’s about both the quantity and quality of ties because of the first, second and third level contacts network effects and having influential people or field experts at the first level of contacts. Doing so adds to the network of second- and third-level contacts people who are equally interesting to their professional field.

Another study, however argued that it was not the quantity of contacts mattered, it was all about frequency of usage that mattered for the receipt of career benefits such as sponsorship and job search assistance. Ah, the attention economy strikes again. You must be before eyeballs if you want to be seen and amplified by the networks you have curated in social media.

Broadly, professional benefits were found to be informational ones, such as timely access to relevant information and being referred for career opportunities. Combining the insights of the previous two studies mentioned, the authors who looked at informational benefits of Linked in observed that posting about work and strategically selecting ties (network composition) predicted higher informational benefits for users. Surprise.

Finally, where is linkedin likely to go from here? Well this one was a no brainer for me because I think that all the platforms we are using will move into augmented reality (AR) and virtual reality (VR) and attempt to build out the mixed reality and immersive reality experiences. I also think that those who are successful in this will attempt to build ecosystems of interaction around social, marketplace and creative/ professional activities rather than sticking to one stream. With this enhanced environment of interaction, I have no doubt norms and etiquettes will once again change. Who knows, maybe LinkedIn will move into professional matchmaking…

Disclaimer: Not written by GenAI… all typos, opinions and observations were made by the composite being that is me. Probably a partial cyborg, definitely digitally-enabled, sort of cute, of the human variety.

Exploring the intersections of Artificial Intelligence and Education

Today I had my first opportunity to share my research into generative AI and thinking through education futures with my new colleagues in the School of Education at La Trobe. It was a great experience and it got me thinking about the fact that there is so much need and desire in this field to talk about AI and its implications for teaching, learning and education futures. With my background as a sociologist of technology, it is almost like walking into a cornucopia of relevance rather than working away in my digital fringes.

Image CreditMF3d

I thought I’d take the time to share some of my formative thinking from engaging with the research on this topic here so that there is some kind of breadcrumb trail back to this point in time as I start moving into a familiar but new research space. Never fear however, I am still watching other frontiers such as Web3 (on which I am currently writing a book!).

First, I chose to grapple with the question of why we are talking about AI now.

The research, tech-focused and public discourse on the topic position Artificial Intelligence (AI) as an emerging and powerful technology with the potential to revolutionise various aspects of our lives. In the field of education, AI offers promising opportunities to enhance teaching and learning experiences, personalise instruction, and improve educational outcomes. However, it has already been flagged to raise considerations regarding ethics, equity, and the evolving role of educators.

In this blog post I’ve decided to delve into the world of AI in education, exploring its definition, historical trajectory, current applications, and the trends we need to consider. 

Let’s start with its historical trajectory.

AI in education has a rich history dating back to the 1980s. It has evolved alongside the growth of the commercial education technology (EdTech) industry and the influence of global technology corporations on education. This technology is intersecting and constituting with other technology trends including the Internet of Things (IoT), Big Data, machine learning, neural networks, platforms and applications. Some of the common threads of critique and conceptualisation across this complex of technologies are notions of ubiquity, platform capitalism and extraction.

From the early focus on simulating human intelligence and expert systems, AI in education has shifted toward data-driven approaches that leverage machine learning, neural networks, and deep learning. The convergence of AI research, edtech industry development, and data-driven policy has contributed to the current landscape of AI in education.

So what are its current applications and considerations?

AI in education, often referred to as Artificial Intelligence in Education (AIED), encompasses two main strands: the development of AI-based tools for classrooms and the use of AI to understand, measure, and improve learning. Examples of AI applications include intelligent tutoring systems, chatbots, and image/video generation tools. These technologies afford personalised feedback, adaptive learning paths, and real-time interaction. However, several considerations must be addressed, including concerns about truth, bias, equity, social justice, and ethical implications. It is crucial to fact-check AI-generated information, be aware of potential biases in training data, and ensure equitable access to AI tools and resources.

What do the tech trend types tell us to focus on?

There is an interesting comparison of trends across the 2021 and 2023 EDUCAUSE Horizon reports that highlight the rapid evolution of key technologies and practices shaping the future of education. The 2021 report highlights Artificial Intelligence, learning analytics amongst the top five trends, while the most recent 2023 report identifies AI-enabled applications for personalised learning and generative AI as having the potential to transform teaching and learning experiences. 

What’s in focus for me in all of this?

I am currently focusing on the opportunities and challenges of generative AI for education futures, which is technically referred to as Generative Pre-Trained Transformer (GPT) AI. Generative AI is conversational AI technology that is activated through user-supplied prompts and conversational turn-taking between the user and the AI. It uses supervised and reinforcement learning techniques to understand and model human and non-human languages. Examples in education include chatbots, intelligent tutoring systems, and image and video generation tools. I’ll have more to say on this in the future.

A quick recap on what I think the take home points are.

AI holds significant promise for education, offering innovative tools and approaches to enhance teaching and learning experiences. By leveraging AI technologies, educators can personalise instruction, provide real-time feedback, and create adaptive learning environments. However, careful consideration must be given to the ethical implications, potential biases, and the impact on equity and social justice. It is essential to navigate this evolving landscape thoughtfully, ensuring that AI in education aligns with the goals of fostering critical thinking, nurturing creativity, and empowering learners to thrive in an AI-driven world.

Senses of/in the city @ ISA2023

I’m thrilled to be attending the World Congress of Sociology here in Melbourne this week. Being in a bevy of sociologists (I wonder what the collective noun actually is? Swarm?) is unusual and heady. I am usually working and collaborating within interdisciplinary teams and am rarely in the thick of sociological heartlands.

For this conference, I’m presenting on some exploratory work that combines my research into playful creative cities with mixed-reality environments and digital pleasures. It is part of what looks set to be a very interesting session on “The Sense of Data and the Data of Sense: Bodies, Technologies, Spaces“.

I’m thrilled to be collaborating on this work with some excellent colleagues, including Naomi Smith (digital pleasures and desire lines in the city), Jacinthe Flores (creative cities and immersive spaces) and Luke Heemsbergen (mixed-reality environments). The backbone of this work was developed during my recent creative cities research residency, with the many conversations that took place during it as inspiring as the setting itself in Olot, Spain.

You can read a the work in progress paper here: Senses of/in the city: A speculative and conceptual exploration of sensory spaces of play in the digital city

The article abstract below will give you a taster for what to expect.

The digital city is a space of sensory play that contains the visceral embodiment of digital pleasures (Smith et al 2019), extended or mixed-reality environments (Heemsbergen 2021) and aesthetic encounters (Maddox et al 2022). We observe the shift from only the platform economy to the mediated experience of a city, and articulate the ways in which that can be achieved (via the big platforms and otherwise) in social digital-physical connections. An embodied sense of the social in these spaces can be derived through unpacking conceptual work on social effervescence (Olaveson 2001), sensory playfulness, and digitally-mediated intimacies. We characterise these social moments of interaction, engagement and participation through intensities and immediacies of experience that involve intention and symbolic focus. Focusing on the sensory, playful, and digitally intimate cities space can also be understood as a way of inscribing desire lines in urban environments that can often be hostile to pleasurable and non-commercial forms of engagement (Smith and Walters 2018). Our approach creates a vector between the urban geography of a city, its digital architectures and a playful and pleasure-ful built environment that speaks. We argue that these vibrant moments of encounter in the digital city can drive creativity, place-making and a sense of belonging that manifest in localised ways. 

How should our politicians deal with online abuse in social media?

Social media is an environment where we share content with each other, consume news, follow our friends, pursue our interests, tell jokes and share memes, and engage in political discussion (by text, images, videos, .gifs and emoji). In social media, public figures and influencers vie for attention and engagement in an environment of high visibility and social contention.

Researchers talk about social media as a space for political engagement, where people discuss and engage with politicians and political issues as well as undertaking social activism (sometimes referred to as connective action and, perhaps more cynically, clicktivisim).

However it’s not only people that are active in this space. We have all sorts of technologies influencing our discussions and information exposure online. These range from algorithms that recognise patterns and target ads to bots that recirculate content or interact with users. There are also more uncanny or malicious user-produced experiences that shape our information and news environment such as deep fakes and disinformation.

Beyond this, political dialogue and activity takes place within the context of digital cultures. Some of these cultures are toxic and, despite the many benefits, there are risks for being active online. Public figure activity on social media brings with it visibility but also becomes a lightning rod for hate and extremism associated with social instability. 

In summary, the online environment is reactive and aware and characterised by an attention economy, with a side serving of backlash. Attention across social media occurs when people, monitoring and tracking technologies, and bots recognise and respond to online interactions, events and public figures. Public figures such as politicians use social media to be visible and engaged with constituents, share their position on topical debates and build their reputation in the public domain. Both their rise in public esteem and downfall is captured in social media, along with public backlash.

Online risks are quite broad and have some differences to those we face in person. Risks faced by people active online include cyberbullying (including death threats), trolling, impersonation, deep fakes, pile-ons, bot attacks, disinformation, targeting and scapegoating, hacking and leaks, and image-based abuse. These become intensified and occurring at scale for public figures such as politicians.

In the light of Jacinda Ardern’s recent resignation as New Zealand’s prime minister, we must consider how online abuse affects our public figures and politicians

Researchers at the University of Auckland found that she faced online vitriol at a rate between 50 and 90 times higher than any other high-profile figure in NZ over the time of social media observation. In their analysis of what was said, they pointed out that misogyny was a key part of it, particularly because Ardern attracted backlash for being a left-wing woman in power who “symbolically or otherwise was taking a number of steps to undermine structures of patriarchy, racial hierarchies and structures within society,”. How ugly is that? Yes, toxic masculinity and the manosphere have a lot to answer for.

NZ Police reported that threats against the Ardern had nearly tripled over three years and that anti-vaccination sentiment was a driving force of a number of threats.  In Australia, the Australian Federal Police reportedreceiving more than 500 reports of threats last to the safety of politicians – including online threats –  last year. 

The pandemic only exacerbated online hate and conspiracism. Within the social media ecologies studied by the authors of the report on mis- and disinformation in Aotearoa, New Zealand, key individuals and groups producing mis- and disinformation capitalise on growing uncertainty and anxiety amongst communities, related to Covid-19 public health interventions, including vaccination and lockdowns, to build fear, disenfranchisement and division.

With rhetoric from international groups trickling into Australia – particularly in Melbourne – lockdowns put the population into a pressure cooker that was intensified by a media environment of uncertainty, disinformation and misinformation. Mis- and disinformation is transmitted within and across platforms, and often very rapidly reaching large audiences, who have likely been targeted. We know that this occurs in the political domain because of the revelations from the Cambridge Analytica scandal.

The authors of the report on mis- and disinformation observe that mis- and disinformation is also particularly targeting and scapegoating already marginalised or vulnerable communities – for whom distrust of the state is the result of intergenerational trauma and lived experience of discrimination or harm, which can increase engagement with conspiratorial explanations and disinformation.

In Australia we saw this social rent come to light as the far left and far right converge on issues, which inflamed social media activity around these topics. We also saw the online polarisation spill into the streets with protests organised through social media.

The point I wanted to make clearly in a recent RMIT expert alert on this topic is that we need to remember that despite the ‘new normal’ and ‘post pandemic’ messaging, we are still experiencing the pandemic and the intensity is still very much alive. 

This means we still have a great deal of social instability, which increases the risks that public figures will receive online abuse (beyond the usual disagreements and name calling). Legally, the ‘serious harm’ threshold for adult cyber abuse investigations is set deliberately high so that it balances freedom of speech, or legitimate expressions of opinion, against the need to protect everyone’s ability to participate online.

People are fatigued by the pandemic and rising mental health issues, including our politicians. However the public also have a low tolerance for government intervention, so our politicians will continue to be in the firing line.

There are many strategies that we can take to deal with online abuse targeting adults, but for politicians it’s more complex as they are public figures whose real names and professional (and personal) lives are in the (social) media. While platforms may deploy content moderation strategies and enforcement of their terms and conditions by banning people, how real is an online threat or abuse to the personal safety, emotional safety of the people who are on the receiving end? 

The Dangerous Speech project defines “dangerous speech” as any form of expression (e.g. speech, text or images) that can increase the risk that its audience will condone or participate in violence against members of another group. So that’s a fair bit more than an ego battering or an attempt to take someone down a notch or two (as is the Australian way).

We need to collectively keep this discussion going about how we can support a healthy digital political domain and curtail its more toxic aspects. Politicians and those quality people who we would like to enter into political life do need preparation to go into the social media ring. Many may now have grown up in the digital environment, sometimes referred to as digital natives, but operating as a public figure brings vastly different risks. 

Digital frontiers, digital cities and the metaverse

It has been a long time between drinks, my friends. I haven’t blogged for quite some time. However I have finally hit a piece of experimental thinking that needs to be expressed in a fluid free form rather than something targeted to particular audiences. If you like it, then you are my audience. At the very least, I write here to be creative and to write for me, for my thinking.

Of recent times, my research practice and writing has been focusing heavily on Web3 spaces, but also on the digital city. You can see my latest work on digital infrastructures for the city here. For a recent speaker slot at the LawTech Summit, I was invited to speak about Web3 and the Metaverse. In the digital infrastructures report, the discussion of the metaverse and its potentials was led by Kelsie Nabben, whose excellent work on the topic can be found here.

As I researched thought around the topic, I’ve realised that there are so many different visions and angles into what the metaverse is and what we can use it for. I thought a great write up with a slightly different angle to Kelsie’s essay was presented by Hatch Quarter on the topic. This essay is useful because it navigates well through the extended environment discussion and signals that it is not only e-commerce, socialising and the gaming industry that is pushing the metaverse, which was well highlighted by the Grayscale Research report on the concept, but also education and fitness.

I thought about all the different discussions and writing that I’ve had and done over the last 6-8 months while working on the Digital CBD project and also on the NFTs, DAOs, cryptocurrency and blockchain technology (decentralised ledger technology) we cover. I started pondering what a sociologist of technology would speak to in this environment… or more particularly, me, what I would speak to.

My research practice was begun in researching the digital community surrounding reptiles and snakes, aka ‘the Herpers’, that I researched and wrote about for my PhD research, then my book. This is my first book and it profiled a community, who were also early adopters of the internet, and how to research such digital communities. Since this book, I have articulated different aspects of research interest that include the study of the communities surrounding cryptomarkets and cryptocurrencies. These frontier socio-technical spaces are highly experimental.

I’ve also built a collaboration with wonderful scholars surrounding digital pleasures that articulate another type of digital frontier more focused on the human-technology encounters aspect. Thinking of the body as a digital instrument that is played through digital media to achieve altered states has been a wonderful creative space. This centering of the body and pleasure has also hooked into my emerging interest in play and gamification. You’ll need to keep up with me on this because these themes resurface when I begin thinking about what the metaverse is and what it could be for us as a built (build) environment.

The notion of digital cultures of care and care more generally has pervaded my work and provides an intellectual link, in some ways, between the digital pleasures research and the EyeStory project.

A research collaboration that has been going on for quite a while now that focuses on play, is the EyeStory project. This project has involved digital storytelling and an interdisciplinary collaboration that has allowed my research to extend to working with children, app design and collaborating with researchers whose expertise is in animation research, game design and optometry.

These projects have all involved some form of ethnographic practice and I have had the pleasure of winding them together into a book chapter on ethnography and digital society that will come out in 2023 in this Sage Handbook on Digital Society. But how do all of these threads weave together in my most recent work? For the digital city report, I foregrounded the importance of play and creativity for enlivening the vibrancy of a city, alongside the more traditional analysis of work and innovation as drivers of a city’s engine.

I am now starting to feel ready for my second book and I can see it forming along the lines of how we can incorporate insights from my research into digital frontiers into a diverse range of research. My most recent affirmation that my vocabulary and conceptual thinking had value to work beyond internet scholarship was at the very inspirational Creative Cities Symposium at Biella Italy. Aside from being incredibly inspired by the venue, the Cittadellarte, I was deeply engaged by the research into Creative Cities, Creative Industries and their links with urban regeneration and social policy. I went there to present from the Digital CBD survey, a city-wide survey led by Annette Markham and a collaborative program of research drawing together DERC and BIH scholars. The discussions I was able to have that helped urban geographers, social scientists and policy researchers to move their thinking into digital frontiers relevant to the focus of their work were really serendipitous.

This was my first international conference after the long lockdowns. I went for many reasons, but one of them was to pop my head out of the Australian bubble and see how Europe had faired through the pandemic. In many of our discussions, the pandemic impacts loomed large and it was enlightening and of course traumatic to hear how different countries and people had faired. I still have no words for this, but believe deeply that until we have begun moving through the grief and trauma of the pandemic, we will struggle to imagine a different future and to shape the future we have incoming to be conducive for us to thrive in.

This thought and point takes me to my LawTech Metaverse presentation, because here I am actively saying the next wave of our future is incoming and we need to engage and shape it. For me, the metaverse is the next logical step that we will take, or something of its ilk. We are currently operating our social life and digital lives largely through the relational spaces of social media. These platforms shape social connection and eCommerce through what they afford. Digital environments are now deeply embedded in our physical environments and we use a range of technologies to connect across place and virtual space. For example, our smartphones are pocket computers that allow us to overlay digital information onto physical environments, we can locate ourselves through them, visually record space and sound, connect with people and do so many other activities. When we game online, we can also experience immersive environments through VR technologies. So what will this mixed or extended reality space of the metaverse look, feel and behave like?

I don’t think we’re really looking at much new to start with, as most of the technologies, including AI/ML and IoT are in-play and starting to mature as a connected ecosystem. Web3 is adding specific affordances through technical infrastructures such as payment networks, decentralised finance, sovereign goods, decentralised governance and portable identities. The GrayScale Research report I linked to earlier provides some really helpful discussion on this, but also the recent McKinsey report on Value Creation in the Metaverse seems to be getting some good air-time on this topic. These are some of the highlights in the report if you want the TLDR version. I’ve also been inspired by the 20 day Web3 festival being held by the House of Beautiful Business, which has pulled together some great thinkers on the topic and fascinating metaverse companies, such as Journee.

So, what do I have to add to all this fantastic thinking? I thought I’d keep it focused and look at the implications for people, data and what risks were involved. It seems that demystifying this area and the terms used for a Web3 version of the metaverse would also be useful. Now I do not pretend to have technical expertise here, but I have the fantastic opportunity to engage with the latest developments and events in the Web3 space through my work colleagues at BIH. So, I’ve got a bit of a handle on the basics. But I’m not going to do that work for you here. If you’re an industrious active kind of reader, then the trail of links I’ve provided in this text will take you where you need to go.

For the presentation I used Kelsie’s Metaverse essay linked above to produce a word cloud that should get you started on the associations that this term has with digital life and the city. I’ve popped it here so you can have a dig through the grab bag of words that are usually associated with the topic.

My first provocation was to ask what questions we need to ask when thinking about people and the metaverse. I started with the ones that we’re asking right now on the social impacts of social media. For example how we are being conditioned in these digital spaces through social engineering, distraction and dopamine responses. We already know that social norms are evolving through our digital interactions, with my recent media engagement on friendship and social media for international friendship day speaking loudly to this. Not to be a prophet of doom, I like to keep things balanced and ask questions about playful and pleasurable interaction in the metaverse, particularly given that online gaming will be a bit motivator of metaverse adoption. Drawing back to the digital pleasures research, I was inspired by conversations with my collaborator, Naomi Smith, about ASMR and how we are now looking at multichannel experiences of it. When you bring in our work on binaural beats as digital drugs… you can see the amplification of pleasure and play moving into another experiential realm. Place these visceral body experiences in an immersive environment, and up the sensory input, and you are really starting to look at digital pleasure experiences in a very different way.

For me, the next significant area of focus is on the rich/thick data that is produced by both human and non-human actors in this environment. Sentience needs data through which to interpret and learn, as much as interactivity and liveliness produce data. I want to know who owns the data we would produce in such spaces and also, who uses it and what we can learn from it. In my thinking around data, I really enjoyed unpacking the question of how we remember in the metaverse, specifically in terms of archiving. Because if you can’t remember, you don’t learn and you can’t generate wisdom. So the knowledge holding in this space is a public utility that is crucial to our ability to build conducive strategies for thriving into the future and for adapting the space to meet our needs and contain our excesses. The internet has the internet archive, and with its wayback machine. Online communities, particularly those I encountered in the cryptomarkets space, publicly archived their own spaces, but they were also actively archived by researchers and no doubt by law enforcement.

Finally, in the talk, I pondered the risks in a Web3 Metaverse. The risks that I highlighted reflect my research wanderings. I kicked off with the key point working the Web3 space now and a persistent question for the early cryptocurrency communities that I encountered during my research in the dark net/cryptomarket space. This question is around how we deal with the grey zones that arise within policy and regulation around emerging technologies. It also links into the second question as to how we seek accountability but allow for experimentation in business models that plagues Web3 start ups and their reputation. The words scammy and Ponzi scheme are thrown around regularly in the web3 space, so I decided to play a little with the work of one of my favourite scholars in the space, Lana Swartz, on networked scams. She built this understanding of scams in the Web3 space through the Wild West context of ICOs in 2017. From her work, I covered off the two basic scams – the ‘exit scam‘ and ‘pump and dump‘. The exit scam I already new from the crypto market space. What I enjoyed about Swartz’s work up of the networked scam is this sense of a scam as sometimes being overt, and legal, business practice or just more mundane.

Swartz observes that ‘scams are capitalism out of place: what gets called a scam is used to perform boundary work that delegitimates certain forms of economic activity (and exploitation) and legitimates others.’ I love this definition and feel that it is appropriately slippery and embracing of the smoke and mirrors nature of this space.

As she observes, and I have heard many times, most blockchain projects are vaporware and punters are attracted to cryptocurrencies and DAO projects in part to make money – the get rich quick hustle. I’ve written a little about this culture here. The next point Swartz makes on this outs a piece of crypto argot that I continually re-encounter in everyday discussions with people who trade, invest in or speculate around cryptocurrencies, hodl.

She quotes a common refrain that ‘it’s not a scam if you hodl long enough’—which she interprets as meaning ‘that if enough people hold (or “hodl,” in crypto argot) their investment long enough, if they believe or suspend disbelief long enough, the promised future will rush to meet reality.’ However, the contradiction that these traders and speculators sit in is that, as Swartz points out, ‘no one wants to be the only one holding a worthless asset that has already been dumped; no one wants to be the only one holding out for a future that will never come.’

I would usually paraphrase these quotes but in this case she writes it best. I can assure you this is an authentic sentiment that I hear expressed on many an occasion.

I finished off the presentation with the point that as we move forward and commercially viable use case are generated for the metaverse concept, we can either have more of what we have now, with a Meta version of the metaverse, or we can embrace and shape a more diverse ecosystem with social values that we care about. These being, in my mind, a sustainably and socially, environmentally and economically regenerative future that appropriately harnesses the inclusive affordances of a digital environment whilst managing and mitigating for its risks and unintended consequences.

Now, this was a long essay. So I’ll be very impressed if you stayed with me to the end. Sadly, I don’t know what kind of world we are going to create through these technologies, but I would like to engage you in a call to arms to actively shape it. I am telling you now, it is happening. I want to see you care enough to engage.

The great de-acceleration

Apparently we are living in times of the great acceleration, with quantum computing and artificial intelligence. But the question is, who are we in these times and what do we value? We view the past as characterised by simplicity and that now we are moving through a time of complexity and polarization of political views and economic circumstances.

I would put forward that we are moving backwards, towards a conservative sense of security and control despite the increased pace and breadth of what we can do with technology and online. We have the optimism and idealism of Silicon Valley conflicting with the adolescent stage adoption that we have in how we use and express ourselves through social media and the digital public domain.

Is change scary and bewildering? Are we afraid of being on the wrong side of history? Can technology solve our problems rather than enabling our expressions of the self within the flawed moment that we are. These are questions that we ask without knowing that we are.

Healthcare, education and how we consume information will be profoundly shaped through artificial intelligence. However we are at the early stages of this. How do we imagine our future in this context?

Over time we will be in much more intelligent systems. Will we be more productive and efficient or will we be on a beach siping cocktails? What is human potential and what do we see as our ‘destiny’, if there was to be one.

Climate change tells us that we are over consuming our welcome and that our relationship to nature is grounded in competition and control rather than flourishing concept of collaboration and living with. Safety and risk is troubled by our depth of interconnection and global mobility.

Nothing is given. We want prosperity and improved quality of life. While we see the benefits, how can technology take us there? What do we need to understand on the spectrum between zero and one to realize this ‘there’? We don’t yet understand how. Simulations versus real world effects remain hypothetical.

Is progress our goal when there is no teleology. The modern agenda is fragile and only a desire.

Does privacy matter for us and is it a human right? How does trust factor in?

If we frame ourselves as stewards, where does responsibly lie? Is getting rich about owning the tools that profile mass populations or about providing mechanisms of privacy in response to the mass fear of the loss of the liberty of the self?

Is there a balance between encryption for privacy, platformised business models, monopoly behavior and forms of governance and global regulation. How does innovation flourish across these tensions?

Knowledge asymmetry has been the underpinning aspect of continued inequalities, particularly in its relationship to power. More so, however is its relationship to capitalism. Economic power is deeply imbricated with social stratifications such as class. For example, systemic reinforcement of capitalist drives towards economic growth makes a laughing stock of the concept of trickle down economics. The rich get richer and the poor get poorer. Can we ask, what alternative models may be possible?

Does our next generation need to adapt to technology? They have grown up with it, what does this mean?

How do our visions of the future replay our past rather than construct an alternative, adaptive and resilient future?

With this thought stream I have asked questions. In doing so the narrative is not aiming for coherence. Instead, I ask you, is getting faster the way to get to where want we to go? While the literature points to accelerationism, I’d like to point to the regressive forces of division, inequality and greed (or over consumption) that seem to be eating us out from the core. A future world needs scaffolding and care from the present one. Seeking to let us destroy what we have, provides no bridge to the future. I am not a fan of nihilism.

Slow down and let diverse voices and overlapping values have the time to work themselves out. We mature through our technology rather than jump from the past to the future.

There is no way to avoid the uncertain messy present. It is our only line to what the future brings. Let’s not forget it.

EOI – Funded PhD position: Digital Ethnography of Children with Vision Impairment (Update)

Excitingly, I’m involved in the opportunity to provide a funded PhD position for someone to join our interdisciplinary team at Deakin and support a Digital Ethnography of Children with Vision Impairment.

You can find out more about the research through this video and our website where we are piloting a digital storytelling application.

EOIs are due July 30th, 2021.

The opportunity is a funded PhD position for someone to join our interdisciplinary team at Deakin and support a Digital Ethnography of Children with Vision Impairment project. 

  • The deadline for full applications is August 9th, 2021 

The HDR candidate would need to commence in 2021. 

Access to EOI form: Faculty Expression of Interest form 

Candidates please send their initial interest to Rosemary Woodcock and myself so we can arrange a time to talk through the project and your experience.

Instructions on how to apply for HDR candidature here: https://www.deakin.edu.au/research/become-a-research-student/how-to-apply-research-degrees

About the project:

This PhD project involves the development and application of an ethnographic probe to examine the journey children with a vision impairment undertake from first diagnosis through treatment. The project will provide insight into the lived experiences of visually impaired children that will inform educational media for research stakeholders such as Vision Australia. The research will be conducted within the context of an interdisciplinary research team and situated within the disciplines of the Deakin Motion Lab (DML) through its focus on advancing visual methodologies and gamified and playful approaches to engagement. The project will also build upon child-centred methodologies that can engage families remotely and is designed with COVID19 conditions in mind through its use of a digital methodology to allow connectedness. The probe at the heart of the research methodology involves a toybox and an app that are designed to engage young children and provide prompts for journaling and digital story telling.

About the people:

We are an interdisciplinary team from DML (a creative research hub within the School of Communication and Creative Arts), the Social Sciences, Computer Sciences and Optometry.

About the PhD student:

We are looking for someone interested in understanding the relationships between digital data collection, ethnographic practice and an interest in applied research in the area of vision health and engaging children through play.

Activites involved include

  1. involvement in the qualitative data collection for the methodological pilot,
  2. engagement with families and children,
  3. working within interdisciplinary team
  4. Engaging in research outputs such as publication and multimedia education outputs
  5. Codesign workshops.
  6. App development (we have a tech person doing this but the whole team is engaged in how the app works and the user experience.

Please share this opportunity across your networks and if you are interested, email the team our me a.maddox@deakin.edu.au

Social media and community building

This blog is a response to the questions Seriously Sassy podcast host, Earvin Cabalquinto sent me in preparation for discussion about the role of social media channels in mediating communities. Now one of my research stomping grounds is community theory and practice, but I have never had to think about how to translate these insights for students who are making social media. I found out that I had something to say.

The questions kicked off with a bit of a historical rear view, asking what the definition of a community was before the advent of social media channels. I pondered this, as one does, and then just went totally old-school based upon my disciplinary roots in the Chicago school of urban studies.

Originally a community was defined by physical proximity and in person relationships. Think a neighbourhood or local sports organisation. Communities offered a sense of shared identity, a buffer or safety net for its members (social support) and a unique culture which informed who were thought insiders and outsiders.

The boundaries of a community at its very fundamental level are expressed through its coordinates in time and place, as much as by its geographic or social materiality. They also were the first formations of complex sociality, beyond direct kinship ties for example. Anthropologist, Jolynna Sinanan, describes this as ‘the village’.

To help us translate from place-based communities to online communities, Barry Wellman coined the idea that computer networks were social networks, early on. 

Online communities were then first defined by sociability articulated through co-presence on a platform. The platform became the site of co-location and co-presence could be synchronous or asynchronous. The big difference here was that people didn’t need to be in the same geographic location and have known each other through in person contexts prior to being a part of an online community. Community forums were, and still are, a common way to think about communities online.  

Digital communities are a bit more location-gregarious when it comes to online interaction and exchange. Instead of defining them by a specific platform, I draw on work that refers to them as foci of activity groups. Foci of activity groups are where the interlocking relationships and overlapping values of a community begin from a mutual interest in a topic or activity. This interest is what draws people initially into a community, however the complexity of these relationships builds into a critical mass and density of interconnections that holds a sense of continuity over time. This is what makes communities different to social movements online.

When we think of online communities, we think of sticky networks over time that support the development of a group identity, roles and norms (Cohesion & Belonging). In these instances, the people participating may change but the collective persists.

The next question lined up was about whether the advent of social media channels shaped the formation and maintenance of a community. I would suggest that it has shaped the re-formation and maintenance of online communities, given that communities can occur within one platform such as Tumblr or across multiple platforms. Again, I begin with a basic definition to outline how the configuration of social media reforms the basic shape of an online community.

Social media is a relational technology defined by user generated profiles affiliated with social networks of friends and followers. Communities emerge through practices of peer-support and connection that are organised through hashtags or other social media organising principles and collectivising practices such as shared memes and the development of a common language or digital argot. Social media also foregrounds visual cultures and animated content.

Often communities expressed through social media are trans-modal in that they engage with each other across multiple modes of interaction (trans media).

They are also generally self-organising however are usually subject to and shaped by content moderation by platforms. In response to these curating and surveillance practices, social media communities can patch together permissive spaces where their common interests and connection can occur. They are always shifting and responding to environmental constraints and emerging social media spaces.

Now that the basic premise of the role of social media in mediating communities has been covered off (ha ha), we turn to the classic question of why a soul would engage in a community online in the first place. Generally, it starts from a place of seeking.

The benefits of being a part of an online community include acceptance, a sense of belonging and being able to engage with people that you share a common interest with. They get you and you get them. 

It’s also a space for creating your own identity narrative in ways that respond to the culture of the community. If the community is online only, you can safely experiment with new ideas and expressions of yourself in a supportive environment. 

If the community spans place based and online interactions, you can keep in touch and continue the conversation even when you can’t be together. This is about being able to maintain relationships and connection to the community experience.

Communities hold forms of social capital and tend to support social mobility that means you may have greater access to targeted and community vetted information, experience social support and get exposure to people and places you may not have access to within your own local and geographical context. The increase of both strong and weak social ties may mean more opportunities become available for you. This premise was first introduced by Mark Granovetter through the strength of weak ties hypothesis. You may have heard of the six degrees of separation principle? Or the Kevin Bacon index? Same idea. 

Now the inter-connectivity of digital communities is not always an awesome feature and can lead to challenges. Just because an idea, sentiment or act can flow along social networks does not mean it’s a good one.

The challenges occur when a community culture is not inclusive, where practices including hazing of newbies turn toxic. The increasing polarisation of views that we find online mean that collaboration, compromise and understanding the other’s point of view become a rare experience. Filter bubbles are argued to reinforce narrow perspectives into a situation or event or other people’s lives. Consequently, sometimes people think that just because they say it online, it doesn’t matter in “the real world”. Hate speech and ranty combatative behaviour is rife and at times unchecked. It can have devastating consequences.

Engaging within a community takes time and it can swallow up a lot of your available time. FOMO is a killer.

Everything you do in an online community may be a product of the synergy of your creativity with the community culture and people, but the platforms you do this through claim ownership of your content. 

Privacy challenges, platform surveillance, and social surveillance and regulation occurs, while hacks and scams abound, making it difficult to know who or what to trust or what will be done with your personal data.

So what next? If this is it, is this all we’re ever going to have? Earvin asks what the future of communities online is, especially in an era wherein online platforms harvest personal data. Well, the data ownership and personal data awareness rant is strong in me, particularly as I attempt to help my students critically think about what they are doing online. However, I got a bit vague here. No idea why.

It’s difficult to say, I go, but there is a data sovereignty movement which raises awareness and advocates for personal data ownership. In the data justice area, the idea that communities can create platforms that mean they own their own data and can decide collectively what to do with it is an interesting idea.

Government and corporate surveillance is rife on the clear web. I study communities who reject this and move to the anonymous spaces of Dark Web. While the site admin still gets access to all the digital traces of users in their platform and social surveillance occurs, people’s identities and practices are not being on-sold to advertisers and third-party data brokers.

So finally, Earvin goes out on a flourish with the questions and asks whether we can do this thing differently. He asks in what ways we can re-imagine and produce an equitable and progressive online community. Now that one had me momentarily stumped… until I just went back to basics and thought about all that I’ve seen and read. For me, it comes down to the following.

A community can liberate or suffocate a person, and usually does a bit of both. Foundational principles of tolerance and diversity need to be at play in both the way people engage with each other and what the platforms of interaction afford. We must re-imagine online communities from the code up. That way we can embed shared and overlapping values into the digital architectures and social practices. We must ask of the features a platform makes possible, how does this afford, for whom and under what circumstances (Thank you Jenny Davis). We must ask of ourselves, what do we want our communities to be, for whom and under what circumstances. I argue that social inclusion creates richer, more enduring and innovative communities that will lead us out of a self-destructive future. 

The connected vehicle: data ownership and cybersecurity

In the HQ TechTank mobility series, one provocation I respond to is the question of the social impacts of a fully connected self driving car, where the computer and the car are connected to other cars (V2V), the internet and to transportation infrastructures (V2I). What could possibly go wrong? This blog will consider the question of data ownership and cybersecurity questions such as hacking autonomous vehicles.

In terms of car ownership, autonomous vehicles will initially be expensive and only those in the luxury market are likely to own their own vehicle. The greater majority of people are likely to access AVs through fleet access and a sharing economy structure, much like GoGet in Australia.

There are several implications that are raised if drivers do not own the vehicle they are travelling in and are not responsible for what the vehicle does. If the users do not own the autonomous vehicles they travel in the questions raised for me are who is liable for accidents (insurance and legal implications) and who owns the users’ activity data. The following case study will focus on this question of personal data ownership.

Case study: Personal data ownership

Ratnam (2019) accounts that a car can generate about 25 gigabytes of data every hour and as much as 4,000 gigabytes a day, according to some estimates. Drawing on data from consulting business McKinsey, he anticipates that the data trove in the hands of car makers could be worth as much as US$750bil (RM3.11tril) by 2030. 

At the time of writing in 2019, (Ratnam) records that consumer groups, aftermarket repair shops and privacy advocates argued that the data belongs to the car’s owners and the information should be subject to data privacy laws. In line with this, the European Union had already ruled that data generated by cars belonged to their owners and is subject to privacy rules under the EU’s General Data Protection Regulations or GDPR.  

In 2019 the Auto Alliance, a trade group representing the world’s largest car makers, was seeking for California to legislate that the companies be allowed to provide only summary information to consumers as opposed to the specific pieces of personal information a business has collected about them. This raises the question as to who owns the data for the user of an autonomous vehicle, particularly if the user does not own the vehicle.

The answer to this question will become more urgent with the deeper levels of data that automated vehicles will collect. Andrejevic (2020) argues that the creation and deployment of autonomous vehicles will transform cars into fully mediated devices, packed with sensors that collect and process a growing range of information. 

To find out more about who will own these autonomous users, we can turn to the Director of Product for Lyft, the ride-sharing platform who spells out how this company anticipates the future with autonomous vehicles (Swisher 2017). Taggart Matthieson detailed the collaborative model proposed in 2017 that Lyft is pursuing through the provision of an open platform with a number of partners who were seeking to produce level 5 vehicles (fully autonomous).  In his description he implies open data flows of user activity between the companies involved.

From the consumer experience point of view, the user of an autonomous vehicle will encounter two brands. The vehicle itself will be branded to the manufacturer, this will include the “brain” of the car, and the interface that the customer engages with and trip experience be curated by Lyft. The integration of the Lyft interface with the vehicle would include gathering intelligence from the car in order to facilitate the passenger validation process; for example gathering sensor data on the weight of the passenger on the car seat, that they have closed the door and connected the seatbelt. It will be an “integrated experience” between the sensors in the vehicle and the trip experience that Lyft provides. Summary taken from Taggart Matthieson, 2017, Recode Decode podcast.

More recently, Lyft and Aptiv launched a robotaxi pilot in January 2018 in Las Vegas. The program, which puts Aptiv vehicles on Lyft’s ride-hailing network, surpassed 100,000 rides this month. Human safety drivers are always behind the wheel and the vehicles do not drive autonomously in parking lots and hotel lobby areas (Korosec 2020).

In her article detailing Lyft’s current AV strategy, Korosec (2020) notes that in 2019, Lyft reported to the Californian DMV that they had 19 autonomous vehicles testing on public roads in California. Those 19 vehicles, which operated during the reporting period of December 2018 to November 2019, drove nearly 43,000 miles in autonomous mode. The report showed that Lyft is doing more than partnering with autonomous vehicle companies like Aptiv. 

These shifts towards MAAS, Pizzuto et al. (2019) argue, will change the rules of the game across the entire mobility space, as software and data become fundamental differentiators when building and operating cars. They observe that the mobility sector will become ground zero for a convergence of industries that include automotive, transportation, software, hardware, and data services. These trends point to the impending question that whilst private car ownership may decrease, who will own our data?

These two case studies demonstrate some of the perhaps hidden or less obvious social implications of autonomous vehicles that do need to be broadly considered.

Is the cybersecurity sector ready for our cars becoming all connected? 

Driverless cars are seen as one of the key disruptors in the next technology revolution. However, Kaur and Rampersad (2018) argue that the main barrier to adoption is the lack of public trust. Drawing on quantitative evidence, their study found that the ability of the driverless car to meet performance expectations and its reliability were important adoption determinants. Significant concerns included privacy (autonomy, location tracking and surveillance) and security (from hackers). The discussion of user privacy concerns was discussed in one of the initial case studies on personal data ownership. This section will focus on the issue of data security. 

In a report on the state of autonomous vehicles, West (2016) notes that autonomous cars depend on vehicle to vehicle (V2V) communications and vehicle to infrastructure (V2I) connections. Similarly, Vassallo and Manaugh (2018) observe that AVs are vulnerable to malicious attacks through many channels such as attackers physically tampering with a vehicle’s hardware, intercepting a vehicle’s communication signals, or hijacking a vehicle’s connection to a centralized server.

Case study: Hacks on AVs

Automated vehicles are equipped with multiple sensors (such as LiDAR, radar and camera) enabling local awareness of their surroundings. Researchers Jonathan Petit and Steven Shladover (2015) outline a number of security threats to connected cars. This includes hacking, jamming, data theft, ghost vehicles, or malicious actions such as using bright lights to blind cameras, radar interference, or sensor manipulation. Any one of these activities could disrupt communications and create false readings for artificial intelligence algorithms. Their study identifies GNSS (global navigation satellite systems) spoofing and injection of fake messages as the most dangerous attacks (i.e., most likely or most severe). Manipulating this type of information puts passengers are risk and potentially can lead to serious accidents. 

Cybersecurity experts already have demonstrated a capacity to remotely hack a Jeep Cherokee. In a report published in Wired magazine (Greenberg 2016), they tampered with the vehicle’s steering, brakes, radio, windshield wipers, and climate controls, and showed that this vehicle was easy to disrupt through its Uconnect software. This example shows that designers need to take vehicle security very seriously in order to avoid unnecessary risks. 

Vassallo and Manaugh (2018) argue that malicious software (malware) is also a hurdle to autonomous vehicle (AV) adoption and a serious threat to AV occupant safety. They observe that by removing the need to pay attention to the road, AVs will allow drivers to conduct Internet browsing activity that increases malware infection risk (like pirating media or viewing pornography) that falls outside the limited browsing options vehicle infotainment systems offer today. They also note that it is also possible that immobile information broadcast points could infect vehicles driving within signal range to these. Interestingly they suggest that it is possible that AVs could avoid malware-prone areas when planning a route or suggesting a destination of interest such as a gas station.

Within a MAAS system the vulnerabilities of centralized platforms coordinating personalized trip planning, pick up and payment (e-commerce) are also likely to be a target for client/server security threats as well as cyber identity thefts (Sharma, Singh & Sharma 2009). 

Case study: Hacks on infrastructure

Identity theft and credit card theft is common in the online environment, with marketplaces for stolen data established in many spaces, including cryptomarkets (Aldridge & Décary-Hétu 2014). E-commerce platforms regularly have their cyber security ‘tested’ by hackers seeking passwords, credit card information, personal identifiers for identity theft or other valuable data points. This has already been the case for Uber, who was reported to have concealed a massive global breach of the personal information of 57 million customers and drivers in October 2016 and paid the attackers $100,000 to delete the data and keep the breach quiet (Wong 2017). Similarly, in 2018, a ride-hail app called Careem based in Dubai reported hackers for stealing data belonging to 14 million riders and drivers, including customer names, email addresses, phone numbers and trip history, but no evidence of password or credit card information (Dickey 2018). E-commerce platforms holding credit card information and personal information are subject to identity and financial theft by malicious actors. 

Not only are the ride-share platforms already demonstrated to be vulnerable and targeted, so too is the public transport system. In 2016, a breach of the Western Australian transport systems initiated the organization shutting down real-time train tracking, amongst other systems, in response (Coyne 2016). Another instance, this time related to payment systems, the San Francisco public transit system was hacked with commuters unable to pay for their journeys and a ransom demaned. Monitors in station agent booths were seen with the message, “You Hacked. ALL data encrypted,” and the culprit allegedly demanded 100 Bitcoin (about $73,000). In response, the public transit service turned off the payment machines and opened the gates as a precaution. 

As Bergal (2018) reports in her article for a public service IT audience discussing instances of hacks of public transportation infrastructures, transportation systems are ripe targets for cyber criminals. From the smart cities perspective, journalist Ian Hardy (2016) draws on cybersecurity experts who say that it’s only a matter of time before hackers become interested in smart city transportation clouds and taking control of parking, traffic lights, signage, street lighting, automated bus stops and many other systems. He provides an existing example from Moscow, which has already experienced its first major transportation hack. Denis Legezo, a researcher with Kaspersky Lab, was able to manipulate traffic sensors and capture data simply by looking up a hardware user manual that was readily available online from the sensor manufacturer. With public infrastructure data likely hosted on an array of cloud servers, some within the jurisdiction and others not, this risk of malicious online activity and inability to apply local jurisdictional controls over the data (Ward & Sipior 2010) is only increased.

From these case studies we can see that there are several implications for an Internet of Vehicles that follow along the same concerns as the Internet of Things (of which they are a part) however hold their own unique implications that must be a part of the consideration of the implementation of these emerging technologies and mobility futures.

How do new generation technologies, like 5G, end up as conspiracies?

The new generation of connected self-driving cars is a convergence of a long list of technologies. This list includes AI, Big data, 5G, Cloud computing, IoT .. and some others. One technology in particular is interesting because it is stirring a lot of controversy (including being blamed for Covid 19 a few months ago), and that is 5G.


This observation that I will be called on to comment about for the upcoming industry focused Mobility panel, held for Hatch Quarter’s “Tech Tank” series, sent me on a rampage through the scholarly literature, from which I emerged with my sanity mostly intact but my hopes for a better future complicated. However, to bring in social change is to understand humanity in both its capacity for beauty and innovation, alongside how our darkness and anxieties that pass amongst us like wild fire shape that potential.

First, I begin with a sanity statement to pin us to the beauty and innovative capacities that we hold and how we develop technologies to realise this. Researchers, Ahmed et al (2020), state that compared to the current 4G networks, 5G wireless communications provide high data rates, have low latency, and increase base station capacity and perceived quality of service. They observe that the popularity of this technology arose because of the burst in smart electronic devices and wireless multimedia demand, which created a burden on existing networks. A key benefit of 5G, they suggest, is that some of the current issues with cellular networks such as poor data rates, capacity, quality of service, and latency will be solved.

Despite these promising innovations of this emerging telecommunications technology, Mansel & Plantin (2020) observe that fifth Generation mobile technology is at the centre of multiple controversies. They note that this controversy has gained momentum in early 2020 when conspiracy theories conflated the global spread of Covid-19 with China and 5G networks. From this point, we take a deep dive into the conspiracy theories and paranoia that are embedded in the cultural logics of our pandemic plagued times.

Ahmed et al (2020) observe that since the beginning of December 2019, the coronavirus disease (COVID-19) has spread rapidly around the world, which has led to increased discussions across online platforms. These conversations have also included various conspiracies shared by social media users. Amongst them, a popular theory has linked 5G to the spread of COVID-19, leading to misinformation and the burning of 5G towers in the United Kingdom.

In March 2020, Dr. Tedros Adhanom Ghebreyesus, the Director General of the World Health Organization (WHO), warned that the world is not just fighting an epidemic but an infodemic (Ghebreyesus, in United Nations, 2020), with real world implications and actions. These fears have fuelled specific conspiracies connecting 5G with COVID-19, animating protests and acts of vandalism that have occurred during the pandemic (Meese et al 2020).

When considering this tendency towards real world impacts of conspiracy theories, researchers Imhoff & Lamberty (2020) observe from existing research on the topic that a conspiracy-prone worldview does not only reduce trust in official versions and adherence to norms but is also linked to a stronger acceptance of violence. They suggest that conspiracy worldviews also make it more plausible to engage in illegal, nonnormative forms of action to reach one’s goals. Drawing from their own prior research, they argue that people high in conspiracy mentality see it as more defensible to use force and other illegal means to pursue one’s political goals.

However it is not just those active on social media the spread 5G related conspiracy theories. Prior to the pandemic relationship, conspiracies related to 5G pointed to Huawei, as a Chinese 5G equipment manufacturer, as being implicated in international trade wars, leading to suspicions of foreign interventions in domestic affairs in the UK (Mansel & Plantin 2020). This conspiracist mentality was pervasive in the British press coverage of the topic in 2017. Researchers Mansel and Plantin highlight the role of British news media coverage of 5G as containing forms of bias that make it difficult for citizens to assess the benefits and risks in an informed way. They found that whilst references to 5G in relation to urban life generally present an optimistic view, reporting on UK government policy in relation to 5G was frequently is linked to applications such as “Surveillance”, “Military Uses” or to the discussion of “Security”, “Data Control”, “Mobile Coverage”, “Personalised Services” and “Virtualisation”.

Building on this finding of the key role of news media in the spread of conspiracy theories, alongside the previous observation of the role of social media in dissemination networks, Australian researchers Bruns et al ( 2020) worked across media formats to trace the dissemination dynamics of rumours that the pandemic outbreak was somehow related to the rollout of 5G mobile telephony technology in Wuhan and around the world. Initially they observed that the rumour built upon a series of related narratives regarding the possible health and environmental impacts of 5G technology, that likely flourished in large part due to pre-existing networks and misinformation surrounding it.

In their analysis, Bruns and co-authors traced the rumour on Facebook from its obscure origins in pre-existing conspiracist groups through greater uptake in more diverse communities to substantial amplification by celebrities, sports stars and media outlets. This finding suggests that pointing the finger at social media, is a diversion from the role of more mainstream and traditional players in the amplification of information.

Past research shows that the increase in conspiracy theories during a pandemic is not a new phenomenon. DeCook (2020) argues that conspiracies surrounding the Covid-19 pandemic are also not unique or “new” in how they mobilize and rally against government institutions, science, and both “democrats” and “liberals.”

In a survey of the scholarly literature it appears that conspiracy thinking increases substantially during times of uncertainty, especially in times of crises and thrive in environments of low confidence and low trust. Conspiracy worldviews have also been connected to refusal to trust science, the biomedical model of disease, and legal means of political engagement and are reflective of the ever-evolving fears of technology and of bodily invasion.

Providing explanations is psychologically advantageous for several reasons, with one sticking out in the previous literature: granting an illusion of control. Considering this reasoning, it is not surprising that a lack of control has been identified as one of the key drivers of conspiracy beliefs. When people are not able to gain control in the real world, they compensate for this lack by perceiving patterns, even if they are an illusion.

The current coronavirus crisis is an almost ideal breeding ground for conspiracy thinking, as there is no easily comprehensible mechanistic explanation of the disease, it is an event of massive scale, it affects people’s life globally and leaves them with lots of uncertainty. Additionally, Mansel & Plantin (2020) identify that new infrastructure projects are always accompanied by conflicting visions or imaginaries and political and economic interests.

5G is also linked to long-standing concerns about potential health hazards of electromagnetic frequencies. Meese et al (2020) detail long-standing concerns around mobile technologies and infrastructures and how they translate to specific worries about 5G technology. They argue that a productive way to understand what is happening with 5G is to look beyond conspiracy theories to a larger set of concerns. DeCook 2020 observes that conspiracy theories often reveal something about the underlying anxieties and fears of the sociohistorical period they are created in as well as the personal anxieties of the adherents themselves

In the current situation, disinformation and conspiracy theories are being pushed from people all along the political spectrum, although DeCook 2020 argues this is more so in far-right information networks. She also observes that all over the world, and particularly in the United States, the anti-science, anti-intellectualist, and anti-establishment ideologies which fuel these movements have been long ongoing.

However, of our current moment, Evans (2020) observes a strange convergence of extremist politics between the new age movements and the far right. From this, perhaps we can say that in times of chaos, everybody regardless of political beliefs is grasping for a nugget of truth to make sense of the situation.

Aupers (2012) has argued that conspiracy theories and paranoia are embedded in the cultural logic of modernity – and that conspiracy culture is not a fringe phenomenon but rather has been absorbed into the mainstream. From this position, DeCook observes that from their pervasive influence, through wide ranging spread due to digital technologies and continual evolution, conspiracy theories as adaptive epistemologies have real and dangerous consequences on societies around the world.

Sources and further reading

Ahmed, W., Vidal-Alaball, J., Downing, J., & López Seguí, F. (2020). COVID-19 and the 5G Conspiracy Theory: Social Network Analysis of Twitter Data. J Med Internet Res, 22(5), e19458. doi:10.2196/19458
Aupers, S. (2012). ‘Trust no one’: Modernization, paranoia and conspiracy culture. European Journal of Communication, 27(1), 22-34. doi:10.1177/0267323111433566
Bruns, A., Harrington, S., & Hurcombe, E. (2020). ‘Corona? 5G? or both?’: the dynamics of COVID-19/5G conspiracy theories on Facebook. Media International Australia, 1329878X20946113. doi:10.1177/1329878X20946113
DeCook, J. R. (2020). The culture of conspiracy and the radical right imaginary. The American Ethnologist. Retrieved from https://americanethnologist.org/features/pandemic-diaries/making-sense-of-things/the-culture-of-conspiracy-and-the-radical-right-imaginary
Evans (2020) Nazi Hippies: When the New Age and Far Right Overlap: Both the New Age and the far right are drawn to conspiracy theories. https://gen.medium.com/nazi-hippies-when-the-new-age-and-far-right-overlap-d1a6ddcd7be4
Imhoff, R. and Lamberty, P. (2020). A bioweapon or a hoax? the link be- tween distinct conspiracy beliefs about the coronavirus disease (covid-19) out- break and pandemic behavior. Social Psychology and Personality Science, DOI: https://doi.org/10.1177/1948550620934692.
Mansell, R., & Plantin, J.-C. (2020). Urban futures with 5G: British press reporting. London School of Economics and Political Science, June. eprints.lse.ac.uk/105801/ ISBN 978-1-909890-65-7
Meese, J., Frith, J., & Wilken, R. (2020). COVID-19, 5G conspiracies and infrastructural futures. Media International Australia, 1329878X20952165. doi:10.1177/1329878X20952165