Going to Netsci in Boston in June? Interested in access to data from the National Internet Observatory, including our RFPs on AI chatbots, browsing behavior, search, and more to come over the next few months? Sign up for our workshop at Netsci:
national-internet-observatory.github.io/beyondapi_ne...
We will also have interactive activities and hands-on sessions with real network datasets that demonstrate NIO's capabilities for enabling novel cross-disciplinary and cross-platform research across web and social network environments!
We will discuss NIO's informed data donation process, participant demographics and behavioral traces, secure computing infrastructure and pathways for data access, and examples of analyses and innovative research with this new source of data for the network science community.
This is a vital source of data, and a crucial methodology for collecting data for academic research in the post-API age!
Track existing datasets open to request for research proposals at nationalinternetobservatory.org/researchers....
an alternative data-collection framework and infrastructure to help researchers study online behavior, with a particular focus on content viewing — the predominant (and a very understudied) form of online activity.
This satellite will introduce the interdisciplinary conference participants to the National Internet Observatory (NIO) (nationalinternetobservatory.org),
Very excited to announce a new satellite coming to NetSci 2026 @netsciconf.bsky.social, co-organized with Scott Cambo and @davidlazer.bsky.social!
More details in this thread and the website (national-internet-observatory.github.io/beyondapi_ne...)
Sign up here: forms.gle/sgjVPMSNWYeY...
AI data centers are straining already fragile power and water infrastructures in communities around the world, leading to blackouts and water shortages. “Data centers are where environmental and social issues meet,” says Rosi Leonard, an environmentalist with @foeireland.bsky.social.
LLMs are often used for text annotation, especially in social science. In some cases, this involves placing text items on a scale: eg, 1 for liberal and 9 for conservative
There are a few ways to accomplish this task. Which work best? Our new EMNLP paper has some answers🧵
arxiv.org/pdf/2507.00828
ICYMI, our DomainDemo dataset, which describes how different demographic groups share domains on Twitter, is now available to download!
📄 Data descriptor: doi.org/10.1038/s415...
📈 Interactive app to explore the data: domaindemo.info
💽 Dataset: doi.org/10.5281/zeno...
For more, check out the paper!
nature.com/articles/s41562-025-02223-4
arxiv.org/abs/2308.06459
For journalists and especially headline writers: even if a discrete piece of information is true, you've got to think carefully about whether the way you're presenting it is useful for promoting narratives that aren't.
Big picture: misleading claims are both *more prevalent* and *harder to moderate* than implied in current misinformation research. It's not as simple as fact-checking false claims or downranking/blocking unreliable domains. The extent to which information (mis)informs depends on how it is used!
If you want to advance misleading narratives — such as COVID-19 vaccine skepticism — supporting information from reliable sources is more useful than similar information from unreliable sources, if you have it.
This calls for a reconsideration of what misinformation is, how widespread it is, and the extent to which it can be moderated. Our core claim is that users are *using* information to promote their identities and advance their interests, not merely consuming information for its truth value.
We find that mainstream stories with high scores on this measure are significantly more likely to contain narratives present in misinformation content. This suggests that reliable information — which has a much wider audience — can be repurposed by users promoting potentially misleading narratives.
We do this by looking at co-sharing behavior on Twitter/X. We first identify users who frequently share information from unreliable sources, and then examine the information from reliable sources that those same users also share at disproportionate rates.
Our paper uses this dynamic — users strategically repurposing true information from reliable sources to advance misleading narratives — to move beyond conceptualizing misinformation as source reliability and measuring it by just counting sharing of / exposure to unreliable sources.
Take, for example, this headline from the Washington Post. The source is reliable and the information is, strictly speaking, true. But the people most excited to share this story wanted to advance a misleading claim: that the COVID-19 vaccine was ineffective at best.
But users who want to advance misleading claims likely *prefer* to use reliable sources when they can. They know others see reliable sources as more credible!
When thinking about online misinformation, we'd really like to identify/measure misleading claims; unreliable sources are only a convenient proxy.
In our new paper (w/ @jongreen.bsky.social , @davidlazer.bsky.social, & Philip Resnik), now up in Nature Human Behaviour (nature.com/articles/s41562-025-02223-4), we argue that this tension really speaks to a broader misconceptualization of what misinformation is and how it works.
There's a lot of concern out there about online misinformation, but when we try and measure it by identifying sharing of/traffic to unreliable sources, it looks like a tiny share of users' information diets. What gives?
Goel et al. examine why some factually correct news articles are often shared alongside fake news claims on social media. @pranavgoel.bsky.social @jongreen.bsky.social @davidlazer.bsky.social
www.nature.com/articles/s41...
If you are at #WebSci2025, join our "Beyond APIs: Collecting Web Data for Research using the National Internet Observatory" - a tutorial that addresses the critical challenges of web data collection in the post-API era.
national-internet-observatory.github.io/beyondapi_websci25/
If you study networks, or have been stuck listening to people who study networks for long enough (sorry to my loved ones), you may have heard that open triads – V shapes – in social networks tend to turn into closed triangles. But why does this happen? In part, because people repost each other.
Congratulations!