The double life of being chronically ill at work, slow librarianship, and checking in as an expression of care / Meredith Farkas
Developing a long-term illness, whether chronic or acute, is like being dropped into a country completely unfamiliar to you. You don’t know the language, the customs, the cuisine, the people. You feel alone, isolated, and totally out of your depth. Eventually, you start to learn the language, the customs. You find community, fellow travelers, people who can help you understand your new life better. It doesn’t stop being hard, but the learning curve becomes less steep and the isolation less intense.
However, unlike when you’re immersed in a new country and culture, you’re falling into this new place and experiencing that painfully slow acculturation while you’re trying to still live your regular life in parallel. You’re expected to be a good parent, partner, family member, friend, employee, housekeeper, bill-payer, etc. But you’re living in two different realities now and because of that, it’s easy to feel alienated from your regular life, especially if you don’t feel like you can bring that other part of yourself to your interactions at work, at home, or out with friends. The cognitive dissonance can be jarring.
It’s hard enough to live that double life, but adding in the vagaries of seeking out a diagnosis and often not being believed, not to mention coping with the symptoms themselves, can make life feel completely untenable. Before my autoimmune diagnosis, I spent more than a year seeing medical professionals who didn’t believe there was anything wrong with me other than the normal discomforts of aging. I kept asking doctors if they thought my symptoms could be autoimmune and was told “no” over and over again, though that felt wrong to me. One PA suggested that some of my symptoms might stem from anxiety since I have a history of anxiety (like I wouldn’t know at this point what anxiety feels like). Within five minutes of talking with the first rheumatologist I saw, after waiting five months for the appointment, he said to me “this doesn’t sound rheumatological at all.” Luckily he still did all the standard testing which showed that he was very wrong. But not being believed by so many doctors for so long stays with you. It leaves a scar. Every time I see a doctor now, I feel like I’m going to court and I’m ready to be cross-examined, to be picked apart. I’m a bundle of nerves.
And my experience is painfully common, especially for women, as poet Meghan O’Rourke writes in her amazing book about chronic illness, The Invisible Kingdom (a meditation on and journalistic exploration of chronic illness and how it is positioned in our social fabric):
And so it is a truth universally acknowledged among the chronically ill that a young woman in possession of vague symptoms like fatigue and pain will be in search of a doctor who believes she is actually sick. More than 45 percent of autoimmune disease patients, a survey by the Autoimmune Association found, “have been labeled hypochondriacs in the earliest stages of their illness.” Of the nearly one hundred women I interviewed, all of whom were eventually diagnosed with an autoimmune disease or another concrete illness, more than 90 percent had been encouraged to seek treatment for anxiety or depression by doctors who told them nothing physical was wrong with them. (p. 103)
Once I got on meds for my condition that finally started working (most first-line meds for autoimmune conditions take three months on average to produce any effects on the immune system), I thought I was past the worst of it. Other than occasional much smaller flares, I was essentially in remission. I learned my limits. I protected my spoons with my life. I kept my stress low. I felt like I had it figured out. I felt good even. And then I got sicker with new symptoms that have stolen so much from me, including my sleep. The past 10 months have honestly been a nightmare with a carousel of doctors who all have completely different theories of what is going on and with a condition that is constantly evolving so what they see in the moment isn’t the full picture. And each doctor I’ve seen hyperfocuses on something different and ignores every other aspect of my case. It’s all making me feel like I’m going crazy.
Each wrong diagnosis brought me to another country, another reality, another identity, that I lived in for a short while. And in each of these countries, I spent countless hours learning, learning, learning all I could, going down research and subreddit rabbit holes, and spending way too much money on products that did nothing because none of those diagnoses were correct. The dermatologist I see who specializes in autoimmune conditions at a research university seems to have given up even trying to diagnose me and she’s basically my last best hope in the state. She wouldn’t even give me differential diagnoses last time beyond “it’s clearly autoimmune given your systemic symptoms.” She’s the one who has put me on a serious immunosuppressant with debilitating side effects, which I guess at least shows she’s taking it seriously, but if she doesn’t know what she’s treating me for, how does she know that this drug that is making me feel terrible is even going to help? Atul Gawande said of doctors that “nothing is more threatening to who you think you are than a patient with a problem you cannot solve” (quoted in The Invisible Kingdom, page 209) and I feel that in how I’m treated at every appointment I go to. With the exception of the charlatan immunologist who was desperate to diagnose me with MCAS though there was no evidence in support of it, not one doctor has seemed at all interested in figuring out what this is – it’s felt more like a game of “not it.”
Of a similar liminal moment in her own illness, O’Rourke wrote, “in my illness I was moored in an unreachable northern realm, exiled to an invisible kingdom, and it made me angry. I wanted to rejoin the throngs. In dark moments I continued to wonder if the wrongness was me” (99). And of course people will feel like that wrongness is them when we live in a culture that views chronic illness as some sort of weakness or something we caused through through our own bad habits. Like O’Rourke, I feel both exiled from my world and forced to be in it at the same time, which is a unique form of torture. Going to work, pretending things are ok, doing my job, meeting deadlines, helping students, smiling, all the while my body is attacking itself, I’m barely sleeping, I’m spontaneously bleeding from my skin and under my skin, and I’m so itchy I sometimes have to wear gloves to bed or I’ll scratch myself raw in my sleep. You feel like you’re play-acting being yourself, being a person in the world, because you’re not really there anymore. And when you’re suffering, and don’t know what your illness is, and you feel abandoned, it’s easy to go down rabbit holes of self-loathing along with those rabbit holes of fruitless research that make you feel like an unhinged obsessive with a murder board and yarn. As Meghan O’Rourke wrote, “your sense of story is disrupted” (p. 259) and you feel like a stranger to yourself.
I started writing about slow librarianship long before I got really sick, but even then, I knew the importance of fostering a work culture where you can be a whole person. I knew how it felt to have a child and feel like you couldn’t prioritize family obligations over work ever (though working during family time? Totally ok, right?). In a workplace that encourages people to be whole people, workers feel like they can prioritize the things in their lives outside of work that are important – their caregiving responsibilities, their health, the people they love, etc. They feel like they can talk about these things – that they don’t make them liabilities. They can be vulnerable and real. And feeling like you can be vulnerable and real about who you are and how you’re doing means that you can also be vulnerable and real in your work, which makes us better employees who are energized to try new things.
I think a lot of people in positions of power might even want a culture like this, but very few actively create it. They might think that saying “take what time you need” when someone is facing illness is enough. But I think two pieces are missing from this. First, managers need to not only say “take what time you need” but work with their direct reports to address the work that would otherwise pile up. If you say “take all the time you need” but all the work with its stressors and deadlines is still there, you’re not really giving people space. Can you take good care of yourself while you watch the work pile up and up and up? How many of us have come back to work while still not fully recovered from an illness because of the work that was piling up or a class they needed to teach?
Also, managers need to model vulnerability, transparency, and being whole people themselves. If they put up a false front of strength, if they’re not willing to be vulnerable and human and real themselves, if they do not model transparency, there’s no way that others will feel safe doing so. I was lucky to have a boss in my first academic library job who was deeply human in her interactions with her employees, so I got to see what that looked like. And it was her humanity that engendered fierce loyalty in her employees – we all thought the world of her. Even when she made decisions that people didn’t like (which was rare as she really did take our insights to heart), she explained her thinking in a transparent way. Given my later experiences, her way of being feels vanishingly rare. I think a lot of managers feel like they need to project strength, not explain their decisions, not let their direct reports really know them as people with full lives, but I don’t think that’s true. A lot of managers operate out of a place of fear or insecurity, but my first academic library director was confident enough to be her full self, flaws and all.
In a culture where we don’t feel like we can bring ourselves fully to work, I don’t feel like I can talk about my illness. I feel selfish and weak for even considering it. Like, we all have shit going on, right? The world is pretty awful right now. People’s lives are complicated and messy and there’s probably a lot of suffering I know nothing about happening all around me. If they don’t talk about it, who am I to talk about it? I’m not special. While I’ve mentioned being sick at work in the context of being immunocompromised and needing to protect myself and not participate in large, crowded events, even that has felt really uncomfortable. Everyone should feel like they are important enough to their places of work and valued enough to bring up these things without feeling embarrassed or like they’re asking for “special treatment.” I read recently (can’t remember the source) that close to 60% of people with chronic illnesses have not told people at work about it. Imagine hiding something that is such a pivotal and ineluctable piece of many people’s identities and think about the double life that forces them to live.
At work I feel a lot of shame about being sick and I work more than I should given how I’ve been feeling (this is common). I feel like I need to mask how I’m really doing, that people don’t want to hear it. And it’s true that most don’t. There are three people at work I can talk to about my illness, but others seem so incredibly uncomfortable when I mention it, so I’ve learned to just pretend it’s all ok or say nothing. I know that some of my reticence and shame comes from my own internalized ableism as it’s the water we all swim in, but when I worked in that library where the culture encouraged vulnerability, humanity, and care, I remember how different it felt. How much less distance there was between the person I really was and the person I was at work.
In a meeting last year, our Dean was talking about making the next all-library meeting in-person only. Previously, they had always offered them hybrid, but she didn’t like that most people were choosing not to come in-person. And I totally get it, even though it sucks to always be the outlier. She wants it to be a team-building experience and that’s really hard to do when most people are participating from home with their cameras off. At that meeting, for the first time, I disclosed my illness in front of a bunch of people and talked about how important it is to always offer an online option for folks who are medically vulnerable or at least find ways to make indoor spaces safer for those who can’t afford to get sick. My boss then asked to meet with me to talk about how to make our spaces safer. I talked about airflow, encouraging and providing masks, and, during temperate months, having the meetings at places where windows and doors can remain open or even holding them outside (we’d had one meeting at a park a few years ago which had been the best one ever from both a health and team-building perspective). I suggested that we make the Winter meeting fully virtual since it’s the height of flu season and you can usually get people participating more in a fully virtual meeting than a hybrid one, where the online people feel like weird lurkers. I was thanked for my feedback and didn’t hear anything after that. This September, the all-library meeting was held in our campus library where the windows don’t open (albeit with a couple of HEPA filters scattered around, but I know we do have large college spaces where doors and windows can be opened because I went to an all-day union meeting in one where they did just that and we’d used that space for library meetings in the past) and our February meeting is going to be held in-person during the worst flu season in decades. Obviously, none of this was personal or intended to cause harm, but, at the same time, how should I feel under the circumstances? Clearly speaking out in that meeting, something that I had to really steel myself to do, had been pointless. Why would I ever bring it up or ask for anything again?
When we come back from winter break, people inevitably ask each other how it was, but do they really want to know? I think they want to hear “good,” “fun,” “restful,” etc. How do you talk about a “vacation” in which you spent most of it doubled over in pain after eating anything thanks to the toxic meds you have to take and your father-in-law was in the hospital dying the entire time? I wish that I felt I could share what an absolute shitshow it’s been, to feel like I could be a full person at work, but when you know no one really wants to hear it, when it just makes people uncomfortable, it becomes so much easier to smile and say “it was good!” and move on. Who wants to be a buzzkill?
Slow librarianship puts worker well-being over productivity and deadlines, allows workers to be whole people at work, and supports a culture of care. While a radical idea, this even makes good business sense because depleted and burned out workers have been shown to be a major drain on the organization and negatively impact the culture. If you’re a manager and you’re not actively fostering a culture where people can bring their whole selves to work, then you are fostering a culture where people do not feel safe being vulnerable and having needs. Workers, if you know a colleague is struggling with something (an illness, losing a loved one, a difficult caregiving situation, etc.) and you don’t check in to see how things have been going for them, you’re sending the message that you don’t want to hear about these things, that they make you uncomfortable, that they’re not appropriate to take to work. I think we’ve all probably been guilty of this at some point in our lives and maybe we even thought that not asking was the right thing to do. I can imagine some people think that asking is invasive or reminds the person that they are sick, but it is an expression of care. As Philip Hoover writes in his excellent Sick Times article entitled “You know someone with Long COVID. They need you to ask about it genuinely”
Approach us with empathy and curiosity. Ask questions that show a sincere desire to understand. How are you feeling this week? works because it acknowledges the chronic, fluctuating nature of my health. Or a version of Nunez’s question, and one I’ve longed to be asked since I’ve been ill: What has this been like for you?
If our response is tough to hear, try not to smother it in optimism. And tread lightly. While some long-haulers may appear okay in public, much of our suffering occurs in private, shade-drawn rooms, across lonely afternoons, stuck in bed. When in doubt, remember that the act of asking never hurts — but never being asked certainly does.
While I don’t have Long COVID, this piece so perfectly encapsulates how I feel as someone with a mostly invisible chronic illness who would just love to be asked “how are you feeling this week?” instead of feeling like I have to pretend I’m okay. I told my manager at the start of Fall term that my autoimmune disease had become significantly worse and that I didn’t know what my capacity might look like going forward. She expressed sympathy, told me to take what time I needed, and never checked in with me after that. That September, I was taking over a very important committee chair role that was made enormously more time-consuming and onerous by the departure of the three colleagues most involved in supporting this work (two of whom had more than 20 years of institutional knowledge locked up in their heads). I didn’t feel like I had the leeway or support to let things drop as more and more kept piling up on my plate related to my chair role and it became clear that I was expected to do a lot of onboarding for the new people in this role even though I was new to my role and was no one’s manager. While I know my boss is extraordinarily busy, the message that not checking in with how I was doing sent me was very different from what I assume she’d wanted to convey. Checking in with a colleague or direct report seems like such a small thing, and it is in terms of the effort it requires, but the impact it can have in making someone feel cared for and less like they have to live a painful double life is enormous.
Weekly Bookmarks / Ed Summers
These are some things I’ve wandered across on the web this week.
🔖 Lou Reed/Laurie Anderson/John Zorn January 10th, 2008 Stone, NYC pt 3
🔖 John Zorn’s Naked City - The Marquee Club, New York City, NY, 1992-04-09
John Zorn’s Naked City April 9, 1992 The Marquee Club, New York, NY pro-shot (a neckey - voltarized upgrade)
Personnel: John Zorn: Alto Sax Bill Frisell: Guitar Wayne Horvitz: keyboards Fred Frith : Bass Joey Baron: drums Yamatsuka Eye: vocals🔖 Pho & Banh Mi Saigonese
🔖 About Standard Ebooks
🔖 The Rime of the Ancient Maintainer
Every culture produces heroes that reflect its deepest anxieties. The Greeks, terrified of both mortality and immortality, gave us Achilles. The Victorians, haunted by social mobility, gave us the self-made industrialist. And Silicon Valley, drunk on exponential curves and both terrified and entranced by endless funding rounds, has given us the Hero Developer: a figure who ships features at midnight, who “moves fast and breaks things,” who transforms whiteboard scribbles into billion-dollar unicorns through sheer caffeinated will.
We celebrate this person constantly. They’re on the front page of TechCrunch et al. They keynote conferences. Their GitHub contributions get screenshotted and shared like saintly relics.
Meanwhile, an unsung developer is updating dependencies, patching security vulnerabilities, and refactoring code that the Hero Developer wrote three years ago before moving on to their next “zero to one” opportunity.
They will never be profiled in Wired.
But they’re doing something far more important than innovation.
They’re preventing collapse.🔖 One Number I Trust: Plain-Text Accounting for a Multi-Currency Household
🔖 Glamorous Christmas: Bringing Charm to Ruby
Today, Ruby 4.0 was released. What an exciting milestone for the language!
This release brings some amazing new features like the experimental Ruby::Box isolation mechanism, the new ZJIT compiler, significant performance improvements for class instantiation, and promotions of Set and Pathname to core classes. It’s incredible to see how Ruby continues to thrive and be pushed forward 30 years after its first release.
To celebrate this release, I’m happy to announce that I’ve been working on porting the Charmbracelet Go terminal libraries to Ruby, and today I’m releasing a first version of them. What better way to make this Ruby 4.0 release a little more glamorous and charming?🔖 29 Finding a broken trace on my old Mac with the help of its ROM diagnostics
🔖 Public Domain Day 2026
🔖 Pulled 60 Minutes segment on CECOT
🔖 Two Years After Cormac McCarthy’s Death, Rare Access to His Personal Library Reveals the Man Behind the Myth
Cormac McCarthy, one of the greatest novelists America has ever produced and one of the most private, had been dead for 13 months when I arrived at his final residence outside Santa Fe, New Mexico. It was a stately old adobe house, two stories high with beam-ends jutting out of the exterior walls, set back from a country road in a valley below the mountains. First built in 1892, the house was expanded and modernized in the 1970s and extensively modified by McCarthy himself, who, it turns out, was a self-taught architect as well as a master of literary fiction.
I was invited to the house by two McCarthy scholars who were embroiled in a herculean endeavor. Working unpaid, with help from other volunteer scholars and occasional graduate students, they had taken it upon themselves to physically examine and digitally catalog every single book in McCarthy’s enormous and chaotically disorganized personal library. They were guessing it contained upwards of 20,000 volumes. By comparison, Ernest Hemingway, considered a voracious book collector, left behind a personal library of 9,000.🔖 How uv got so fast
🔖 Post-Platform Digital Publishing Toolkit
🔖 Giambattista Vico
🔖 Robots Can Be Hacked in Minutes, Chinese Cybersecurity Experts Warn
Commercial robots have widespread and exploitable vulnerabilities that can allow hackers to take over within hours or even minutes, according to Chinese cybersecurity experts.
Security in the robotics industry is “riddled with holes,” said Xiao Xuangan, who works at Darknavy, an independent cybersecurity research and services firm based in Singapore and Shanghai. Xiao noted that when testing low-level security issues in quadruped robots, his team gained control of one of Deep Robotics’ Lite-series products in just an hour.2025-12-31: Review of WS-DL's 2024 / Web Science and Digital Libraries (WS-DL) Group at Old Dominion University
Better late than never, right? This review of WS-DL's 2024 is incredibly late, but some family concerns delayed my writing and then I never quite got back on track. We had quite a productive 2024, graduating a record three MS students and four PhD students.
Students and Faculty
We did not add or lose any faculty this year, but Dr. Jayarathna received tenure early! Congratulations to Sampath!
We graduated four PhD students:
Bathsheba Farrow defended her dissertation on 2024-06-28. Bathsheba already had a position with the Naval Surface Warfare Center, and will continue with them after her graduation.
Yasith Jayawardana defended his dissertation on 2024-07-09. Yasith took a research scientist position at Georgia Tech.
Gavindya Jayawardena defended her dissertation on 2024-07-10. Gavindya took a post-doc position at the Information eXperience lab at the UT Austin School of Information.
Muntabir Hasan Choudhury defended his dissertation on 2024-11-06. Muntabir took a research fellow position at the Food and Drug Administration (FDA).
Congratulations to Dr. Bathsheba Farrrow @sheissheba for successfully defending her dissertation "A Microservices Approach to EEG Research in the Public Cloud" She is the very first student from the @NirdsLab research group @WebSciDL @oducs to defend the Ph.D. pic.twitter.com/K5wDRa7qu4
— Sampath Jayarathna (@OpenMaze) June 28, 2024
Congratulations to Dr. Yasith Jayawardana @yasithdev for successfully defending his doctoral dissertation "A Realtime Biosignal Processing Framework for Lab Scale Experimentation." Second Ph.D to come out of @NirdsLab research team @WebSciDL @oducs @ODUSCI. pic.twitter.com/zMYQ9Tq6KP
— Sampath Jayarathna (@OpenMaze) July 9, 2024
Congratulations to Dr. Gavindya Jayawardena @Gavindya2 for successfully defending her dissertation “RAEMAP: Real-time Advanced Eye Movement Analysis Pipeline”. 🥳@NirdsLab @WebSciDL @oducs @ODUSCI @OpenMaze pic.twitter.com/35NeNDHT1L
— Yasasi (@Yasasi_Abey) July 10, 2024
— Jian Wu (@fanchyna) November 6, 2024
And we also graduated three MS students, two of whom stayed and entered the PhD program:
Caleb Bradford finished his MS in 2024.
Dominik Soós defended his MS thesis and joined the PhD program.
Lesley Frew defended her MS thesis and joined the PhD program.
My student Dominik Soós successfully defended his Master's thesis today. His thesis title is "Who Wrote the Scientific News? Improving the Discernibility of LLMs to Human-Written Scientific News". Thanks @vikas_daveb @Meng_CS. Dominik will be a PhD student in Fall 2024 @WebSciDL pic.twitter.com/7yVXdoOT0w
— Jian Wu (@fanchyna) July 18, 2024
Congrats to @lesley_elis for a successful MS thesis defense! In "Surfacing Text Changes in Archived Webpages," she showcased a new interface to search web archive collections for added or deleted terms. 🧵@WebSciDL @phonedude_mln @OpenMaze @oducs @odusci pic.twitter.com/onrxQTGuto
— Michele Weigle (@weiglemc) July 29, 2024
We added three new PhD students, two of whom joined from our MS program:
Lawrence Obiuwevwi (@LObiuwevwi) joined WS-DL in Spring 2024 and is working with Dr. Jayarathna.
Dominik Soós (@DomSoos) continued with WS-DL and Dr. Wu in the PhD program in Fall 2024.
Lesley Frew (@lesleyelisabeth.bsky.social) continued with WS-DL and Dr. Weigle in the PhD program in Fall 2024.
We also had six PhD students advance their status:
Yasasi Abeysinghe advanced to candidacy
Nithiya Venkatraman advanced to candidacy
Lamia Salsabil advanced to candidacy
Tarannum Zaki advanced to candidacy
Yash Prakash advanced to candidacy, and also defended his prospectus
Xin Wei defended her prospectus
Publications and Presentations
To generate our annual publication list, we use our tool "Scholar Groups", which scrapes Google Scholar profiles and merges and deduplicates publications. As such, our list is limited by the accuracy of Google Scholar, which is pretty good but not perfect. It looks like for 2024 we published about 28 refereed conference papers, 12 journal articles, and one patent (congratulations, Dr. Ashok!).
Conferences have mostly returned to f2f, but frequently there are still virtual/remote options. Below is a partial list of trip reports for the events where we presented our work:
Digital Humanities and the Study of Mediterranean Mobilities
Modeling, Simulation and Visualization Student Capstone Conference (MSVSCC 2024)
3rd International Conference on Science of Science and Innovation (ICSSI 2024)
ACM Conference for Reproducibility and Replicability (ACM REP 2025)
ACM SIGIR Conference on Research and Development in Information Retrieval (SIGIR 2024)
ACM Symposium on Eye Tracking Research & Applications (ETRA 2024)
ACM International Conference on Information and Knowledge Management (CIKM 2024)
International Conference on Theory and Practice of Digital Libraries (TPDL 2024)
IEEE Visualization and Visual Analytics Conference (VIS 2024)
International Symposium on Electronic Theses and Dissertations (ETD 2024)
9th Computational Archival Science Workshop (CAS 2024)
🏆Received "Best Presentation Award" back to back for another year. This time for the AI & Autonomous Systems track! 🥳 pic.twitter.com/e1DcbUYngV
— Tarannum Zaki (@tarannum_zaki) April 11, 2024
Tonight at the Poster Session during the #CIKM2024 Welcome Reception, @lesley_elis presented her short paper titled “Retrogressive Document Manipulation of US Federal Environmental Websites”. @weiglemc @phonedude_mln @WebSciDL @cikm2024 pic.twitter.com/MalRAFLGdB
— Himarsha R. Jayanetti (@HimarshaJ) October 23, 2024
I just presented our full paper, "Towards Enhancing Low Vision Usability of Data Charts on Smartphones," at #IEEEVIS 🎉. Grateful for the opportunity to share our work on improving data accessibility for low vision users. #DataViz #Accessibility #LowVision @WebSciDL @ieeevis pic.twitter.com/xZblvfkMhw
— Yash Prakash (@LunaticBugbear) October 17, 2024
Research Presentations and Outreach
In addition to the conferences and workshops listed above where we presented papers, we also had an array of presentations and outreach.
We hosted the third and final year of our NSF Disinformation Detection and Analytics Research Experiences for Undergraduates (REU) Program. The cohort of the third year was especially successful, as reflected in the mid-point and final presentations. As you can see from the list above, many of the REU projects resulted in publications.
We also had our 5th annual "Trick or Research" event. Dr. Jayarathna initiated the event five years ago and the entire department has embraced it.
We also had our fourth annual Research Expo, where we have five PhD students give overviews of their work.
In addition, we had the following events and outreach:
Mohan Krishna presented at the Global Summit on Artificial Intelligence
Dr. Nelson and Dr. Ross Gore presented "Establishing and Growing a Research Identity" at OERI.
Dr. Wu was quoted in a JLab article about the Boost Platform Workshop
Dr. Ashok presented his group's research for Global Accessibility Awareness Day (GAAD)
Dr. Jayarathna presented his group's research at the Perry Honors College
We were happy to hose Prof. Michael Herzog of Magdeburg-Stendal University.
Dr. Jayarathna participated in the Fall 2024 Open House, School of Data Science
Dr. Jayarathna gave an invited talk at Ocean Lakes High School (2024-10-16)
Good to see our friend Professor Herzog @maherzog during his annual visit @odu. We had a nice lunch at the Webb Center, got a chance to exchange some ideas, and met our @NirdsLab @WebSciDL students during the weekly lab meeting. pic.twitter.com/Ry0rr99KzJ
— Sampath Jayarathna (@OpenMaze) March 22, 2024
Happenning now: Drs. Michael Nelson @phonedude_mln and Ross Gore @rossgore are presenting Establishing and Growing a Research Identity at OERI.
— Faryaneh Poursardar (@Faryane) May 3, 2024
cc:/ @WebSciDL @vmasc_odu @oeriatodu pic.twitter.com/CzM7Whbv6f
Software, Data Sets, Services
Our scholarly contributions are not limited to conventional publications or presentations: we also advance the state of the art through releasing software, data sets, and proof-of-concept services & demos. Some of the software, data sets, and services that we either initially released or made significant updates to in 2024 include:
Several repositories supporting publications from the Accessible Computing Lab (Dr. Ashok), including: InstaFetch, Discussion_Dataset, Low Vision Graphical Perception of Bar Charts, GraphLite, and ChartSync.
Multi-Domain Scientific Claim Verification Evaluation Corpus (MSVEC)
Awards and Recognition
Many members of our group received awards and other forms of recognition, a sampling of which includes:
Yasasi Abeysinghe received the Levenstein Scholarship, Outstanding Researcher Award, and COVES Fellowship.
Tarannum Zaki won the "Best Presentation Award" at MSVSCC2024.
Dominik Soós won the department's best research assistant and Rochana Obadage won the department's best teaching assistant
David Calano received the SMART Scholarship from the DoD, which includes summer internships
Kenny Ajayi had a summer internship at Amazon
Mohan Kriksha had a summer internship at PRA Group
Dr. Nelson received the Eminent Scholar designation.
Dr. Wu won the Provost's Outstanding Undergraduate Research Mentor
Dr. Jayarathna was Senior Faculty Fellow, Office of Naval Research Summer Faculty Program, at Naval Surface Warfare Center Dahlgren Division, Dam Neck Activity during the summer
Dr. Jayarathna received won “Outstanding Undergraduate Research Mentor" and "John R. Broderick Diversity Champion Award" (in addition to graduating three PhD students this year!)
I am honored to be awarded the 2024 Provost's Outstanding Undergraduate Research Mentor of the Old Dominion University. Thanks President Hemphill and Provost Agho. @WebSciDL @oducs @ODUSCI @ODU pic.twitter.com/sAl2xiM2Dh
— Jian Wu (@fanchyna) April 24, 2024
Thank you for joining with me @hanks_norfolk to celebrate my tenure and promotion. Thank you for the camaraderie! @WebSciDL @storymodelers @NirdsLab @oducs pic.twitter.com/2iSJxWJh1J
— Sampath Jayarathna (@OpenMaze) April 30, 2024
Funding
In 2024, we received > $7.8M in six new externally funded grants:
Dr. Wu (PI) received $564k from IMLS for "Preserving Open Access Datasets and Software for Sustained Computational Reproducibility" (WSDL alumnus Dr. Sawood Alam is Co-PI).
Dr. Poursardar (PI) received $150k from NSF for "Advancing Health Equity: Integrating LLM Technology into Homeless Telehealth Services for Chronic Disease Education"
Drs. Nelson and Weigle (Co-PIs) received $399k from IMLS for "Preserving Personalized Advertisements for More Accurate Web Archives" (WSDL alumnus Dr. Mat Kelly is PI)
Drs. Ashok and Poursardar (PI and Co-PI, respectively) received $50k from CCI for "Tackling Dark Pattern-Induced Online Deception of People with Visual Disabilities"
Dr. Ashok (PI), with Drs. Jayarathna and Wu (Co-PIs), received $117k from IMLS for "Enhancing Accessibility of Electronic Theses and Dissertations"
Dr. Jayarathna is also senior personnel (Karen Sanzo (PI)) for a grant of $6.59M from the Virginia Department of Education for three new lab schools in Newport News and Chesapeake
Summary
2024 was a strong year for us: four PhD and three MS students graduated, three new PhD students, and six PhD students advancing their status. One of our faculty members, Dr. Sampath Jayarathna, received tenure. We continued to publish in prestigious venues, with about 40 refereed publications. We helped generate just over $7M in new external funds, from six different grants. WS-DL continues to grow and thrive, and I am proud of all the members and alumni and their progress in 2024.
If you would like to join WS-DL, please get in touch. To get a feel for our recent activities, please review our previous WS-DL annual summaries (2023, 2022, 2021, 2020, 2019, 2018, 2017, 2016, 2015, 2014, and 2013) and be sure to follow us at @WebSciDL to keep up to date with publications, software releases, trip reports and other activities. We especially would like to thank all those who have publicly complimented some aspect of the Web Science and Digital Libraries Research Group, our members, and the impact of our research.
–Michael
Congratulations to All three Dr. J's from the @NirdsLab Dr. Bathsheba Jackson @sheissheba , Dr. Gavindya Jayawardena @Gavindya2 , and Dr. Yasith Jayawardana. @yasithdev. @WebSciDL PhD Crush board updated, just waiting in the border of alumni cloud until the paperwork is done! pic.twitter.com/dwEMMzOpy5
— Sampath Jayarathna (@OpenMaze) July 10, 2024
After the defense, @lesley_elis continued @WebSciDL's tradition of students sharing a hometown meal with us. Lesley's Philly roots showed up in the cheesesteaks (made authentic with Cheez Whiz), @Tastykake, birch beer soda, and @GrittyNHL decorations. Congrats again, Lesley! 🎉🎓 pic.twitter.com/WP8Ny1d2BY
— Michele Weigle (@weiglemc) July 29, 2024
Public Domain Day 2026: Celebrating human creativity and sharing / John Mark Ockerbloom
I’m glad we’ve reached a new Public Domain Day, and that the works I’ve been featuring in my #PublicDomainDayCountdown, and many more, are now free to copy and reuse. I’ve been posting about works joining the public domain in the United States, which include sound recordings published in 1925, and other works published in 1930 that had maintained their copyrights. (Numerous works from 1930, and later, that had to renew their copyrights, and did not, were already in the public domain, though many of the best-known works did renew copyrights as required.) This is the eighth straight year Americans have seen a year’s worth of works join the public domain, after a 20-year freeze following a 1998 copyright extension.
I intend my countdown not just to celebrate the works joining the public domain, but also to celebrate what people have done with those works. In some posts, I note later creations based on those works. In nearly all my posts, I link to things that people have written about those works. Like the works themselves, those responses may have flaws or quirks, but I value them as human reactions to human creations. Whether they’re reviews, personal blog posts, professionally written essays, scholarly analyses, or Wikipedia articles, they’re created by people who encountered an interesting work and cared about it enough to craft a response to it and share it with the world. Those shared responses in turn pique my interest in the writers and the works.
It wasn’t always easy for me to find such responses online. Sometimes I’d go searching for responses to a promising-sounding work, and only find sales listings on e-commerce sites, social media posts not easily linkable or displayable without logging into a commercial platform, paywalled articles that many of my readers can’t view, or generic-sounding pages that read like they were generated by a large language model or a content farm, but not by anyone who I could clearly tell cared about or even read the work in question. Some works I initially hoped to feature got left off my countdown, replaced by other works where I could more readily link to an interesting response.
The people publishing the responses I link to are often swimming against a strong current online. Many online writing systems– including the one I’ve been writing these posts on— are now urging their users to “improve” their posts by letting “AI” write them. Some writers may be tempted to allow it, when facing an impending deadline or writer’s block or anxiety, even when the costs can include muffling one’s own voice, signing onto falsehoods confidently stated by a stochastic text generator, or abusively exploiting existing content and services. Other writers may feel pushed to put their work behind paywalls or other access controls that makes them less likely to be plagiarized or aggressively crawled by those same “AI” systems. And most writers, myself included, find it easy to dash off a quick short take on a social media platform, be quickly gratified by some “like”s, and then have it forgotten. It’s harder to take the time to craft something longer or more thought-out that will be readable for years, and that might take much longer for us to hear appreciated. The easy alternatives can discourage people from devoting their time to better, more lasting creations.
As I’ve noted before, both copyright and the public domain serve important purposes in encouraging the creation, dissemination, sharing, and reuse of literature and art. One reason I write my public domain posts is to promote a better balance between them, particularly in encouraging shorter copyright lengths to benefit both original creators and the public. Similarly, as I’ve noted in another recent post, I value both human creation and automated processes, but I increasingly see a need to improve the balance between those as well, especially as some corporations aggressively push “generative AI”. While I appreciate many ways in which automation can help us create and manage our work, I treasure the humanity that people thoughtfully put into the creation of literature and art of all kinds, and the human responses that those creations elicit.
Today I’m thankful for all of the people, most no longer with us, who made the works that are joining the public domain today. I’m thankful for the new opportunities we have to share and build on those works now that they’re public domain. I’m thankful to all the people who have responded to those works, whether as brief reactions or as new works as ambitious as the works they respond to. And I hope you’ll keeping making and sharing those responses with the world when you can. I look forward to reading them, and perhaps linking to them in future posts.
New Job: Project Gutenberg / Eric Hellman
Personal Note, January 1 2026: I have a new job: Executive Director of the Project Gutenberg Literary Archive Foundation. Here's what I wrote for PG's January Newsletter.
Greetings from the new Executive Director
Happy Public Domain Day! You might hear people say that books published in 1930 have "fallen" into the US Public Domain, or, that they have lost copyright "protection". This is not quite correct. Rather, books published in 1930 have been FREED of copyright restrictions. They have ASCENDED into the public domain and into the embrace of organizations like Project Gutenberg. They now belong to ALL of us, and we need to take care of them for future generations.
On October 21, Project Gutenberg lost its longtime leader, Greg Newby, to pancreatic cancer. I had agreed to step up as Acting Executive Director so that Project Gutenberg could continue the mission that had become Greg's life work: to serve and preserve public domain books so that all of us can use and enjoy them without restrictions. Although I've been doing development work for Project Gutenberg for the past 8 years, I did not really understand what Greg's job entailed, or how many tasks he had been juggling. Three months in, I'm still discovering mysterious-to-me aspects of the organization. I've also been amazed at the dedication and talent of the many volunteers behind Project Gutenberg and our sister organization, Distributed Proofreaders. And at the large number of donors who make the organization financially viable and sustainable. So as of 2026, with your support, I'm continuing as Executive Director.
In the past three months Project Gutenberg has proven to be resilient; we took a heavy blow and managed to keep going. My top priority going forward is to make Project Gutenberg even more sustainable as well as resilient. In other words, my job is be one runner in a relay race: take the baton and make sure I get it to the next runner. That's what we all have to do with public domain books, too. We want them to still be there in 50 years! Whether you're already a volunteer or booster, an avid reader, or just someone curious about what we do, I hope you'll help us pass that baton.
Sabotaging Bitcoin / David Rosenthal
![]() |
| Source |
In 2024 Soroush Farokhnia & Amir Kafshdar Goharshady published Options and Futures Imperil Bitcoin's Security and:
showed that (i) a successful block-reverting attack does not necessarily require ... a majority of the hash power; (ii) obtaining a majority of the hash power ... costs roughly 6.77 billion ... and (iii) Bitcoin derivatives, i.e. options and futures, imperil Bitcoin’s security by creating an incentive for a block-reverting/majority attack.
![]() |
| Source |
90% of transaction volume on the Bitcoin blockchain is not tied to economically meaningful activities but is the byproduct of the Bitcoin protocol design as well as the preference of many participants for anonymity ... exchanges play a central role in the Bitcoin system. They explain 75% of real Bitcoin volume.Of course, just because they aren't "economically meaningful" doesn't mean they aren't worth attacking! The average block has ~3.2K transactions, so ~$121.6M/block. As a check. $121.6M * 144 block/day = $17.5B. So to recover their cost for a 51% attack would require double-spending about 8 hours worth of transactions.
I agree with their technical analysis of the attack, but I believe there would be significant difficulties in putting it into practice. Below the fold I try to set out these difficulties.
Maciej Cegłowski
First, I should point out that I wrote about using derivatives to profit from manipulating Bitcoin's price more than three years ago in Pump-and-Dump Schemes. These schemes have a long history in cryptocurrencies, but they are not the attack involved here. I don't claim expertise in derivatives trading, so it is possible my analysis is faulty. If so, please point out the problems in a comment.
The Attack
Farokhnia & Goharshady build on the 2018 work of Ittay Eyal & Emin Gün Sirer in Majority is not enough: Bitcoin mining is vulnerable:The key idea behind this strategy, called Selfish Mining, is for a pool to keep its discovered blocks private, thereby intentionally forking the chain. The honest nodes continue to mine on the public chain, while the pool mines on its own private branch. If the pool discovers more blocks, it develops a longer lead on the public chain, and continues to keep these new blocks private. When the public branch approaches the pool's private branch in length, the selfish miners reveal blocks from their private chain to the public.In April 2024 Farokhnia & Goharshady observed that:
...
We further show that the Bitcoin mining protocol will never be safe against attacks by a selfish mining pool that commands more than 1/3 of the total mining power of the network. Such a pool will always be able to collect mining rewards that exceed its proportion of mining power, even if it loses every single block race in the network. The resulting bound of 2/3 for the fraction of Bitcoin mining power that needs to follow the honest protocol to ensure that the protocol remains resistant to being gamed is substantially lower than the 50% figure currently assumed, and difficult to achieve in practice.
Given that the rule of thumb followed by most practitioners is to wait for 6 confirmations, a fork that goes 6 levels deep can very likely diminish the public’s trust in Bitcoin and cause a crash in its market price. It is also widely accepted that a prolonged majority attack (if it happens) would be catastrophic to the cryptocurrency and can cause its downfall.But, as they lay out, this possibility is discounted:
The conventional wisdom in the blockchain community is to assume that such block-reverting attacks are highly unlikely to happen. The reasoning goes as follows:
- Reverting multiple blocks and specifically double-spending a transaction that has 6 confirmations requires control of a majority of the mining power;
- Having a majority of the mining power is prohibitively expensive and requires an outlandish investment in hardware;
- Even if a miner, mining pool or group of pools does control a majority of the mining power, they have no incentive to act dishonestly and revert the blockchain, as that would crash the price of Bitcoin, which is ultimately not in their favor, since they rely on mining rewards denominated in BTC for their income.
| Source |
These huge futures markets enable Farokhnia & Goharshady's attack:
In short, an attacker can first use the Bitcoin derivatives market to short Bitcoin by purchasing a sufficient amount of put options or other equivalent financial instruments. She can then invest any of the amounts calculated above, depending on the timeline of the attack, to obtain the necessary hardware and hash power to perform the attack. If the attacker chooses to obtain a majority of the hash power, her success is guaranteed and she can revert the blocks as deeply as she wishes. However, she also has the option of a smaller upfront investment in hardware in exchange for longer wait times to achieve a high probability of success. In any case, as long as her earnings from shorting Bitcoin and then causing an intentional price crash outweighs her investments in hardware, there is a clear financial incentive to perform such an attack. The numbers above show that the annual trade volume in Bitcoin derivatives is more than three orders of magnitude larger than the required investment in hardware. Thus, it is possible and profitable to perform such an attack.
Assumptions
Farokhnia & Goharshady make some simplifying assumptions:The justification for the first assumption is that it keeps our analysis sound, i.e. we can only over-approximate the cost by making this assumption. As for the second assumption, we note that electricity costs are often negligible in comparison to hardware costs and that our main argument, i.e. the vulnerability of Bitcoin to majority attacks and block-reverting attacks, remains intact even if the estimates we obtain here are doubled. Indeed, as we will soon see, the trade volume of Bitcoin derivatives is more than three orders of magnitude larger than the numbers obtained here.
- We only consider the cost of hardware at the time of writing. We assume the attacker is buying the hardware, rather than renting it and do not consider potential discounts on bulk orders.
- We ignore electricity costs as they vary widely based on location.
Goal
As Farokhnia & Goharshady stress, the success of a block-reverting attack is probabilistic, so the attacker needs to have a high enough probability of making a large enough profit to make up for the risk of failure.My analysis thus assumes that the goal of the attacker is to have a 95% probability of earning at least double the cost of the attack.
Attacker
There are two different kinds of attackers with different sets of difficulties:- Outsiders: someone who has to acquire or rent sufficient hash power.
- Insiders: someone or some mining pool who already controls sufficient hash power.
- Obtaining and maintaining for the duration of the attack sufficient hash power without detection.
- Obtaining and maintaining for the duration of the attack a sufficient short position in Bitcoin without detection.
Hash Power
The outsider's problems are more complex than the insider's.Outsider Attack
The outsider attacker requires three kinds of resource:- Mining rigs.
- Power to run the rigs.
- Data center space to hold the rigs.
Mining rigs
- Could they acquire mining rigs sufficient to provide 30% of the combined insider and outsider hash power, or ~43% of the pre-attack hash power?
- How long would it take to acquire the rigs?
- Would their acquisition of the rigs be detected?
Because the economic life of mining rigs is less than two years, the first part of Bitmain's production goes into maintaining the hash rate by replacing obsolete rigs. The second part goes into increasing the hash rate. If we assume that the outsider attacker could absorb the second part of Bitmain's production, how long would it take to get the necessary 43% of the previous hash power?
![]() |
| Source |
The lack of rigs to increase the hash rate over a period of much less than two years would clearly be detectable.
Power
The Cambridge Bitcoin Energy Consumption Index's current estimate is that the network consumes 22GW. The outside attacker would need 43% of this, or about 9.5GW, for the duration of the attack. For context, Meta's extraordinarily aggressive AI data center plans claim to bring a single 1GW data center online in 2026, and the first 2GW phase of their planned $27B 5GW Louisiana data center in 2030. The constraint on the roll-out is largely that lack of access to sufficient power. The attacker would need double the power Meta's Louisiana data center plans to have in 2030.Access to gigawatts of power is available only on long-term contracts and only after significant delays.
Data centers
![]() |
| Hyperion |
Estimates for AI data centers are that 60% of the capital cost is the hardware and 40% everything else. Thus the "everything else" for Meta's $27B 5GW data center is $10.8B. "Everything else" for the attacker's two similar data centers would thus be $21.6B. Plus say 5 years of interest at 5% or $5.4B.
Operational cost
Ignoring the evident impossibility of the outsider attacker amassing the necessary mining rigs, power and data center space, what would the operational costs of the attack be?It is hard to estimate the costs for power, data center space, etc. But an estimate can be based upon the cost to rent hash power, noting that in practice renting 43% of the total would be impossible, and guessing that renters have a 30% margin. A typical rental fee would be $0.10/TH/day so the costs might be $0.07/TH/day. The attack would have a 95% probability of needing 482EH/s over 34 days or less, so $516M or less.
Thus the estimated total cost for the hash power used in the attack would have a 95% probability of being no more than $7.66B. Plus about $27B in data center cost, which could presumably be repurposed to AI after the attack.
Insider Attack
![]() |
| Source |
The insider's loss of income from the blocks they would otherwise have mined would have a 95% probability of being 4,590 BTC or less, or about $425M.
Short Position
Both kinds of attackers need to ensure that, when the attack succeeds, they have a large enough short position in Bitcoin that would generate their expected return from the attack's decrease in the Bitcoin price. There are two possibilities:- When the attacker's chain is within one block of being the longest, they have ten minutes to purchase the shorts. There is unlikely to be enough liquidity in the market to accommodate this sudden demand, which in any case would greatly increase the price of the shorts. I will ignore this possibility in what follows.
- At the start of the attack the attacker gradually accumulates sufficient shorts. Even assuming there were enough liquidity, and that the purchases didn't increase the price, the attacker has to bear both the cost of maintaining the shorts for the duration of the attack, and the risk of the market moving up enough to cause the position to be liquidated.
Farokhnia & Goharshady note that:
At the time of writing, the open interest of BTC options is a bit more than 20 billion USD. Thus, a malicious party performing the attack mentioned in this work would need to obtain a considerable amount of the available put contracts. This may lead to market disruptions whose analysis is beyond the scope of this work. This being said, if the derivatives market continues to grow and becomes much larger than it currently is, purchasing this amount of contracts might not even be detected.There are two different kinds of market in which Bitcoin shorts are available:
- Regulated exchanges such as the CME offering options on Bitcoin and stock exchanges with Bitcoin ETFs and Bitcoin treasury companies such as Strategy.
- Unregulated exchanges such as Binance offering "perpetual futures" (perps) on Bitcoin.
Unregulated Exchanges
Patrick McKenzie's Perpetual futures, explained is a clear and comprehensive description of the derivative common on unregulated exchanges:Instead of all of a particular futures vintage settling on the same day, perps settle multiple times a day for a particular market on a particular exchange. The mechanism for this is the funding rate. At a high level: winners get paid by losers every e.g. 4 hours and then the game continues, unless you’ve been blown out due to becoming overleveraged or for other reasons (discussed in a moment).So the exchange makes money from commissions, and from the spread against the actual spot price. The price of the perp is maintained close to the spot price by the "basis trade", traders providing liquidity by shorting the perp and buying the spot when the perp is above spot, and vice versa. Of course, the spot price itself may have been manipulated, for example by Pump-and-Dump Schemes.
Consider a toy example: a retail user buys 0.1 Bitcoin via a perp. The price on their screen, which they understand to be for Bitcoin, might be $86,000 each, and so they might pay $8,600 cash. Should the price rise to $90,000 before the next settlement, they will get +/- $400 of winnings credited to their account, and their account will continue to reflect exposure to 0.1 units of Bitcoin via the perp. They might choose to sell their future at this point (or any other). They’ll have paid one commission (and a spread) to buy, one (of each) to sell, and perhaps they’ll leave the casino with their winnings, or perhaps they’ll play another game.
Where did the money come from? Someone else was symmetrically short exposure to Bitcoin via a perp. It is, with some very important caveats incoming, a closed system: since no good or service is being produced except the speculation, winning money means someone else lost.
How else does the exchange make money?
Perp funding rates also embed an interest rate component. This might get quoted as 3 bps a day, or 1 bps every eight hours, or similar. However, because of the impact of leverage, gamblers are paying more than you might expect: at 10X leverage that’s 30 bps a day.A "basis point (bps)" is "one hundredth of 1 percentage point", so 30bps/day is 0.3%/day or around 120%/year. But the lure of leverage is the competitive advantage of unregulated exchanges:
In a standard U.S. brokerage account, Regulation T has, for almost 100 years now, set maximum leverage limits (by setting minimums for margins). These are 2X at position opening time and 4X “maintenance” (before one closes out the position). Your brokerage would be obligated to forcibly close your position if volatility causes you to exceed those limits.Unregulated markets are different:
Binance allows up to 125x leverage on BTC.Although these huge amounts of leverage greatly increase the reward from a small market movement in favor of the position, they greatly reduce the amount the market has to move against the position before something bad happens. The first bad thing is liquidation:
One reason perps are structurally better for exchanges and market makers is that they simplify the business of blowing out leveraged traders. The exact mechanics depend on the exchange, the amount, etc, but generally speaking you can either force the customer to enter a closing trade or you can assign their position to someone willing to bear the risk in return for a discount.The bigger and faster the market move, the more likely the loss exceeds your collateral:
Blowing out losing traders is lucrative for exchanges except when it catastrophically isn’t. It is a priced service in many places. The price is quoted to be low (“a nominal fee of 0.5%” is one way Binance describes it) but, since it is calculated from the amount at risk, it can be a large portion of the money lost. If the account’s negative balance is less than the liquidation fee, wonderful, thanks for playing and the exchange / “the insurance fund” keeps the rest, as a tip.
In the case where the amount an account is negative by is more than the fee, that “insurance fund” can choose to pay the winners on behalf of the liquidated user, at management’s discretion. Management will usually decide to do this, because a casino with a reputation for not paying winners will not long remain a casino.The second bad thing is automatic de-leveraging (ADL):
But tail risk is a real thing. The capital efficiency has a price: there physically does not exist enough money in the system to pay all winners given sufficiently dramatic price moves. Forced liquidations happen. Sophisticated participants withdraw liquidity (for reasons we’ll soon discuss) or the exchange becomes overwhelmed technically / operationally. The forced liquidations eat through the diminished / unreplenished liquidity in the book, and the magnitude of the move increases.
Risk in perps has to be symmetric: if (accounting for leverage) there are 100,000 units of Somecoin exposure long, then there are 100,000 units of Somecoin exposure short. This does not imply that the shorts or longs are sufficiently capitalized to actually pay for all the exposure in all instances.McKenzie illustrates ADL with an example:
In cases where management deems paying winners from the insurance fund would be too costly and/or impossible, they automatically deleverage some winners.
So perhaps you understood, prior to a 20% move, that you were 4X leveraged. You just earned 80%, right? Ah, except you were only 2X leveraged, so you earned 40%. Why were you retroactively only 2X? That’s what automatic deleveraging means. Why couldn’t you get the other 40% you feel entitled to? Because the collective group of losers doesn’t have enough to pay you your winnings and the insurance fund was insufficient or deemed insufficient by management.For our purposes, this is an important note:
In theory, this can happen to the upside or the downside. In practice in crypto, this seems to usually happen after sharp decreases in prices, not sharp increases. For example, October 2025 saw widespread ADLing as (more than) $19 billion of liquidations happened, across a variety of assets.How does this affect the outsider attacker? Lets assume that the attack has a 95% probability of costing no more than $7.5B and would reduce the Bitcoin price from $100K to $80K in a single 4-hour period. With 10X leverage this would generate $200K/BTC in gains.
![]() |
| Source |
![]() |
| Source |
The way liquidation of a short works is that as the market moves up, the initial leverage increases. Each exchange will have a limit on the leverage it will allow so, allowing for the liquidation fee, if the leverage of the short position gets to this limit the exchange will liquidate it.
| Move % | Leverage |
| 0 | 10 |
| 1 | 11.1 |
| 2 | 12.5 |
| 3 | 14.3 |
| 4 | 16.7 |
| 5 | 20 |
| 6 | 25 |
| 7 | 33.3 |
| 8 | 50 |
| 9 | 100 |
In the unlikely event that the attack succeeds early enough to avoid liquidation there would have been one of those "sharp decreases in prices" that cause ADL, so as a huge winner it would be essentially certain that the attacker would suffer ADL and most of the winnings needed to justify the attack would evaporate.
Regulated Exchanges
The peak open interest in Bitcoin futures on the Chicago Mercantile Exchange over the past year was less than $20B, so even if we add together both kinds of exchange, the peak open interest over the last year isn't enough for the attacker.Conclusions
Neither an outsider nor an insider attack appears feasible.Outsider Attack
An outsider attack seems infeasible because in practice:- They could not acquire 43% or more of the hash power.
- Even if they could it would take so long as to make detection inevitable.
- Even if they could and they were not detected, the high cost of the rigs makes the necessary shorts large relative to the open interest, and expensive to maintain.
- These large shorts would need to be leveraged perpetual futures, bringing significant risks of loss of collateral through liquidation, and of the potential payoff being reduced through automatic de-leveraging.
- The attacker would need more than the peak aggregate open interest in Bitcoin futures over the past year.
Insider Attack
The order-of-magnitude lower direct cost of an insider attack makes it appear less infeasible, but insiders have to consider the impact on their continuing mining business. If the assumed 20% drop in the Bitcoin price were sustained for a year, the cost to the miner controlling 30% of the hash rate would be about 15,750 BTC or nearly $1.5B making the total cost of the attack (excluding the cost of carrying the shorts) almost $2B.![]() |
| Source |
![]() |
| Source |
mining-company stocks are still flying, even with cryptocurrency prices in retreat. That's because these firms have something in common with the hottest investment theme on the planet: the massive, electricity-hungry data centers expected to power the artificial-intelligence boom. Some companies are figuring out how to remake themselves as vital suppliers to Alphabet, Amazon, Meta, Microsoft and other "hyperscalers" bent on AI dominance.I wonder why the date is 2028! As profit-driven miners use their bouyant stock price to fund a pivot to AI the hash rate and the network difficuty will decrease, making an insider attack less infeasible. The drop in their customer's income will likely encourage Bitmain to similarly pivot to AI, devoting an increasing proportion of their wafers to AI chips, especially given the Chinese government's goal of localizing AI.
...
Miners often have to build new, specialized facilities, because running AI requires more-advanced cooling and network systems, as well as replacing bitcoin-mining computers with AI-focused graphics processing units. But signing deals with miners allows AI giants to expand faster and cheaper than starting new facilities from scratch.
...
Shares of Core Scientific quadrupled in 2024 after the company signed its first AI contract that February. The stock has gained 10% this year. The company now expects to exit bitcoin mining entirely by 2028.
A 30% miner whose rigs were fully depreciated might consider an insider attack shortly before the halvening as a viable exit strategy, since their future earnings from mining would be greatly reduced. But they would still be detected.
Counter-measures
Even if we assume the feasibility of both the hash rate and the short position aspects of the attack, it is still the case that for example, an attack with 30% of the hash power and a 95% probability of success will, on average, last 17 days. it seems very unlikely that the coincidence over an extended period of a large reduction in the expected hash rate and a huge increase in short interest would escape attention from Bitcoin HODl-ers, miners and exchanges, not to mention Bitmain. What counter-measures could they employ?![]() |
| Source |
- The 6-block rule is just a convention, there is no dial that can be turned.
- Much of the access to the Bitcoin blockchain is via APIs that typically have the 6-block rule hard-codded in.
- Many, typically low-value, transactions do not wait for even a single confirmation.
- Even it were possible, changing from a one-hour to a four-hour confirmation would have significant negative impacts on the Bitcoin ecosystem.
Acknowledgements
This post benefited greatly from insightful comments on a draft from Jonathan Reiter, Amir Kafshdar Goharshady and Joel Wallenberg, but the errors are all mine.Happy New Year / Ed Summers
Happy New Year
Blackeye peas and collard greens for us.
2025 down, 2026 to go / Coral Sheldon-Hess
I don’t think any of us went into 2025 thinking this would be an easy year, and boy wasn’t it. Rather than restate our collective challenges, I’m going to stick to my own little life and keep it short with bullet points and pictures. I’ll also list as many positives as I can, because even in the darkest of times, there are sources of joy and reasons to be grateful.
Real life, in no particular order
- (OK, there’s some order, let’s do the hard one first) My mom passed away in late October – our relationship was complicated, and so are my feelings, but I’m working through it all with a therapist
- (Bittersweet) I adopted Mom’s green cheek conure, Tutu (which we jokingly spell “Teauxtu,” thanks to my brother), who bonded to me quickly and thoroughly, but who is currently plucking all of his feathers anyway — we’re working on it with his vet, and I’m hopeful we’ll get him past this (though I’ll love him anyway, even if he’s a naked bird forever; I cannot overstate how sweet a little guy he is!)
- My brother got married to a really nice lady, and I’m so happy for them!
- In early October, we moved from Midcoast Maine to Western NY, where there is a larger and more active covid-conscious community – because the economy was so unstable, it took months to sell our house in Maine, but we did finally succeed, the day before Christmas

- We joined the mask bloc here for a D&D night and had a lot of fun – looking forward to getting involved in mask distribution and attending more events in 2026
- I finally read all of the Murderbot Diaries series by Martha Wells, spurred on by the release of the TV show – then I listened to the audiobooks, and it became my comfort listen for the whole latter part of the year
- I ran an accessible birding event here in Rochester, and I’m hoping to run more next year
- I took American Sign Language 101 and 102, which I’ve really enjoyed. I’ll be retaking 102 in early 2026 (it was a rough autumn/winter), and then I hope to move on to 103 and 104.
- I learned to darn (as in, mending damaged knitted items), and I’m working on expanding my mending skills repertoire – this is well-timed, since, besides his own feathers and our hair, Tutu loves to bite on shirts until they have holes in them
- We saw the most vivid aurora we’ve ever seen, including during our time in Alaska
- My aunt sent me her scone recipe, and I made so. many. scones. this year
Bird Buddies! (Outdoor birds captured with our feeder cameras.)






The three baby bluebirds in the video below brought us so much joy this spring! Here, they appear to have a little conference, to decide whether it’s time to fly away or not, before a starling shows up.
Bird Buddies who live indoors




Miscellaney!






Work, in no particular order
- I ran the first part of a major, potentially multi-year LibGuides accessibility remediation project (“potentially” because there’s a task force considering how important LibGuides are to us and our patrons, and maybe we will shift focus away from them or have a single accessibility remediator or … something I haven’t thought of, who knows?)
- I co-ran a task force focused on accessibility training for library staff
- We survived a university-wide “Cybersecurity Alignment” (or the first part of one?)
- I learned the developer side of Figma (kind of, I still find parts of it baffling)
- I learned a ridiculous amount about authentication to electronic resources (which would be really satisfying if I didn’t keep running into things I still don’t know)

Coming up in 2026, an incomplete list
- I’ll take more ASL classes
- I’ll mend some things
- I’ll ride my bike more, because there are more safe places to do so where I live!
- I’m hoping to work through more herbal studies and practice more of those skills – it may become vital in the near future
- There will be an IT “alignment” at work mid-year – it could mean anything from “nothing” to “now I report to Central IT (who doesn’t care about my MLIS and might actually consider it a negative) instead of the Library (who doesn’t require an MLIS for my position, but knows it’s a strength, and also my team is so good, I don’t want to lose them, OK, yes, I am very stressed about this possibility)”
- Our landlord will be selling our house, something I think (hope) he didn’t realize he’d be doing when we started renting here – we’ll get first right of refusal on buying it, and he’s willing to write us a lease for as long as we want one if we choose not to buy (NY law means the new owner will have to honor it), so while it’s a stressful situation, it’s not as bad as it could be
- A valued colleague (everyone in my department is valued, truly, but his institutional knowledge is unmatched) will be retiring, and because of budget constraints, that will probably mean some shuffling of responsibilities semi-permanently
- If we decide we aren’t moving (so, we decide to buy this house or to sign a longer lease), I’ll set up some garden beds and grow some things
- I want this to be a thing, but financially, it may not: I really very much want to build or buy a small (like, Class B, or C at the highest) camper so that I can travel again
“Last…” will be first / John Mark Ockerbloom
Ninety-five (or 100) years is a very long time for copyrights to last. But Olaf Stapledon saw a much longer future for us in Last and First Men (reviewed here and here). His book tells a story of successive human species over the next 2 billion years.
Stapledon died in 1950, and his work is already public domain most places outside the US. Tomorrow a copy in Australia will be among the first books to be relisted in my new books listing, finally free for all. #PublicDomainDayCountdown
2025-12-31: From Tables to Triumph: A PhD Journey in Uncertainty-Aware Scientific Data Extraction / Web Science and Digital Libraries (WS-DL) Group at Old Dominion University
In January 2021, I began a journey that would span nearly five years, three children, countless late nights, and a singular focus: teaching machines to extract data from complex scientific tables with confidence—and to know when they're uncertain. On October 29, 2025, I successfully defended my dissertation titled "SCITEUQ: Toward Uncertainty-Aware Complex Scientific Table Data Extraction and Understanding" at Old Dominion University. This milestone represents not just the culmination of intensive research but a testament to perseverance, family support, and the power of focused determination.
Finding My Path at LAMP-SYS
When I joined the Lab for Applied Machine Learning and Natural Language Processing Systems (LAMP-SYS), part of ODU's Web Science and Digital Libraries Research Group (WSDL) under the guidance of Dr. Jian Wu, I knew exactly what problem I wanted to solve: making scientific table data extraction both accurate and trustworthy through uncertainty quantification.
Scientific tables are ubiquitous in research papers, containing critical experimental data, statistical results, and research findings. Yet extracting this data automatically from PDF documents remains surprisingly difficult. Unlike the simple, well-structured tables you might see on Wikipedia, scientific tables are complex beasts—featuring multi-level headers, merged cells, irregular layouts, and domain-specific notations that confound even state-of-the-art machine learning models.
But here's the real problem: existing methods don't tell you when they're wrong. They extract data with the same confidence whether they're processing a simple table or struggling with a complex one. For scientific applications where accuracy is paramount, this means researchers must manually verify every single extracted cell—a task that doesn't scale when you're dealing with thousands of tables.
The Research Challenge
My research addressed a fundamental question: How can we build systems that not only extract data from complex scientific tables but also quantify their uncertainty, allowing us to focus human verification effort only where it's needed?
To tackle this challenge, I formulated four research questions:
RQ1: What is the status of reproducibility and replicability of existing TSR models?
Before building something new, I needed to understand what already existed. I conducted the first systematic reproducibility and replicability study of 16 state-of-the-art Table Structure Recognition (TSR) methods. The results were sobering: only 8 of 16 papers made their code and data publicly available, and merely 5 had executable code. When I tested these methods on my newly created GenTSR dataset (386 tables from six scientific domains), none of the methods replicated their original performance. This highlighted a critical gap in the field. This work was published at ICDAR 2023: "A Study on Reproducibility and Replicability of Table Structure Recognition Methods."
RQ2: How do we quantify the uncertainties of TSR results?
To address this, I developed TTA-m, a novel uncertainty quantification pipeline that adapts Test-Time Augmentation specifically for TSR. Unlike vanilla TTA, my approach fine-tunes pre-trained models on augmented table images and employs ensemble-based methods to generate cell-level confidence scores. On the GenTSR dataset, TTA-m achieved an F1-score of 0.798, with over 80% accuracy for high-confidence predictions—enabling reliable automatic detection of extraction errors. This work was published at IEEE IRI 2024: "Uncertainty Quantification in Table Structure Recognition."
RQ3: How can we integrate uncertainties from TSR and OCR for holistic table data extraction?
I designed and implemented the TSR-OCR-UQ framework, which integrates table structure recognition (using TATR), optical character recognition (using PaddleOCR), and conformal prediction-based uncertainty quantification into a unified pipeline. The results were compelling: the accuracy improved from 53-71% to 83-97% for different complexity levels, with the system achieving 69% precision in flagging incorrect extractions and reducing manual verification labor by 53%. This work was published at ICDAR 2025: "Uncertainty-Aware Complex Scientific Table Data Extraction."
RQ4: How well do LLMs answer questions about complex scientific tables?
To evaluate the QA capability of Large Language Models on scientific tables, I created SciTableQA, a benchmark dataset containing 8,700 question-answer pairs across 320 complex scientific tables from multiple domains. My evaluation revealed that while GPT-3.5 achieved 79% accuracy on cell selection TableQA tasks, performance dropped to 49% on arithmetic reasoning TableQA tasks—highlighting significant limitations of current LLMs when dealing with complex table structures and numerical reasoning. This work was published at TPDL 2025: "SciTableQA: A Question-Answering Benchmark for Complex Scientific Tables."
The SCITEUQ Framework
Putting it all together, SCITEUQ (Scientific Table Extraction with Uncertainty Quantification) represents a comprehensive solution to uncertainty-aware scientific table data extraction. The framework achieves the state-of-the-art performance while providing essential uncertainty quantification capabilities that enable efficient human-in-the-loop verification.
Each component contributes to a more reliable approach:
- GenTSR provides rigorous cross-domain evaluation
- TTA-m quantifies uncertainties in structure recognition
- TSR-OCR-UQ integrates structure and content extraction with uncertainty maps
- SciTableQA enables systematic evaluation of reasoning capabilities
Publications and Research Impact
My research resulted in five first-author publications at top-tier conferences and journals:
- ICDAR 2025: "Uncertainty-Aware Complex Scientific Table Data Extraction."
- TPDL 2025: "SciTableQA: A Question-Answering Benchmark for Complex Scientific Tables."
- IEEE IRI 2024: "Uncertainty Quantification in Table Structure Recognition."
- Nature Scientific Data 2023: "DeepPatent2: A Large-Scale Benchmarking Corpus for Technical Drawing Understanding."
- ICDAR 2023: "A Study on Reproducibility and Replicability of Table Structure Recognition Methods."
Each of these papers faced initial rejection before ultimately being accepted. This taught me an invaluable lesson: rejection is not failure; it's an opportunity to refine and improve your work.
Industry Experience: From Azure to Alexa to Microsoft AI
While my research focused on scientific tables, my internships at Microsoft and Amazon broadened my perspective on applying machine learning at scale.
Microsoft (Summers 2022, 2023, 2025)
My first two summers at Microsoft were with the Azure team, where I worked on infrastructure optimization problems far from my research area. I developed an AI-human hybrid LLM-based multi-agent system for AKS Cluster configuration, reducing cluster type generation time from 2 weeks to 1 hour (link to blog post). I also designed ML anomaly detection systems on Azure Synapse that reduced hardware maintenance costs by over 20% and formulated new metrics for characterizing node interruption rates that decreased hardware downtime by 25% (link to blog post).
In Summer 2025, I joined the Microsoft AI team under the Bing organization, working on problems at the intersection of large-scale search and AI—which is what I'll be doing when I return to Microsoft full-time in January 2026.
Amazon (Summer 2024)
At Amazon, I worked with the Alexa Certification Technology team in California, where I drove 10% customer growth by designing LLM-based RAG systems with advanced prompt engineering techniques and increased the revenue by over 5% by developing LLM Agents on AWS to improve Alexa-enabled applications (link to blog post).
These internships, while not directly related to my dissertation research, taught me how to apply ML thinking to diverse industrial problems and to work effectively in large, complex organizations.
Balancing PhD Life with Family
Perhaps the most challenging aspect of my PhD journey had nothing to do with research—it was combining my studies with raising three young children. My youngest son, Daniel, was born just six month after I was enrolled in the PhD program. Managing research deadlines, experimental runs, paper submissions, and the demands of parenting three boys (Paul, David, and Daniel) required discipline and sacrifice.
I developed a strict routine: work from 9 AM to 3 PM every day at my research lab, then pick up my kids from school and be fully present for them. This meant no late nights in the lab, no weekend marathons of coding—just consistent, focused work during designated hours. It wasn't always easy. Conference deadlines sometimes meant asking my wife, Olabisi, to take on even more, or my mother, Beatrice, to provide extra support. But this routine kept me grounded and taught me that quality of work matters more than quantity of hours.
The Defense
On October 27, 2025, I defended my dissertation before my committee:
- Dr. Jian Wu (Director) - Old Dominion University
- Dr. Michael L. Nelson - Old Dominion University
- Dr. Michele C. Weigle - Old Dominion University
- Dr. Sampath Jayarathna - Old Dominion University
- Dr. Yi He (Co-advisor) - College of William & Mary
Their thoughtful feedback, probing questions, and constructive critiques throughout my PhD journey were instrumental in refining my research and pushing me to think deeper about the implications and limitations of my work.
Lessons Learned
Looking back on nearly five years of doctoral work, several lessons stand out:
1. Embrace Rejection as Refinement
Four of my papers were initially rejected. Each rejection stung, but each one ultimately led to a stronger paper. The review process, while sometimes frustrating, forced me to clarify my arguments, strengthen my experiments, and address weaknesses I hadn't noticed. My TPDL 2025 paper on SciTableQA went through two rounds of revisions, but the final version is significantly better than the original submission.
2. Establish Non-Negotiable Boundaries
My 9 AM to 3 PM schedule wasn't just convenient—it was essential for maintaining my sanity and my family relationships. While some might argue that PhD students need to work 80-hour weeks, I proved that focused, disciplined work during reasonable hours can produce quality research. Those boundaries also made me more efficient: when you only have six hours a day, you learn to prioritize ruthlessly.
3. Build for Reproducibility from Day One
My systematic study on TSR reproducibility taught me the hard way how difficult it is to reproduce other people's work. This experience shaped how I approached my own research. Every framework I built—TTA-m, TSR-OCR-UQ, SciTableQA—comes with comprehensive documentation, publicly available code, and clear instructions for replication. Future researchers shouldn't struggle to build upon my work the way I struggled with others'.
4. Choose Problems That Matter to You
I entered my PhD knowing I wanted to work on table extraction with uncertainty quantification, and I never wavered from that focus. This singular vision helped me navigate the inevitable setbacks and distractions that come with doctoral research. When experiments failed or papers got rejected, I could always return to the core question: How do we make scientific data extraction both accurate and trustworthy?
5. Internships Broaden Your Perspective
While my Microsoft and Amazon internships didn't directly contribute to my dissertation, they fundamentally shaped how I think about research. Working on production systems with millions of users taught me to think about scalability, robustness, and real-world constraints in ways that academic research rarely emphasizes. These experiences make me a better researcher because I can now evaluate my work not just on benchmark performance, but on whether it could actually be deployed at scale.
Looking Forward
In January 2026, I'll be joining Microsoft as a Data Scientist 2 with the Microsoft AI team at the Redmond campus in Washington state. My family and I are excited about this new chapter—moving from Norfolk, Virginia, to the Pacific Northwest, and transitioning from academic research to industry applications.
While I'll be working on different problems at Microsoft, the skills and mindset I developed during my PhD—rigorous experimentation, systematic evaluation, uncertainty quantification, and reproducible research—will continue to guide my work. I'm particularly excited about the opportunity to apply research-driven thinking to real-world problems at a scale that can impact millions of users.
Acknowledgments
This journey would have been impossible without extraordinary support:
To Dr. Jian Wu, my advisor, mentor, and guide—thank you for believing in my research vision, for pushing me to think bigger, and for your patience during the inevitable frustrations of doctoral research. Your mentorship has not only shaped my research but also my approach to solving complex problems.
To Dr. Yi He, my co-advisor at William & Mary, your expertise and thoughtful feedback greatly enriched this research. Thank you for your guidance and support throughout this journey.
To my dissertation committee—Drs. Michael Nelson, Michele Weigle, and Sampath Jayarathna—your constructive critiques and expert insights were essential in refining my ideas and strengthening this work.
To my colleagues in WSDL and LAMP-SYS, the collaborative environment, intellectual exchanges, and camaraderie made this journey both enriching and memorable.
To my wife, Olabisi—you walked beside me every step of this journey with unwavering devotion and love. Your patience during the long hours, your understanding through the challenges, and your constant encouragement when the path seemed difficult made this achievement possible. This accomplishment is as much yours as it is mine.
To my sons—Paul, David, and Daniel—you are my greatest blessings and my constant source of joy and motivation. I hope this work serves as an example that with dedication and faith, you too can achieve your dreams.
To God Almighty, who is the source of all wisdom and strength, I give thanks and praise.
-Kehinde Ajayi (@KennyAJ)
I-80 / Ed Summers
Recorded in a Ohio service center somewhere along I-80.
The muted sound of cars and trucks passing at high speed about 200 feet away.
This is a snippet cut and pasted a few times in REAPER:
Soon graduating into the public domain / John Mark Ockerbloom
Yale’s 1905 commencement ceremonies included a honorary doctorate for British composer Edward Elgar, and a portion of his first “Pomp and Circumstance” march. It’s been a staple of graduation processions ever since. The full suite of five marches that Elgar finished takes about 30 minutes to play, but took nearly three decades to complete. The first march has long been in the US public domain; the last, published in 1930, joins it there in two days. #PublicDomainDayCountdown
Two great blues musicians, and 2000 more records / John Mark Ockerbloom
David Seubert writes that more than 2500 records from 1925 digitized by the UC Santa Barbara Library will soon be freely downloadable there. A full listing of these recordings is online (though note that recordings made in 1925 but not released that year won’t be public domain yet).
A highlight of the collection is “St. Louis Blues”, sung by Bessie Smith with Louis Armstrong on cornet. One of the top selling records of 1925, it will be public domain in 3 days. #PublicDomainDayCountdown
Marlene Dietrich comes to America / John Mark Ockerbloom
Marlene Dietrich enjoyed success on stage and screen in 1920s Berlin, but became an international star in 1930. That year she came to the United States to star in Morocco alongside Gary Cooper. Her performance was nominated for an Academy Award. So was the direction by Josef von Sternberg, who also directed her in The Blue Angel and several other films. Morocco was inducted into the National Film Registry in 1992, and will be inducted into the public domain in 4 days. #PublicDomainDayCountdown
Mind The GAAP / David Rosenthal
![]() |
| Source |
A gauge of risk on Oracle Corp.’s (ORCL) debt reached a three-year high in November, and things are only going to get worse in 2026 unless the database giant is able to assuage investor anxiety about a massive artificial intelligence spending spree, according to Morgan Stanley.Mutua reports that:
A funding gap, swelling balance sheet and obsolescence risk are just some of the hazards Oracle is facing, according to Lindsay Tyler and David Hamburger, credit analysts at the brokerage. The cost of insuring Oracle Corp.’s debt against default over the next five years rose to 1.25 percentage point a year on Tuesday, according to ICE Data Services.
The company borrowed $18 billion in the US high-grade market in September. Then in early November, a group of about 20 banks arranged a roughly $18 billion project finance loan to construct a data center campus in New Mexico, which Oracle will take over as tenant.
Banks are also providing a separate $38 billion loan package to help finance the construction of data centers in Texas and Wisconsin developed by Vantage Data Centers,
![]() |
| Source |
Below the fold I look into why Oracle and other hyperscalers desperate efforts to keep the vast sums they're borrowing off their books aren't working.
Part of the reason the market is unhappy started in mid-September with The Economist's The $4trn accounting puzzle at the heart of the AI cloud. It raised the issue that I covered in Depreciation, that the hardware that represents about 60% of the cost of a new AI data center doesn't last long. It took a while for the financial press to focus on the issuea, but now they have.
The most recent one I've seen was triggered by the outage at the CME (caused by overheating in Chicago in November!). In AI Can Cook the Entire Market Now Tracy Alloway posted part of the transcript of an Odd Lots podcast with Paul Kedrosky pointing out a reason I didn't cover why the GPUs in AI data centers depreciate quickly:
When you run using the latest, say, an Nvidia chip for training a model, those things are being run flat out, 24 hours a day, seven days a week, which is why they're liquid-cooled, they're inside of these giant centers where one of your primary problems is keeping them all cool. It's like saying ‘I bought a used car and I don't care what it was used for.’ Well, if it turns out it was used by someone who was doing like Le Mans 24 hours of endurance with it, that's very different even if the mileage is the same as someone who only drove to church on Sundays.There was a similar problem after the Ethereum merge:
These are very different consequences with respect to what's called the thermal degradation of the chip. The chip's been run hot and flat out, so probably its useful lifespan might be on the order of two years, maybe even 18 months. There's a huge difference in terms of how the chip was used, leaving aside whether or not there's a new generation of what's come along. So it takes us back to these depreciation schedules.
73% of Ethereum miners have just given up: “About 10.6 million RTX 3070 equivalents have stopped mining since the merge.”But this depreciation problem is only one part of why the market is skeptical of the hyperscalers technique for financing their AI data centers. The technique is called Conduit Debt Financing, and Les Barclays' Unpacking the Mechanics of Conduit Debt Financing provides an accessible explanation of how it works:
We strongly recommend that you do not hit eBay for a cheap video card, despite the listings reassuring you that this card was only used by a little old lady to play Minecraft on Sundays and totally not for crypto mining, and that you should ignore the burnt odor and the charred RAM. Unless you’re poor, and the card’s so incredibly cheap that you’re willing to play NVidia Roulette.
How well do miners treat their precious babies? “GPU crypto miners in Vietnam appear to be jet washing their old mining kit before putting the components up for sale.” There are real cleaning methods that involve doing something like this with liquid fluorocarbons — but the crypto miners seem to be using just water.
Conduit debt financing is a structure where an intermediary entity (the “conduit”) issues debt securities to investors and passes the proceeds through to an end borrower. The key feature distinguishing conduit debt from regular corporate bonds is that the conduit issuer has no substantial operations or assets beyond the financing transaction itself. The conduit is purely a pass-through vehicle, the debt repayment relies entirely on revenues or assets from the ultimate borrower.The article continues to examine Meta's deal in great detail, and notes some of the legal risks of this technique:
Think of it this way: Company A wants to borrow money but doesn’t want that debt appearing on its balance sheet or affecting its credit rating. So it works with a conduit entity, Company B, which issues bonds to investors. Company B takes that capital and uses it to build infrastructure or acquire assets that Company A needs. Company A then enters into long-term lease or service agreements with Company B, and those payments service the debt. On paper, Company A is just a customer making payments, not a debtor owing bondholders.
The structure creates separation. The conduit issuer’s creditworthiness depends on the revenue stream from the end user, not on the conduit’s own balance sheet (because there isn’t really one). This is why conduit debt is often referred to as “pass-through” financing, the economics flow through the conduit structure to reach the underlying obligor.
Legal risks when things break: Substantive consolidation (court merges conduit with sponsor), recharacterization (lease treated as secured financing), and fraudulent transfer challenges. The structures haven’t been stress-tested yet because hyperscalers are wildly profitable. But if AI monetization disappoints or custom silicon undercuts demand, we’ll discover whether bondholders have secured claims on essential infrastructure or are functionally unsecured creditors of overleveraged single-purpose entities.The article asks the big question:
Why would Meta finance this via the project finance markets? And why does it cost $6.5 billion more?The $6.5B is the total of the 1% extra interest above Meta's corporate bond rate over the 20 years.
That’s how much more Meta is paying to finance this new AI data center using the project finance market versus what they could have paid had they used traditional corporate debt. So why on earth is this being called a win? And even crazier, why are other AI giants like Oracle and xAI looking to copy it?
![]() |
| Meta data center |
Construction on the project was well under way when Meta announced a new financing deal last month. Meta moved the project, called Hyperion, off its books into a new joint venture with investment manager Blue Owl Capital. Meta owns 20%, and funds managed by Blue Owl own the other 80%. Last month, a holding company called Beignet Investor, which owns the Blue Owl portion, sold a then-record $27.3 billion of bonds to investors, mostly to Pimco.Under GAAP, when would Meta be required to treat it as a finance lease?
Meta said it won’t be consolidating the joint venture, meaning the venture’s assets and liabilities will remain off Meta’s balance sheet. Instead Meta will rent the data center for as long as 20 years, beginning in 2029. But it will start with a four-year lease term, with options to renew every four years.
This lease structure minimizes the lease liabilities and related assets Meta will recognize, and enables Meta to use “operating lease,” rather than “finance lease,” treatment. If Meta used the latter, it would look more like Meta owns the asset and is financing it with debt.
The joint venture is what is known in accounting parlance as a variable interest entity, or VIE for short. That term means the ownership doesn’t necessarily reflect which company controls it or has the most economic exposure. If Meta is the venture’s “primary beneficiary”—which is another accounting term of art—Meta is required to consolidate it.Does Meta have “the power to direct the activities" at the data center it will operate?:
Under the accounting rules, Meta is the primary beneficiary if two things are true. First, it must have “the power to direct the activities that most significantly impact the VIE’s economic performance.” Second, it must have the obligation to absorb significant losses of the VIE, or the right to receive significant benefits from it.
Blue Owl has control over the venture’s board. But voting rights and legal form aren’t determinative for these purposes. What counts under the accounting rules is Meta’s substantive power and economic influence. Meta in its disclosures said “we do not direct the activities that most significantly impact the venture’s economic performance.” But the test under the accounting rules is whether Meta has the power to do so.Does Meta receive "significant benefits"? Is it required to "absorb losses"?:
The second test—whether Meta has skin in the game economically—has an even clearer answer. Meta has operational control over the data center and its construction. It bears the risks of cost overruns and construction delays. Meta also has provided what is called a residual-value guarantee to cover bondholders for the full amount owed if Meta doesn’t renew its lease or terminates early.The lease is notionally for 20 years but Meta can get out every four years. Is Meta likely to terminate early? In other words, how likely in 2041 is Meta to need an enormous 16-year old data center? Assuming that the hardware has an economic life of 2 years, the kit representing about 60% of the initial cost would be 8 generations behind the state of the art. In fact 60% of the cost is likely to be obsolete by the first renewal deadline, even if we assume Nvidia won't actually be on the one-year cadence it has announced.
But what about the other 40%? It has a longer life, but not that long. The reason everyone builds new data centers is that the older ones can't deliver the power and cooling current Nvidia systems need. 80% of recent data centers in China are empty because they were built for old systems.
But the new ones will be obsolete soon:
Today, Nvidia's rack systems are hovering around 140kW in compute capacity. But we've yet to reach a limit. By 2027, Nvidia plans to launch 600kW racks which pack 576 GPU dies into the space one occupied by just 32.Current data centers won't handle these systems - indeed how to build data centers that do is a research problem:
To get ahead of this trend toward denser AI deployments, Digital Realty announced a research center in collaboration with Nvidia in October.If the design of data centers for Nvidia's 2027 systems is only now being researched, how likely is it that Meta will renew the lease on a data center built for Nvidia's 2025 systems in 2041? So while the risk that Meta will terminate the lease in 2029 is low, termination before 2041 is certain. And thus so are residual-value guarantee payments.
The facility, located in Manassas, Virginia, aims to develop a new kind of datacenter, which Nvidia CEO Jensen Haung has taken to calling AI factories, that consumes power and churn out tokens in return.
How does the risk of non-renewal play out under GAAP?
Another judgment call: Under the accounting rules, Meta would have to include the residual-value guarantee in its lease liabilities if the payments owed are “probable.” That could be in tension with Meta’s assumption that the lease renewal isn’t “reasonably certain.”Weil sums it up concisely:
If renewal is uncertain, the guarantee is more likely to be triggered. But if the guarantee is triggered, Meta would have to recognize the liability.
Ultimately, the fact pattern Meta relies on to meet its conflicting objectives strains credibility. To believe Meta’s books, one must accept that Meta lacks the power to call the shots that matter most, that there’s reasonable doubt it will stay beyond four years, and that it probably won’t have to honor its guarantee—all at the same time.
![]() |
| David Sacks Nov 6 |
OpenAI explicitly requested federal loan guarantees for AI infrastructure in an October 27 letter to the White House—which kindly refused the offer, with AI czar David Sacks saying that at least 5 other companies could take OpenAI’s place—directly contradicting CEO Sam Altman's public statements claiming the company doesn't want government support.After this PR faux pas some less obvious way taxpayer dollars could keep the AI bubble inflating had to be found. Just over two weeks later Thomas Beaumont reported that Trump signs executive order for AI project called Genesis Mission to boost scientific discoveries:
The 11-page letter, submitted to the Office of Science and Technology Policy, called for expanding tax credits and deploying "grants, cost-sharing agreements, loans, or loan guarantees to expand industrial base capacity" for AI data centers and grid components. The letter detailed how "direct funding could also help shorten lead times for critical grid components—transformers, HVDC converters, switchgear, and cables—from years to months."
Trump unveiled the “Genesis Mission” as part of an executive order he signed Monday that directs the Department of Energy and national labs to build a digital platform to concentrate the nation’s scientific data in one place.This appears to be a project of David Sacks, the White House AI advisor and a prominent member of the "PayPal Mafia". Sacks was the subject of a massive, 5-author New York Times profile entitled Silicon Valley’s Man in the White House Is Benefiting Himself and His Friends:
It solicits private sector and university partners to use their AI capability to help the government solve engineering, energy and national security problems, including streamlining the nation’s electric grid, according to White House officials who spoke to reporters on condition of anonymity to describe the order before it was signed.
The article quotes Steve Bannon:
- Mr. Sacks has offered astonishing White House access to his tech industry compatriots and pushed to eliminate government obstacles facing A.I. companies. That has set up giants like Nvidia to reap an estimate of as much as $200 billion in new sales.
- Mr. Sacks has recommended A.I. policies that have sometimes run counter to national security recommendations, alarming some of his White House colleagues and raising questions about his priorities.
- Mr. Sacks has positioned himself to personally benefit. He has 708 tech investments, including at least 449 stakes in companies with ties to artificial intelligence that could be aided directly or indirectly by his policies, according to a New York Times analysis of his financial disclosures.
- His public filings designate 438 of his tech investments as software or hardware companies, even though the firms promote themselves as A.I. enterprises, offer A.I. services or have A.I. in their names, The Times found.
- Mr. Sacks has raised the profile of his weekly podcast, “All-In,” through his government role, and expanded its business.
Steve Bannon, a former adviser to Mr. Trump and a critic of Silicon Valley billionaires, said Mr. Sacks was a quintessential example of ethical conflicts in an administration where “the tech bros are out of control.”
“They are leading the White House down the road to perdition with this ascendant technocratic oligarchy,” he said.
![]() |
| David Sacks Nov 24 |
“The way this works”, said an investor friend to me this morning: “is that when Nvidia is about to miss their quarter, Jen Hsun calls David Sacks, who then gets this government initiative to place a giant order for chips that go into a warehouse.”I think the six companies Sacks was talking about are divided into two groups:
I obviously can’t confirm or deny that actually happened. My friend might or might not have been kidding. But either way the White House’s new Science and AI program, Genesis, announced by Executive Order on Monday, does seem to involve the government buying a lot of chips from a lot of AI companies, many of which are losing money.
And David Sack’s turnaround from “read my lips, no AI bailout” (November 6) to “we can’t afford to [let this all crash]” tweet (November 24) came just hours before the Genesis announcement.
- OpenAI, Anthropic and xAI, none of whom have a viable business model.
- Meta, Google and Microsoft, all of whom are pouring the cash from their viable business models into this non-viable business,
So before they need to replace the 60% of the loan's value with the next generation of hardware in 2027 they need to find enterprise generative AI applications that are so wildly profitable for their customers that they will pay enough over the cost of running the applications to cover not just the payments on the loans but also another 30% of the loan value every year. For Meta alone this is around $30B a year!
And they need to be aware that the Chinese are going to kill their margins. Thanks to their massive investments in the "hoax" of renewable energy, power is so much cheaper in China that systems built with their less efficient chips are cost-competitive with Nvida's in operation. Not to mention that the Chinese chip makers operate on much lower margins than Nvidia. Nvidia's chips will get better, and so will the Chinese chips. But power in the US will get more expensive, in part because of the AI buildout, and in China it will get cheaper.
This won't end well
Update: 28th December 2025
![]() |
| Source |
If I were advising Xi, I’d counsel him to go for the jugular by engaging in AI-dumping, a repeat of their aughts steel-dumping playbook. It’s already underway — and working. Eighty percent of a16z startups use open-source Chinese models. Same story at Airbnb. China is registering similar or better performance as the American LLM leaders, but with a fraction of the capex. Flooding the market with competitive, less-expensive AI models will put pressure on the margins and pricing power of the Mag 7, taking down a frighteningly concentrated S&P and likely sending the U.S., possibly the globe, into recession.The average of the three US models' cost is $12.33. The average of the three Chinese models' cost is $1.36. The US models are 9 times more expensive, but they are nowhere near 9 times better.
2025-12-28: IEEE High-Performance Computing 2025 Trip Report / Web Science and Digital Libraries (WS-DL) Group at Old Dominion University
This trip can be broken down into two main parts: a short sightseeing visit to Delhi and Agra, followed by the conference in Hyderabad.
Delhi and Agra
The trip started with a couple of days in New Delhi. Since I had not been to this part of the world before, I wanted to take the opportunity to explore the city before the conference. Delhi is enormous, both geographically and culturally. Over the first two days of the trip, I ended up walking more than 50 kilometers.During this time, I visited several UNESCO World Heritage sites, including the Taj Mahal, Agra Fort, and landmarks within Delhi such as Qutub Minar, which is the world's tallest brick minaret. While in Delhi, I also met up with a few friends from Fermilab. Although we didn't do any sightseeing together, we did manage to go out for dinner one evening, which was a nice break from traveling and a fun way to catch up before the conference started. Starting the trip with sightseeing was a great contrast to the dense technical program that followed.
HiPC 2025
I arrived in Hyderabad on December 17, one day before the main technical program began. HiPC 2025 turned out to be the largest conference I have attended so far, both in terms of attendance and breadth of topics covered, spanning high-performance computing, AI systems, and quantum computing.
Day 1: December 18, 2025
The main conference started on December 18. One notable statistic that stood out in the opening session was that only 29% of all submitted papers were accepted to the main proceedings. I was fortunate that Zeus was part of that small fraction.
The Day 1 keynote was given by Dr. Pratyush Kumar from Sarvam AI. His talk focused on what it actually takes to train large language models from scratch. He walked through the challenges of setting up compute and data infrastructure and shared lessons learned while building LLMs in practice.One part I found especially interesting was his discussion of real-world applications, including examples where language models helped with educational videos with real-time audio in multiple languages, while keeping the same voice as the original speaker. Overall, the keynote gave a very practical view of LLM development, beyond just model architectures.
The rest of the day featured workshops and technical sessions covering HPC systems, AI, and education.
Day 2: December 19th, 2025
- Particle Swarm Optimization to perform an initial global search, identify promising regions of the search space,
- BFGS, a quasi-Newton method, for fast local convergence,
- Automatic Differentiation (AD) to compute gradients accurately without requiring users to manually derive them,
- and massively parallel GPU execution, where hundreds or thousands of independent optimizations run concurrently.
The algorithm operates in two phases. First, a small number of PSO iterations are used to improve the quality of the starting points. In the second phase, each particle independently invokes a BFGS optimization on the GPU, using forward-mode AD to compute gradients efficiently. Once sufficient convergence is reached, all the threads synchronize and terminate to stop early using atomic operations.
By running many independent optimizations in parallel on GPUs, Zeus achieves 10x--100x speedups over a perfectly parallel CPU implementation while also improving accuracy compared to existing GPU-based methods. One of the advantages of the parallel algorithm is that it is less sensitive to poor starting points, whereas for the sequential version, we must repeatedly restart until sufficient convergence is achieved.
In the talk, I also discussed experimental results from both synthetic benchmark functions, such as the Rastrigin and Rosenbrock, and a real-world high-energy physics application. The example plot shows simulated data from proton-proton collisions at the Large Hadron Collider. When protons collide, their quarks and gluons produce sprays of particles called jets. When two jets are produced, their invariant mass can be reconstructed and fitted by minimizing a negative log likelihood. The pull distribution measures how far each data point is from the fit, in units of its expected uncertainty. A good fit should have pulls fluctuating around zero and mostly within ±2σ. This shows agreement between the simulated data and the model prediction. I also touched on current limitations, such as handling objectives with discontinuous derivatives, and outlined future work, including deeper levels of parallelism and improved stopping criteria.
Presenting this work felt especially meaningful because it tied together my internship experience at Fermilab and my growing interest in high-performance computing. It was rewarding to share our ideas with the community and see how the broader themes of the conference connected directly with our contribution.
Day 3: December 20th, 2025
The third and final day focused heavily on AI/ML topics, along with a very interesting keynote speaker, and concluded with a quantum computing workshop.
The Day 3 keynote was given by Dr. Christos Kozryakis from Stanford University and NVIDIA Research. His talk focused on how AI workloads are shaping modern datacenter design. He argued that current AI systems often follow a supercomputing-style approach, which may not be the best fit as models continue to scale.
Instead, he made a case for scale-out AI systems, where efficiency and system-level design play a bigger role. One idea that stayed with me was his discussion of power and energy efficiency, especially the question of how much AI can realistically fit within a gigawatt of power.
Later in the day, I attended the Quantum Computing Workshop, which was one of the highlights of the conference for me. This workshop was particularly exciting for me, as I will be taking the Quantum Computing course in Spring 2026, and I am interested in exploring how Zeus could be mapped into a hybrid classical-quantum optimization algorithm.
To close the workshop, a speaker from Fujitsu presented the current state of their quantum research, including ambitious plans toward a 1000-qubit machine. After the workshop, I had several valuable discussions with experts in the field. In particular, Dr. Anirban Pathak provided initial guidance on how my current algorithm could be adapted toward a hybrid classical-quantum approach.
Additionally, Aravind Ratnam pointed me to Q-CTRL's learning tutorials, which he recommended as an excellent hands-on resource for building a stronger foundation in quantum computing.
To close the conference, I attended the banquet, which featured a cultural program and an Indian Dinner at the rooftop restaurant.
Closing Thoughts
As only my second conference, HiPC 2025 was both intense and deeply rewarding. Compared to my first conference, I felt noticeably more confident presenting my work, asking questions, and engaging with researchers across different fields. At the same time, the experience reinforced a familiar lesson that conferences are just as much about people and conversations as they are about papers and talks.
I am grateful for the opportunity to present this work, for the feedback I received, and for the many discussions that will shape my future research directions. HiPC 2025 was an unforgettable experience, and I hope to return again.
~Dominik Soós (@DomSoos)
One Year of Learning 2025 / Peter Murray
Inspired by Tom Whitwell's 52 things I learned in 2022, I started my own list of things I learned in 2023 and repeated it last year. Reaching the end of another year, it is time for Things I Learned In 2025. Part way through the year I had the brilliant idea of putting a learning at the bottom of my weekly newsletter, and that worked well until the middle of the year when I stopped publishing newsletter issues. So here is a half year of learnings.
What did you learn this year? Let me know on Mastodon or Bluesky.
In Ethiopia, time follows the sun like nowhere else
Because Ethiopia is close to the Equator, daylight is pretty consistent throughout the year. So many Ethiopians use a 12-hour clock, with one cycle of 1 to 12 — from dawn to dusk — and the other cycle from dusk to dawn. Most countries start the day at midnight. So 7:00 a.m. in East Africa Time, Ethiopia&aposs time zone, is 1:00 in daylight hours in local Ethiopian time. At 7:00 p.m., East Africa Time, Ethiopians start over again, so it&aposs 1:00 on their 12-hour clock.—If you have a meeting in Ethiopia, you'd better double check the time, The World from PRX, 30-Jan-2015
This could have easily gone in the Thursday Threads on time standards. There are 12 hours of daylight, numbered 1 through 12. Then 12 hours of night, numbered 1 through 12. What could be easier?
From Thursday Threads issue 104 on Long Term Digital Storage.
A biographer embedded with the Manhattan Project influenced what we think about the atomic bomb
In early 1945, a fellow named Henry DeWolf Smyth was called into an office in Washington and asked if he would write this book that was about a new kind of weapon that the US was developing. The guy who had called him into his office, Vannevar Bush, knew that by the end of the year, the US was going to drop an atomic bomb that had the potential to end the war, but also that as soon as it was dropped, everybody was going to want to know what is this weapon, how was it made, and so forth. Smyth accepted the assignment. It was published by Princeton University Press about a week after the bomb was dropped. It explained how the US made the bomb, but it told a very specific kind of story, the Oppenheimer story that you see in the movies, where a group of shaggy-haired physicists figured out how to split the atom and fission, and all of this stuff. The thing is, the physics of building an atomic bomb is, in some respects, the least important part. More important, if you actually want to make the thing explode, is the chemistry, the metallurgy, the engineering that were left out of the story.—Wars Are Won By Stories, On the Media, 22-Jan-2025
The quote above comes from the transcript of this podcast episode. I've thought about this a lot in the past week as the Trump administration's flood-the-zone strategy overwhelms the senses. In a valiant effort to cover everything that is news, I can't help but wonder about the lost perspective of what isn't being covered. And I wonder where I can look to find that perspective.
From Thursday Threads issue 105 on Facial Recognition.
The origin of the computer term "mainframe" comes from "main frame" — the 1952 name of an IBM computer's central processing section
Based on my research, the earliest computer to use the term "main frame" was the IBM 701 computer (1952), which consisted of boxes called "frames." The 701 system consisted of two power frames, a power distribution frame, an electrostatic storage frame, a drum frame, tape frames, and most importantly a main frame.—The origin and unexpected evolution of the word 'mainframe', Ken Shirriff's blog, 1-Feb-2025
"Mainframe" is such a common word in my lexicon that it didn't occur to me that its origins was from "main frame" — as in the primary frame in which everything else connected. I've heard "frame" used to describe a rack of telecommunications equipment as well, but a quick Kagi search couldn't find the origins of the word "frame" from a telecom perspective.
From Thursday Threads issue 106 on How much do you know about the credit card industry?.
It takes nearly 3¢ to make a penny, but almost 14¢ to make a nickel
FY 2024 unit costs increased for all circulating denominations compared to last year. The penny’s unit cost increased 20.2 percent, the nickel’s unit cost increased by 19.4 percent, the dime’s unit cost increased by 8.7 percent, and the quarter-dollar’s unit cost increased by 26.2 percent. The unit cost for pennies (3.69 cents) and nickels (13.78 cents) remained above face value for the 19th consecutive fiscal year—2024 Annual Report, United States Mint
I knew pennies cost the U.S. mint more than one cent to make, but I didn't realize that the cost of nickels is so much more out of whack. I also learned a new word: seigniorage — the difference between the face value of money and the cost to produce it.
From Thursday Threads issue 107 on the humble battery. Also this year, the U.S. mint stopped pressing pennies in November.
It is much harder to get to the Sun than it is to Mars
The Sun contains 99.8 percent of the mass in our solar system. Its gravitational pull is what keeps everything here, from tiny Mercury to the gas giants to the Oort Cloud, 186 billion miles away. But even though the Sun has such a powerful pull, it’s surprisingly hard to actually go to the Sun: It takes 55 times more energy to go to the Sun than it does to go to Mars.—It’s Surprisingly Hard to Go to the Sun, NASA, 8-Aug-2018
I suppose it that headline above needs some nuance. It is easy to get to the Sun...just escape Earth's gravity and point yourself there. It is hard to get to the Sun in a controlled way that means you won't burn up along the way.
From Thursday Threads issue 108 on Educational Technology.
There are now 23 Dark Sky Sanctuaries in the World
Rum, a diamond-shaped island off the western coast of Scotland, is home to 40 people. Most of the island — 40 square miles of mountains, peatland and heath — is a national nature reserve, with residents mainly nestled around Kinloch Bay to the east. What the Isle of Rum lacks is artificial illumination. There are no streetlights, light-flooded sports fields, neon signs, industrial sites or anything else casting a glow against the night sky. On a cold January day, the sun sets early and rises late, yielding to a blackness that envelopes the island, a blackness so deep that the light of stars manifests suddenly at dusk and the glow of the moon is bright enough to navigate by.—Take a Look: A Dark Scottish Isle Where Starlight Reigns Supreme, New York Times, 24-Feb-2025
The pictures that accompany this article from the New York Times are stunning (gift link). And to think that there are only 23 places in the world that have reached this level of commitment to the environment.
From Thursday Threads issue 109 on Generative AI in Libraries.
Mexico has only one gun store for the entire country
Mexico notes that it is a country where guns are supposed to be difficult to get. There is just one store in the whole country where guns can be bought legally, yet the nation is awash in illegal guns sold most often to the cartels.—Mexico faces off with U.S. gunmakers at the Supreme Court, NPR, 4-Mar-2025
And not only is there one gun store, the single store in Mexico is located on an army base and is run by soldiers, according to an article in the Associated Press from 2016.
From Thursday Threads issue 110 on Research into Generative AI.
Plants reproduce by spreading little plant-like things
This is where pollen comes in. Like sperm, pollen contains one DNA set from its parent, but unlike sperm, pollen itself is actually its own separate living plant made of multiple cells that under the right conditions can live for months depending on the species... So this tiny male offspring plant is ejected out into the world, biding its time until it meets up with its counterpart. The female offspring of the plant, called an embryosac, which you're probably less familiar with since they basically never leave home. They just stay inside flowers. Like again, they're not part of the flower. They are a separate plant living inside the flower. Once the pollen meets an embryosac, the pollen builds a tube to bridge the gap between them. Now it's time for the sperm. At this point, the pollen produces exactly two sperm cells, which it pipes over to the embryosac, which in the meantime has produced an egg that the sperm can meet up with. Once fertilized, that egg develops into an embryo within the embryosac, hence the name, then a seed and then with luck a new plant. This one with two sets of DNA.—Pollen Is Not Plant Sperm (It’s MUCH Weirder), MinuteEarth, 7-Mar-2025
From Thursday Threads issue 111 on End-to-end Encryption.
Most plastic in the ocean isn't from littering, and recycling will not save us
Littering is responsible for a very small percentage of the overall plastic in the environment. Based on this graph from the OECD, you can see littering is this teeny-tiny blue bar here, and mismanaged waste, not including littering, is this massive one at the bottom. Mismanaged waste includes all the things that end up either in illegal dump sites or burned in the open or in the rivers or oceans or wherever. The focus on littering specifically, it's an easy answer because obviously there's nothing wrong with discouraging people from littering, but it focuses on individual people's bad choices rather than systemic forces that are basically flushing plastic into the ocean every minute. Mismanaged waste includes everything that escapes formal waste systems. So they might end up dumped, they might end up burned, they might end up in the environment.—You're Being Lied To About Ocean Plastic, Business Insider via YouTube, 26-Sep-2024
Contrary to popular belief, most plastic in the Great Pacific Garbage Patch stems from the fishing industry, with only a small fraction linked to consumer waste. The video highlights that mismanaged waste, rather than individual littering, is the primary contributor to plastic pollution, with 82% of macroplastic leakage resulting from this issue. It emphasizes the ineffectiveness of recycling as a solution, noting that less than 10% of plastics are currently recycled, and the industry has perpetuated the myth that recycling can resolve the plastic crisis. Microplastics, which are increasingly recognized as a major problem, originate from various sources, including tires and paint, with new data suggesting that paint is a significant contributor.
From Thursday Threads issue 112 on Social Media Research.
"But where is everybody?!?" — the origins of Fermi's Paradox
The eminent physicist Enrico Fermi was visiting his colleagues at Los Alamos National Laboratory in New Mexico that summer, and the mealtime conversation turned to the subject of UFOs. Very quickly, the assembled physicists realized that if UFOs were alien machines, that meant it was possible to travel faster than the speed of light. Otherwise, those alien craft would have never made it here. At first, Fermi boisterously participated in the conversation, offering his usual keen insights. But soon, he fell silent, withdrawing into his own ruminations. The conversation drifted to other subjects, but Fermi stayed quiet. Sometime later, long after the group had largely forgotten about the issue of UFOs, Fermi sat up and blurted out: “But where is everybody!?”—All by ourselves? The Great Filter and our attempts to find life, Ars Technica, 26-Mar-2025
This retelling of the Fermi Paradox coms from this story about why, despite the vastness of the universe, we have yet to encounter evidence of extraterrestrial civilizations. Enrico Fermi famously posed the question, "Where is everybody?" suggesting a disconnect between the expectation of abundant intelligent life and the lack of observable evidence. With this comes the Great Filter notion...proposing that there may be significant barriers preventing intelligent life from becoming spacefaring. The article goes on to speculate where we are relative to the "Great Filter" — are we past it, or is it yet in front of us? In other words, have we survived the filter or is our biggest challenge ahead of us?
From Thursday Threads issue 113 on Copyright and Foundational AI Models.
The pronoun "I" was capitalized to distinguish it from similarly typset letters
In fact, the habit of capitalizing “I” was also a practical adaptation to avoid confusion, back in the days when m was written “ııı” and n was written “ıı.” A stray “i” floating around before or after one of those could make the whole thing hard to read, so uppercase it went. And now it seems perfectly logical.—I Have a Capital Suggestion for a New Pronoun, New York Times, 27-Mar-2025
I'm not buying the opinion author's underlying premise (capitalizing “they” in writing when it refers to a nonbinary person), but the origins of why we capitalize "I" and not other pronouns are fascinating.
From Thursday Threads issue 114 on Digital Privacy.
The word "scapegoat" originated in a 1530 bible translation
Early English Christian Bible versions follow the translation of the Septuagint and Latin Vulgate, which interpret azazel as "the goat that departs" (Greek tragos apopompaios, "goat sent out", Latin caper emissarius, "emissary goat"). William Tyndale rendered the Latin as "(e)scape goat" in his 1530 Bible. This translation was followed by subsequent versions up through the King James Version of the Bible in 1611: "And Aaron shall cast lots upon the two goats; one lot for the Lord, and the other lot for the scapegoat."—Scapegoat, Wikipedia

From Thursday Threads issue 115 on Public and Private Camera Networks.
"Leeroy Jenkins!!!!" was staged
It was one of the first memes ever, a viral sensation that went mainstream back when people still used dial-up internet. Yet the cameraman behind “Leeroy Jenkins” still seems stupefied that anyone fell for it.—The Makers Of 'Leeroy Jenkins' Didn't Think Anyone Would Believe It Was Real, Kotaku, 25-Dec-2017
First posted on May 10, 2005, this year marks the 20th anniversary of this bit of internet folklore. I remember when this first came out, and I totally believed it was real until earlier this year.
From Thursday Threads issue 116 on Government Surveillance.
Ammonium chloride may be the 6th basic taste
Ammonium chloride is a slightly toxic chemical most notably found in “salmiak,” a salt licorice candy, which is popular in northern Europe. In a new study, researchers found that the compound triggers a specific proton channel called OTOP1 in sour taste receptor cells, which fulfills one of the key requirements to be considered a primary taste like sweet, salty, sour, bitter, and umami. Ammonium is commonly found in waste products and decaying organic matter and is slightly toxic, so it makes sense that vertebrates evolved a specific taste sensor to recognize it.—Ammonium chloride tastes like nothing else. It may be the sixth basic taste, Big Think, 11-Oct-2023
From Thursday Threads issue 117 on Local Government Surveillance.
Banned in Texas / John Mark Ockerbloom
Struggle over academic freedom in Texas state universities has a long history. Today it’s often over race and gender; in the 1940s, it was over things like John Dos Passos’s USA trilogy. When the University of Texas Board of Regents banned it from classrooms, university president Homer Price Rainey objected to their interference. After they fired him, thousands protested on campus. The first part of the USA trilogy, The 42nd Parallel, joins the public domain in 5 days. #PublicDomainDayCountdown
Editorial Demand for AI Research: Evidence from Taylor & Francis Special Issue Calls for Papers / Journal of Web Librarianship
In the public domain soon, in libraries now / John Mark Ockerbloom
The Penn Libraries, where I work, has first editions of many of the works featured in my #PublicDomainDayCountdown . From today through Public Domain Day, the Libraries social media will feature photos of some distinctive books from 1930.
One of featured photos, also shown here, is of the 1930 edition of W. H. Auden‘s Poems, which made his work known to readers worldwide. Glynn Young writes about the “brilliant collection” of 30 poems, and one verse drama, joining the public domain in 6 days.

O come, let all 4,850 of us adore him / John Mark Ockerbloom
In 1925 the Associated Glee Clubs of America put on a concert like no other. 15 choral groups, with over 850 singers in all, came together in New York’s Metropolitan Opera House to sing a program broadcast on radio across America. Portions were electrically recorded, including “Adeste Fideles”, where the audience of 4000 joined in the carol. Lloyd Winstead writes about the record, which is now in the National Recording Registry, and joins the public domain in 7 days. #PublicDomainDayCountdown
The debut of a dramatic duo / John Mark Ockerbloom
Moss Hart wrote the first draft of Once in a Lifetime, a comedy about Hollywood’s transition to “talkies”, as a 25-year-old unknown. Established playwright George S. Kaufman helped revise it into a Broadway hit. Steve Vineberg calls it the “the finest comedy ever written by Americans”, and discusses how it began a long successful collaboration between the two writers. Still making theatergoers laugh in the 21st century, the 1930 play joins the public domain in 8 days. #PublicDomainDayCountdown
2025-12-24: Beyond Alt-Text: Surprising Truths About How Blind Users Experience Online Discussions / Web Science and Digital Libraries (WS-DL) Group at Old Dominion University
Introduction
Study Overview
Figure 2: Illustration of the interview study process.(Figure 2 in M.J. Ferdous et al.)
Key Findings
Design Implications
- Thread summarization could help users quickly understand the main points of a discussion without listening to every post.
- Context-aware navigation could allow users to follow specific reply chains or sub-conversations more easily.
- Text normalization could convert slang, abbreviations, and informal language into more screen-reader-friendly forms while preserving original content.
- Auditory differentiation could improve speaker identification and conversational flow in multi-participant discussions.
Conclusion
-- Md Javedul Ferdous (@jaf_ferdous)
Reference
See Dick and Jane free / John Mark Ockerbloom
Given how much “Dick and Jane” have been used sardonically, one might think Zerna Sharp’s schoolbook characters were already public domain. But you can’t copyright names, expressive style, or stock situations, and fair use allows limited copying for purposes like criticism, analysis, and parody.
In 9 days, the 1930 Elson Basic Readers introducing Dick and Jane join the public domain, and their stories’ full texts and original artwork can be reused without limit. #PublicDomainDayCountdown
An impressive body of work / John Mark Ockerbloom
Four writers get credit on the 1930 copyright registration for “Body and Soul”: composer Johnny Green, and lyricists Robert Sour, Edward Heyman, and Frank Eyton. But many more artists shaped the perennial jazz standard we know today. Among its more than 1700 recorded interpretations and variations are covers by Coleman Hawkins, Frank Sinatra, Billie Holiday, Ella Fitzgerald, Tony Bennett and Amy Winehouse. The song surrenders itself to the public domain in 10 days. #PublicDomainDayCountdown
Library Futures: Information Literacy in the Age of GenAi: my slides, notes, and an addendum / Mita Williams
Is Tweaking Enough?: A Follow-up Usability Study of Primo VE at an Academic Library / Journal of Web Librarianship
A controversial bestseller / John Mark Ockerbloom
Michael Gold’s bestselling novel Jews Without Money depicts the plight of poor East European immigrants in New York. It resonated with readers in 1930 facing not-yet-fully-acknowledged impacts of the Depression. Gold hoped it would counter antisemitic propaganda, but many readers grew repelled by his own Communist agitprop, and a later edition deleted his ending calling the workers’ revolution the “true Messiah”. Gold’s uncut original goes public domain in 11 days. #PublicDomainDayCountdown
Clouds / Ed Summers
Clouds
Weekly Bookmarks / Ed Summers
These are some things I’ve wandered across on the web this week.
🔖 Principles of AI use at the National Library of Finland
🔖 Harry Towell Yearbooks
🔖 Welcome to L5
🔖 extremely cool shell trick I did not know about
🔖 Blackbird & Jesu, Joy of Man’s Desiring - Hiromi Uehara
🔖 Prediction: AI will make formal verification go mainstream
🔖 Software Foundations Series
The Software Foundations series is a broad introduction to the mathematical underpinnings of reliable software.
The principal novelty of the series is that every detail is one hundred percent formalized and machine-checked: the entire text of each volume, including the exercises, is literally a “proof script” for the Coq proof assistant.
The exposition is intended for a broad range of readers, from advanced undergraduates to PhD students and researchers. No specific background in logic or programming languages is assumed, though a degree of mathematical maturity is helpful. A one-semester course can expect to cover Logical Foundations plus most of Programming Language Foundations or Verified Functional Algorithms, or selections from both.🔖 LISTERS: A Glimpse Into Extreme Bird Watching
🔖 Computers Are Bad
🔖 Rust By Example
Rust is a modern systems programming language focusing on safety, speed, and concurrency. It accomplishes these goals by being memory safe without using garbage collection.
Rust by Example (RBE) is a collection of runnable examples that illustrate various Rust concepts and standard libraries. To get even more out of these examples, don’t forget to install Rust locally and check out the official docs. Additionally for the curious, you can also check out the source code for this site.🔖 Climate Yap Recap: AI, Energy, Facilities, and Planning
🔖 duckdb-web-archive
🔖 State of HTML 2025
If you want to get an idea of how capable the web platform has gotten over the past few years, look no further than the list of categories covered in this survey.
Forms, graphics, performance features, methods of interacting with various system and device APIs, accessibility… It’s a wonder anybody is able to keep track of it all!🔖 Subscribe INTERTAPES is a collection of found cassette tapes from over the world. The audio fragments include: voice memos, field recordings, mixtapes, bootlegs & more.
🔖 fighting erasure: digitizing Gaza’s genocide and the War on Lebanon
🔖 World’s largest Internet Domain Database
🔖 Scholarly Communication Analytics
🔖 How we got hit by Shai-Hulud: A complete post-mortem
🔖 Free Gifts: Capitalism and the Politics of Nature
🔖 Marcin Wichary’s Keyboard, Typewriting, and Type Collection
The debut of Disney’s distinctive dog design / John Mark Ockerbloom
Norm Ferguson joined Disney’s animation team in 1929, and quickly had a major influence on the studio’s style. In The Chain Gang, the bloodhounds he designed to chase Mickey Mouse impressed Disney with their expressiveness. Ferguson went on to animate Minnie’s dog Rover with a variant of that design in The Picnic. The two cartoons, joining the public domain in 12 days, show a clear development into the dog who Ferguson would also animate in later cartoons named Pluto. #PublicDomainDayCountdown
The ever transforming public domain / John Mark Ockerbloom
The Internet Archive has multiple events in January for Public Domain Day, including a contest for film transformations of public domain works.
Those works themselves often involved transformation. Four songs that joined the public domain in 2024 were written for the stage musical Animal Crackers, which went public domain in 2025. The film adaptation, joining the public domain in 13 days, drops all four, but adds Groucho Marx’s signature song “Hello, I Must Be Going”. #PublicDomainDayCountdown
Birds in hand / John Mark Ockerbloom
HathiTrust has created a 1930 Publications Collection for Public Domain Day. It already has over 20,000 items opened, many by their Copyright Review Program, which finds works without copyright renewals.
Over 50,000 more items will open in 14 days, including the book version of The Maltese Falcon. The serial version was in last year’s #PublicDomainDayCountdown, but I haven’t to date found its now-rare magazine issues. I thank HathiTrust and all others bringing the public domain to the public!
🚀 Trustworthy AI for the Open Data Ecosystem – OKFN Newsletter End-of-Year 2025 / Open Knowledge Foundation
Annual Report, New Partnerships, Digital Public Goods, CKAN and Open Data Editor, Open Science and more.
The post 🚀 Trustworthy AI for the Open Data Ecosystem – OKFN Newsletter End-of-Year 2025 first appeared on Open Knowledge Blog.
Top Library Links of 2025 / Artefacto
No soccer skills required / John Mark Ockerbloom
Duke’s Center for the Public Domain has its Public Domain Day post out, listing many works joining the public domain in 2026, and explaining the complicated factual and legal determinations sometimes needed to verify their status.
Among the listed artworks researcher Jason Rosenberg found will join the public domain in 15 days is the Jules Rimet cup, created for the FIFA World Cup in 1930. The original trophy is now missing, but anyone can soon freely make their own. #PublicDomainDayCountdown
2025-12-17: Paper Summary: "Understanding Low Vision Graphical Perception of Bar Charts" / Web Science and Digital Libraries (WS-DL) Group at Old Dominion University
Motivation
Bar charts are a widely used method for presenting categorical data in digital content, education, and professional reports. While previous research has examined how sighted users and blind screen reader users interact with these charts, little is known about the experiences of low-vision users who rely on partial sight and screen magnifiers. This group faces unique visual challenges, including reduced acuity, contrast sensitivity issues, zoom-related blurring, and the need for constant panning. These factors can increase cognitive effort and reduce accuracy when interpreting visual information. In particular, the presence of stacked bars, misaligned elements, or irrelevant distractor bars can significantly impair perception. Without a clear understanding of these barriers, designers lack the evidence needed to create accessible visualizations. This study addresses that gap by analyzing how low-vision users perceive different bar chart layouts and identifying design factors that contribute to perceptual errors.
Background and Related Work
Bar charts are a widely used form of data visualization due to their simplicity and effectiveness in presenting categorical comparisons. Decades of research, starting with Cleveland and McGill, have examined how visual features such as alignment, spacing, and bar length influence accuracy in graphical perception. These findings have informed design guidelines for improving chart readability for sighted users. Later studies have built on this work by exploring graphical perception in different populations, including children and domain experts.
Research on accessibility for users with visual impairments has primarily focused on blind individuals using screen readers. These efforts have led to the development of alternative text descriptions and sonification tools for accessing chart content. However, low-vision users who retain partial sight and often use screen magnifiers have not received the same level of attention in visualization research.
Some studies have explored how low-vision users read text under different visual conditions, such as changes in font, color, or contrast. While informative, these findings do not address the unique challenges involved in interpreting graphical data. Issues like zoom-induced blurring, difficulty in scanning across the chart, and increased visual clutter from distractors can significantly impact comprehension. Prior work has not systematically examined these effects in chart reading tasks. This study addresses that gap by analyzing the visual perception of bar charts among low-vision users and identifying design factors that influence error rates.
Methodology
Participants
The study involved 25 low-vision participants (13 male, 12 female), aged between 17 and 37, all of whom regularly use screen magnifiers and have prior familiarity with bar charts. Participants were recruited via mailing lists and word of mouth. Eligibility criteria included proficiency with screen magnifier software and the ability to perform visual tasks involving bar charts.
Apparatus
A custom web-based chart viewer was developed to simulate interaction with bar charts under magnification. The system incorporated a built-in screen magnifier modeled after ZoomText to ensure user familiarity and eliminate learning overhead. The application logged all user interactions, including zoom and pan activity, in real time for detailed analysis. Participants viewed the interface on a 1440x1024 resolution monitor.
Experimental Design
Participants performed a percentage estimation task on various types of bar charts. The charts included simple bar charts, aligned and unaligned stacked bar charts, and divided single-column bar charts. Annotated bars were accompanied by distractor bars of varying heights, either short or tall. The goal was to estimate the percentage height of a shorter bar relative to a taller reference bar. Differences in bar height and separation distance were controlled based on prior graphical perception literature.
Procedure
After providing informed consent, participants completed a 20-minute practice session to become familiar with the interface. Each participant was then asked to perform the estimation task on a randomized set of charts. A time constraint of 15 seconds per chart was enforced to focus on perceptual rather than cognitive processing. Participants responded verbally, and a think-aloud protocol was used to capture their reasoning. An exit interview was conducted to gather qualitative feedback on strategies and challenges.
Data Analysis
Participant responses were compared against ground-truth values, and the log of absolute error was used to quantify perceptual discrepancies. Data points with a Pearson correlation below 0.7 were excluded to remove outliers, such as those resulting from fatigue or guessing. Both quantitative error rates and qualitative insights from verbal feedback and interviews were used to evaluate perception patterns and identify contributing factors.
Results
Evaluation
Standard Bar Charts
Most participants (90%) preferred having both bars visible within the same screen viewport. When bars were separated, the average log absolute error was 0.87 (median 0.91). For adjacent bars, the error dropped to 0.65 (median 0.63). The presence of tall distractors significantly increased errors. Charts with short distractors had an average error of 0.78, while those with tall distractors averaged 0.96. For separated bars with tall distractors, the error averaged 0.95. For adjacent bars with tall distractors, it averaged 0.72. However, for 12 participants, this pattern was reversed. In those cases, adjacent bars produced more errors (average 0.89) than separated bars (average 0.73).
Multi-Column Stacked Bar Charts
Stacked bar charts resulted in a higher average error (0.83) compared to unstacked bars (0.74). Within stacked charts, aligned bars had a lower error (0.64) than unaligned ones (0.77), indicating that alignment helped reduce estimation difficulty.
Single-Column Stacked Bar Charts (Divided Bars)
Divided bars showed the highest error rate, averaging 0.91. Participants struggled with separation and distractors. In some cases, adjacent segments were perceived as a single visual unit. Error rates were slightly higher for adjacent bars than for separated ones, with distractors contributing approximately 0.3 additional points of error.
Analyses
Separation Effect
In charts without distractors, separated bars were harder to compare due to the need for zooming and panning. Users often had to move back and forth to estimate bar heights, leading to increased confusion and error. In charts with distractors, participants tried to avoid panning by fitting both task bars into one view. This strategy increased blur, which resulted in higher error rates. For some participants, tall distractors in adjacent bars led to higher errors than in separated bars.
Influence of Distractors
Tall distractors affected perception more than short ones. They introduced visual clutter and blur, particularly for adjacent bar comparisons. Participants tended to focus on the tallest elements in the chart, which pulled attention away from relevant bars.
Blurring Effect
Most participants (80%) described difficulty estimating bar heights due to blurred edges at high zoom. For example, a bar that is 50 pixels high may appear to be anywhere between 40 and 60 pixels. Participants often needed to magnify and inspect each bar individually before comparing, which increased cognitive effort.
Unalignment Effect
Unaligned bars in stacked charts led to higher errors, consistent with prior findings. Participants had difficulty comparing bars without a common baseline. Some errors were also attributed to the "Parallel Line Illusion," where the length of a bar appears distorted when placed next to a much taller or shorter bar.
Discussion
The study highlights several perceptual barriers low-vision users face when interpreting bar charts. One major finding is that strategies effective for sighted users do not always transfer to low-vision settings. For example, adjacent bars typically improve comparison accuracy for sighted users, but in our study, adjacent bars combined with tall distractors caused more errors for some participants. This effect was linked to increased blurring at high zoom levels and difficulty distinguishing task bars from clutter.
Distractor bars played a central role in shaping perception. Tall distractors consistently drew visual attention away from the task-relevant bars, even when users attempted to compensate by adjusting magnification. Participants also experienced fatigue from repeated panning and zooming, especially during longer sessions with complex chart layouts.
Stacked and divided bar charts introduced additional challenges. Unaligned stacked bars resulted in greater estimation errors, consistent with prior research. In divided bars, participants sometimes interpreted adjacent segments as a single visual unit, particularly when separator lines were faint or obscured by magnification. These results emphasize the importance of alignment and segment clarity for accessible chart design.
Participants relied heavily on screen magnifiers but noted that these tools were not well suited for structured visual tasks like chart reading. Continuous adjustments, frequent re-orientation, and loss of context were recurring concerns.
Limitations
- The annotation dot on bars was placed near the lower end rather than centered, which may have affected bar identification and height estimation.
- Charts were shown in a fixed color and contrast scheme, without allowing users to adjust visual settings based on their preferences or needs.
- The participant pool, while diverse in age and diagnosis, did not capture the full range of low-vision profiles, especially with respect to contrast sensitivity and field loss.
- We did not collect data on participants’ visual field characteristics, which could have provided further insight into perception strategies.
- Think-aloud and interview data provided valuable qualitative insights but lacked the precision that could be obtained through eye-tracking or gaze-based methods.
Conclusion
This study examined how low-vision users perceive bar charts using screen magnifiers, revealing key factors that affect visual accuracy, including bar alignment, spacing, and the presence of distractors. Results from a lab study with 25 participants showed that tall distractors, unaligned bars, and divided stacked layouts significantly increased estimation errors, often due to blurring and visual overload during magnified viewing. Unlike prior research focused on sighted or blind users, this work highlights the distinct challenges low-vision users face in interpreting visual data. These insights can guide the design of more accessible chart layouts and inform future development of adaptive tools, such as intelligent magnifiers with automated panning. Further research will include eye-tracking to better understand visual attention patterns and support more effective accessibility solutions.
References
- AKSHAY KOLGAR NAYAK (@AkshayKNayak7)
2025-12-17: Paper Summary: "Towards Enhancing Low Vision Usability of Data Charts on Smartphones" / Web Science and Digital Libraries (WS-DL) Group at Old Dominion University
The IEEE Visualization Conference (VIS) is the premier forum for research in visualization and visual analytics. It brings together a global community of scholars, practitioners, and designers committed to advancing visual methods for exploring, analyzing, and communicating data. This year, the 35th IEEE VIS Conference (VIS 2024) was held in St. Pete Beach, Florida, USA, from October 13 to 18, 2024. In this blog post, I highlight my work “Towards Enhancing Low Vision Usability of Data Charts on Smartphones,” which addresses the accessibility challenges faced by low-vision users when interacting with data charts on mobile devices and introduces GraphLite, a mobile assistive technology that transforms static charts into customizable interactive visualizations.
Motivation
Data charts are a powerful way to communicate trends, comparisons, and summaries, and they are now commonly found across news articles, financial tools, health dashboards, and social media platforms. For people with low vision, interacting with these charts on smartphones remains a significant challenge. While screen magnifiers allow users to enlarge content, doing so often hides important visual context, such as axis labels or relationships between data points. As a result, users are forced to pan repeatedly, remember individual values, and mentally piece together information, which can be both frustrating and cognitively demanding. Most existing accessibility solutions focus on blind screen reader users and do not address the specific needs of low-vision magnifier users. To bridge this gap, we introduce GraphLite, a mobile assistive technology that converts static charts into interactive and customizable visualizations tailored for low-vision users.
Background and Related Work
Research on chart accessibility has primarily focused on blind users, offering solutions such as textual descriptions (e.g., Alt-Text generation), sonification (e.g., audio-based data exploration), and table conversions (e.g., replacing charts with structured tables). These approaches are designed for non-visual interaction and do not support the needs of users who depend on residual vision and prefer visual formats.
For low-vision users, who typically rely on screen magnifiers, the challenges are different. Magnification can obscure spatial relationships, disconnect data points from axis labels, and require constant panning, which increases cognitive effort. Prior studies have proposed general techniques like space-preserving magnifiers (e.g., SteeringWheel) and responsive visualization frameworks for small screens (e.g., MobileVisFixer, Dupo). While these techniques reduce layout disruption, they are not optimized for analytical tasks like comparing distant values or identifying trends in visual data.
These gaps highlight the need for solutions that not only preserve visual context during magnification but also support selective focus, customization, and seamless navigation across chart elements. Our work builds on this insight by designing an approach that directly targets these unmet needs in mobile chart interaction for low-vision users.
Uncovering Usability Issues with Chart Interaction
To better understand the challenges low-vision users face when interacting with data charts on smartphones, we conducted a semi-structured interview study with 14 participants who regularly use screen magnifiers. Participants had a variety of eye conditions and reported different levels of visual acuity. The interviews explored participants’ everyday encounters with charts, typical usage contexts, and the specific difficulties they experience when using magnification to interpret chart content. All sessions were conducted remotely and recorded with consent. We applied open and axial coding to extract recurring themes that directly shaped our design decisions.
Participants described three core challenges. First, they found it difficult to associate data points with axis labels when zoomed in, often losing context due to the limited field of view. Second, making comparisons between distant values in a chart was described as mentally exhausting, requiring repeated panning and memorization. Lastly, many resorted to memorizing individual values and mentally reconstructing comparisons, especially when dealing with larger datasets. These insights revealed a clear need for features that minimize panning, support selective data access, and preserve visual relationships between chart elements. This feedback directly informed the design principles behind our solution.
GraphLite Design
Key Features
GraphLite introduces three core features designed to address the chart interaction challenges identified in our user study: selective data viewing, personalized visual styling, and simplified gesture-based navigation.
Selective Data Viewing
Participants expressed difficulty in comparing distant data points within magnified views. To support targeted analysis, GraphLite allows users to selectively choose specific bars or line segments for focused viewing. These selections are grouped into distinct views, enabling users to compare relevant data without the need to pan across the entire chart. Users can switch between these views using simple swipe gestures, reducing cognitive load and helping retain context during visual comparisons.
Personalized Visual Styling
To accommodate varying visual preferences and improve clarity, GraphLite provides customization controls for color, contrast, font size, and background themes. Users can tailor the appearance of charts to match their visual comfort, which was particularly important for those who found high contrast or dark themes more readable. Styling adjustments are applied directly to the interactive chart interface, allowing for more legible and less fatiguing interaction.
Simplified Gesture-Based Navigation
Instead of relying on complex multi-finger gestures often required by default magnifiers, GraphLite supports one-finger tap, swipe, and long-press interactions for common tasks such as selecting data, switching views, or opening configuration menus. This design reduces the need for precision input and lowers the physical effort involved in exploring chart content on small screens.
GraphLite Architecture
The design of GraphLite was shaped by insights from our interview study and prior research on low-vision interaction. The system architecture supports automatic chart detection, data extraction, and personalized rendering within a mobile browser interface. Below, we outline the key components of the system.
Design Considerations and Requirements
GraphLite was designed to support selective focus, reduce cognitive load, and preserve spatial context during magnified interaction. Our goal was to enable users to control both the content and appearance of charts. To support these needs, we focused on three design principles: enabling selective data views to minimize unnecessary information, allowing visual customization to improve clarity, and offering simplified gestures to reduce physical effort. These principles guided both the system’s architecture and interface behavior.
Overview
When a user opens a webpage in the GraphLite browser, the system identifies and processes chart images on the page. A trained classifier recognizes whether an image is a data chart and determines its type, such as bar or line. Once detected, the chart is transformed into a structured, interactive version that replaces the original static image. The user can tap a chart to activate GraphLite’s proxy interface. From here, they can select specific data points, customize appearance settings, and create multiple views that group relevant values together. Navigation between views is performed using simple swipe gestures. This structure lets users focus on comparing small subsets of data without needing to pan across the entire chart.
Chart Data Extraction
GraphLite uses ChartOCR, a hybrid deep learning model, to extract data from chart images. For bar charts, the system detects the corners of each bar and calculates height values. For line charts, it clusters key pivot points and reconstructs the full line segments. OCR is performed using AWS-Rekognition to extract axis labels, titles, and legends. The final output is a structured JSON representation containing all chart elements, which is used to render the interactive proxy.
Proxy Interface Design
The proxy interface is built for screen magnifier users and avoids reliance on multi-finger gestures. Users interact with charts using tap, swipe, and long-press actions. A swipe up gesture opens the theme picker, allowing users to adjust color, font, and contrast settings. A long press on the chart opens a list of x-axis labels as checkboxes, enabling selection of specific data points to include in a focused view. Users can create multiple such views and move between them with left or right swipes. To reduce panning, GraphLite applies space compaction, optimizing layout spacing while preserving interpretability.
Implementation Details
GraphLite is implemented as an Android mobile browser using the Flutter framework. The app extracts page content and sends it to a back-end server for chart processing. Image classification is handled by a custom-trained Inception-V3 model, and chart data extraction is performed using the ChartOCR pipeline. The server communicates with the app using JSON-based APIs. For rendering charts in the proxy interface, we use the Syncfusion Flutter charting library, which supports full customization and interaction. All components are modular, allowing future upgrades to chart models or OCR engines without rewriting the interface logic.
Evaluation
To assess the effectiveness of GraphLite, we conducted a comprehensive user study comparing it against three baseline methods: the default screen magnifier, a table-based chart conversion tool, and a space compaction-only interface. The study aimed to evaluate GraphLite’s impact on task performance, usability, workload, and user satisfaction when interacting with data charts on smartphones.
Participants and Study Design
We recruited 26 low-vision participants who regularly use screen magnifiers on mobile devices. The participant pool included individuals with varying eye conditions and levels of visual acuity. The study followed a within-subjects design, where each participant completed chart-based tasks across five conditions: screen magnifier (SM), tabular representation (TBL), space compaction only (SC), space compaction with customization (SCC), and the full GraphLite system (SCCF). Tasks included pairwise comparisons, selective filtering, trend prediction, and trend comparison, using both bar and line charts.
Procedure: Data Collection and Analysis
Participants performed 12 tasks across the five study conditions. Tasks were designed to simulate realistic data interaction scenarios, such as comparing stock values or filtering sales figures. We measured task completion time, task success rate, and accuracy. After each condition, participants completed the System Usability Scale (SUS) and NASA-TLX questionnaires, followed by a semi-structured interview to gather qualitative feedback. Quantitative data was analyzed using non-parametric statistical tests due to the non-normal distribution of some measures. Interview responses were coded thematically to identify common experiences and suggestions.
Figure 4 Prakash et al.: Task completion times in seconds across all conditions and tasks. For the MBF task, only TBL and SCCF results are shown, as participants could not complete the task in the SM, SC, and SCC conditions within the allotted time.Task Performance
GraphLite’s full system (SCCF) consistently led to faster completion times across all tasks. For example, participants completed pairwise comparison tasks nearly twice as fast with SCCF compared to the default magnifier. In more complex filtering and trend analysis tasks, SCCF outperformed both the tabular baseline and partial GraphLite conditions, especially when multiple data points had to be evaluated. Task success rates with SCCF reached 100 percent, compared to 61.5 percent under the default magnifier.
Usability and Workload Scores
SCCF received the highest SUS scores (average = 85.8), indicating strong perceived usability. By contrast, the screen magnifier condition received the lowest average score (51.3). NASA-TLX scores showed a significant reduction in cognitive load with SCCF (average = 22.7), compared to high workload ratings in the SM condition (average = 70.1). The TBL condition showed moderate performance, with usability and workload scores improving for simpler tasks but degrading as task complexity increased.
Error Patterns and Observations
In the screen magnifier condition, participants often lost track of axis labels or misestimated bar heights due to limited field of view. Errors were also common in the SC and SCC conditions, especially when users attempted to manually compare distant data points. In the TBL condition, errors were typically due to mental overload while scanning long tables. The SCCF condition showed the fewest errors, aided by selective data views and tooltip support.
Qualitative Feedback
Participants described GraphLite as significantly easier to use compared to other methods. Many appreciated the ability to focus on a subset of data without panning, and the option to adjust contrast and font size. Users noted that they felt more in control when they could select and compare relevant values directly. Several participants mentioned that it was the first time they could analyze trends visually without assistance. Suggestions for future improvement included adding vertical compaction, auto-focus gestures, and quick-access tooltips.
Discussion
Limitations
- GraphLite currently supports only bar and line charts, which limits its applicability to other common formats like pie, scatter, or stacked charts.
- The system assumes clean and well-aligned input images. Charts with low resolution, visual clutter, or unusual layouts may lead to extraction errors.
- Our study was conducted on a controlled device, not on users’ personal smartphones, which may affect the generalizability of usage behavior.
- While GraphLite supports visual customization, it does not currently integrate audio feedback, screen reader compatibility, or other multimodal access features.
Future Work
One promising direction is to enhance data comprehension through details-on-demand (DoD) interactions. These techniques allow users to request specific information as needed, keeping the interface clean while still providing access to deeper insights. Future implementations could explore how selection-based and zoom-based DoD methods can reduce the need for constant panning and support a more efficient workflow for low-vision users. A controlled study that simulates various DoD techniques could help identify which specific interactions are most beneficial for tasks like trend recognition or value comparison, and how they can be optimized for mobile magnifier users.
Another avenue for future research is predictive magnification, which involves automatically guiding the user’s focus to salient regions of a chart. Participants in our study expressed a desire for such behavior, especially to minimize the number of manual gestures. Building on prior work in content saliency and gaze tracking, future systems could use predictive models to pan toward important chart elements like axis labels or data peaks. This could reduce navigation effort and cognitive load during analysis. Integrating predictive magnification with user-driven control mechanisms may lead to a more fluid and supportive chart navigation experience for low-vision users.
Conclusion
GraphLite demonstrates how mobile-first, visualization-specific accessibility tools can improve data comprehension for low-vision users. By transforming static charts into interactive, customizable, and compact interfaces, the system supports faster, more accurate, and less effortful interaction. User studies confirm significant improvements over screen magnifiers and tabular alternatives, validating the importance of selective focus and visual personalization. While current limitations include chart type support and generalizability to real-world images, GraphLite lays the groundwork for future research in predictive navigation and adaptive visualization design. It represents a step toward more inclusive visual interfaces, where accessibility is embedded as a foundational element rather than an add-on.
References
A psychoanalyst’s desire for a saner world / John Mark Ockerbloom
In Civilization and its Discontents, Sigmund Freud turned his psychoanalytic attention from the troubles of the individual to those of the world. The book, joining the US public domain in 16 days in both German and English, diagnoses inherent conflicts between individual drives for happiness and society’s needs to tame those drives for survival. Kristin Dorsey suggests ways to read Freud’s work in the context of its time, and in dialogue with works on similar questions. #PublicDomainDayCountdown
Data Centers In Spaaaace! / David Rosenthal
![]() |
| Pigs in Spaaaace! |
To buy into SpaceX’s audacious $1.5 trillion valuation in a listing next year, investors will need to have faith in Elon Musk’s equally galactic vision for his rocket and satellite maker, from orbital data centers to lunar factories to human settlements on Mars.I chose one that ought to be more credible than Musk from Scientific American. Jeremy Hsu's Data Centers in Space Aren’t as Wild as They Sound reports that:
In early November Google announced Project Suncatcher, which aims to launch solar-powered satellite constellations carrying its specialty AI chips, with a demonstration mission planned for 2027. Around the same time, the start-up Starcloud celebrated the launch of a 60-kilogram satellite with an NVIDIA H100 GPU as a prelude to an orbital data center that is expected to require five gigawatts of electric power by 2035.To do Hsu justice, he did point out a few of the problems. But follow me below the fold for more.
A couple of weeks ago an anonymous engineer posted Datacenters in space are a terrible, horrible, no good idea:
In the interests of clarity, I am a former NASA engineer/scientist with a PhD in space electronics. I also worked at Google for 10 years, in various parts of the company including YouTube and the bit of Cloud responsible for deploying AI capacity, so I'm quite well placed to have an opinion here.You should absolutely read the whole thing, it is magnificent. I don't have this background, but I don't need it to know that they are a terrible idea. All I need is arithmetic.
The short version: this is an absolutely terrible idea, and really makes zero sense whatsoever. There are multiple reasons for this, but they all amount to saying that the kind of electronics needed to make a datacenter work, particularly a datacenter deploying AI capacity in the form of GPUs and TPUs, is exactly the opposite of what works in space. If you've not worked specifically in this area before, I'll caution against making gut assumptions, because the reality of making space hardware actually function in space is not necessarily intuitively obvious.
Steven Clark reports that Investors commit quarter-billion dollars to startup designing “Giga” satellites:
K2 is designing two classes of satellites—Mega and Giga .... The company’s first “Mega Class” satellite is named Gravitas. Gravitas will also deploy twin solar arrays capable of generating 20 kilowatts of power.Gravitas can launch on a Falcon 9, but Giga-class satellites need Starship. In Power: The answer to and source of all your AI datacenter problems Tobias Mann reports that:
...
K2 says Gravitas is “on par with the largest satellites that have ever been produced.” But K2 won’t stop there. The firm’s next satellite iteration, known as Giga, .... Underpinning Giga is its ability to generate up to 100 kilowatts of power per satellite ... Examples of missions Giga can support include AI computing
Today, Nvidia's rack systems are hovering around 140kW in compute capacity. But we've yet to reach a limit. By 2027, Nvidia plans to launch 600kW racks which pack 576 GPU dies into the space once occupied by just 32.So the largest satellites ever can power 14% of a current Nvidia rack. The next generation that requires Starship to launch can power 70% of a current Nvidia rack. By the time they can launch it, it will power 17% of a contemporary Nvidia rack. See, progress!
Starcloud's "5GW" satellite could power over 8000 2027 Nvidia racks in 2035 but it would be 50,000 times larger than a satellite that is much larger than one "on par with the largest satellites that have ever been produced".
The next question is "How many racks does one data center hold?" Yasir Shinaishin of ABB wrote in New Strategies in Design to Meet the Demands of AI Data Centers:
It is not uncommon to see data centers of 1 GW with liquid-cooled racks with densities over 100 kW.Say 7,000 current Nvidia racks. So they need 10,000 Giga-class satellites to match one 2025 data center. Lets say they start launching in 2027 and launch 100 satellites/year. They can match one 2025 data center in 2127. Maybe the AI bubble will have burst by then...
![]() |
| Source |
For longer duration missions, which would be the case with space based datacenters because they would be so expensive that they would have to fly for a long time in order to be economically viable, it's also necessary to consider total dose effects. Over time, the performance of chips in space degrades, because repeated particle impacts make the tiny field-effect transistors switch more slowly and turn on and off less completely. In practice, this causes maximum viable clock rates to decay over time, and for power consumption to increase.But lets ignore that and asume a useful life of 5 years. Thus to keep one 2025 data center operating in space they need to launch 2000 enormous satellites per year, or over 5 per day. It is true that solar power in space is free, but launches aren't. Adding the cost of 10,000 launches to the cost of replacing the 60% of the cost of the data center that is the racks is going to make the already impossible economics of AI orders of magnitude worse.
That is not to mention that they will be burning vast amounts of fuel and burning up more than 5 huge satellites a day in the upper atmosphere, which would cause vastly more pollution and damage to the ozone layer than described in Global 3D rocket launch and re-entry air pollutant and CO2 emissions at the onset of the megaconstellation era by Barker et al or Near-future rocket launches could slow ozone recovery by Revell et al.
![]() |
| Thiele et al Fig. 2 |
But even that isn't the biggest problem. In An Orbital House of Cards: Frequent Megaconstellation Close Conjunctions Thiele et al show that we are close to the Kessler syndrome:
Here we propose a new metric, the CRASH Clock, that measures such stress in terms of the time it takes for a catastrophic collision to occur if there are no collision avoidance manoeuvres or there is a severe loss in situational awareness. Our calculations show the CRASH Clock is currently 2.8 days, which suggests there is now little time to recover from a wide-spread disruptive event, such as a solar storm. This is in stark contrast to the pre-megaconstellation era: in 2018, the CRASH Clock was 121 days.Co-author Prof. Lawler wrote:
One of the scariest parts of this project was learning more about Starlink's orbital operations. I had always assumed they had some kind of clever configuration of the satellites in the orbital shell that minimized conjunctions, and we would see the number of conjunctions grow over time in our simulations. But no! It's just random! There's no magic here, it's just avoiding collisions by moving a Starlink satellite every 2 minutes. This is bad.Even if we haven't rendered low earth orbit unusable in the next few years, a constellation of 10,000 huge satellites in low earth orbit would rapidly self-destruct and guarantee humanity lost access to space.
What are these investors smoking?
Announcement of Strategic Funding to Enhance Multilingual, Sector-specific AI Literacy and Develop Trustworthy AI for Open Data / Open Knowledge Foundation
At the Open Knowledge Foundation (OKFN), we are excited and grateful to announce an important initiative for the next year, with the generous support of the Patrick J. McGovern Foundation (PJMF). With over two decades of transforming digital infrastructure to drive impact worldwide, we have identified two emerging problems over the last few years: The...
The post Announcement of Strategic Funding to Enhance Multilingual, Sector-specific AI Literacy and Develop Trustworthy AI for Open Data first appeared on Open Knowledge Blog.
Come Join the 2025 Winter Holiday Hunt / LibraryThing (Thingology)
It’s December, and we’re hosting our second annual Winter Holiday Hunt!
This hunt is meant to celebrate the season of light, and the holidays it brings. We wish all our members a Merry Christmas, Happy Hanukkah, and an entertaining hunt!
We’ve scattered a stand of evergreen trees around the site. You’ll solve the clues to find the trees and gather them all together.
- Decipher the clues and visit the corresponding LibraryThing pages to find some evergreen trees. Each clue points to a specific page on LibraryThing. Remember, they are not necessarily work pages!
- If there’s an evergreen tree on a page, you’ll see a banner at the top of the page.
- You have a little more than three weeks to find all the trees (until 11:59pm EST, Tuesday January 6th).
- Come brag about your stand of evergreen trees (and get hints) on Talk.
Win prizes:
- Any member who finds at least two evergreen trees will be
awarded an evergreen tree Badge (
). - Members who find all 15 evergreen trees will be entered into a drawing for one of five LibraryThing (or TinyCat) prizes. We’ll announce winners at the end of the hunt.
P.S. Thanks to conceptDawg for the European goldfinch illustration! ConceptDawg has made all of our treasure hunt graphics in the last couple of years. We like them, and hope you do, too!
Kafka becomes more accessible / John Mark Ockerbloom
Franz Kafka‘s work is now known around the world, but it couldn’t be read in English until after he died, and there’s still limited access to good English translations of much of his work. The English Kafka books I list are copyrighted translations generously shared online by David Wyllie and Ian Johnston.
Soon the first English Kafka books will enter the US public domain. The Castle, one of several Kafka works translated by Willa and Edwin Muir, gets there in 17 days. #PublicDomainDayCountdown
Author Interview: Loretta Ellsworth / LibraryThing (Thingology)

LibraryThing is pleased to sit down this month with Minnesota-based author Loretta Ellsworth, whose published work includes books for both juvenile and adult audiences. A former middle grade Spanish teacher, Ellsworth received her MFA in Writing for Children and Young Adults from Hamline University, and made her authorial debut in 2002 with the young adult novel The Shrouding Woman. She has had three additional young adult novels published, as well as a picture book for younger children, Tangle-Knot, in 2023. These books have won many accolades, including being named as ALA and IRA Notables, and being nominated for prizes such as the Rebecca Caudill Young Readers’ Award. Ellsworth published her first work for adults, the historical novel Stars Over Clear Lake, in 2017, followed in 2024 by The French Winemaker’s Daughter. Her third historical novel for adult readers, The Jilted Countess, which follows the story of a Hungarian countess who makes her way to Minnesota following World War II, in pursuit of her American GI fiancé, is due out from HarperCollins this coming January. Ellsworth sat down with Abigail this month to discuss the book.
The Jilted Countess was apparently inspired by a true story of a Hungarian countess who emigrated to Minnesota after the Second World War. Tell us a little bit about that original story. How did you discover it, and what made you feel you needed to retell it?
In 1948, a penniless Hungarian countess came to Minnesota to marry the GI fiancé she’d met abroad, only to find out he’d recently married someone else. Determined to stay in the U.S., she appealed to newspaperman Cedric Adams to help her find a husband before she’d be deported in two weeks back to Hungary, which was under Communist control. He agreed, using a fake name for her, and putting her picture in the newspaper, citing her circumstances. She received almost 1800 offers of marriage! And in two weeks she narrowed it down, went on a few dates, chose a husband, and was never heard from again. Fast forward to 2015, when someone found an old copy of that article in their attic and asked columnist Curt Brown if he knew what had happened to her. Curt Brown wrote a short article asking if anyone could provide an answer. Unfortunately, no one could. But that article made me wonder how a Hungarian countess could disappear like that, and I also wondered if she ever encountered her former fiancé again. She was, after all, the first Bachelorette, before the show was even a concept.
Did you do any kind of research, historical or cultural, in order to write the book? What were some of the most interesting things you learned?
I spent an exorbitant amount of time at the Minnesota History Center researching old microfiche articles to find anything I could about her. I examined marriage records for Minneapolis and St. Paul for any Hungarian-sounding names, and I searched for clues as to her whereabouts. Without a name, though, it was very difficult, and I never found her. I also had to research Hungary during and after the war, and the life of aristocrats, which I knew little about.
Contemporary readers might be surprised at the idea of a “Bachelorette” dating program taking place in the 1940s. How do you think Roza’s experience would tally with and differ from that of contemporary women seeking a spouse in this way?
After her marriage, she was approached by Look Magazine and other outlets for interviews, all of which she turned down as she wanted a private life. With social media today, there’s no way Roza would have been able to disappear like she did in 1948. And most likely her search would have taken place on social media rather than through the newspaper and mail.
World War II stories remain perennially popular with readers, despite the passage of the years. Why is that? What is it about this period that continues to speak to us?
I think it was such a pivotal time in the world, and one we’re still struggling to understand. And there are so many hidden stories that we’re constantly discovering about that time period that continue to speak to us. Also, the last of WWII veterans are disappearing, and their stories will be gone as well.
Tell us about your writing process. Do you write in a particular place, have a specific schedule you keep to, or any rituals that help you? Do you outline your stories, or discover them as you go along?
Because I worked as a teacher and had four children of my own, I had to learn to write in short intervals and adapt my writing schedule to be flexible. I wrote everywhere: at soccer practices and coffee shops and the library. Now that I no longer teach and my children are grown, I have a more disciplined schedule and usually write in the mornings in my home office, sometimes stretching into the afternoon. I also have learned to outline, whereas I used to write from the seat of my pants before. It’s helped to save me from a great deal of revision, although I still revise, just not as much as before.
What’s next for you? Will you be writing more historical novels for adults, or perhaps returning to the world of young adult books?
I am working on a young adult novel as well as another historical novel, so I hope to keep my foot in both genres as long as I’m able to. I enjoy both and read both.
Tell us about your library. What’s on your own shelves?
I have one full shelf of books on the craft of writing–I’m still drawn to how others write and am curious about their process. I have a mix of memoir, middle-grade, YA, and a lot of historical fiction. I still buy physical books, and my shelves are always overflowing. I donate a lot of books to our local Friends of the Library group for their annual book sale. And I have so many signed copies of books that I can’t part with. But that’s a good problem to have, isn’t it?
What have you been reading lately, and what would you recommend to other readers?
I read a great deal–I just finished reading the first two books of the Westfallen series by Ann and Ben Brashares with my grandson, and I’m reading The Correspondent by Virginia Evans, The Ivory City by Emily Bain Murphy, and The Gospel of Salome by Kaethe Schwehn. And I just finished James by Percival Everett. There are so many good books out there!
Letter from the Editors / Information Technology and Libraries
We announce the call for proposals for a future special issue on how Generative AI could transform our professional landscape, and summarizing the content of the December 2025 issue.
Neurodivergent People in the Library Workplace / Information Technology and Libraries
From the Field column for December issue
Relationality Over Neutrality / Information Technology and Libraries
Bridging the disciplinary gaps between library and information science and communication/cultural studies, this column utilizes LIS's interest in generative AI to frame a critical discussion regarding neutrality towards technology within librarianship.
AI-Infused Discovery Environments / Information Technology and Libraries
Although still in its infancy, artificial intelligence (AI) is rapidly making inroads into most facets of the library and education spheres. This paper outlines steps taken to examine Primo Research Assistant, an AI-infused discovery environment, for potential deployment at a large US public research university. The researchers aimed to evaluate the quality and relevance of the AI results in comparison to sources retrieved from the conventional search functionality, as well as the AI system’s multi-paragraph overview reply to the search query. As a starting point, the authors collected 103 search strings from a Primo Zero Result Searches report to approximate a corpus of natural language search queries. For the same research area, it was discovered that there was only limited overlap between the titles returned by the AI tool versus the current discovery layer. The researchers did not find appreciable differences in the numbers of topic-relevant sources between the AI and non-AI search products (Yes = 46.3% vs. Yes = 45.6%, respectively). The overview summary is largely helpful in terms of learning more details about the recommended sources, but it also sometimes misrepresents connections between the sources and the research topic. Given the overall conclusion that the AI system did not constitute a clear advancement or decline in effective information retrieval, the authors will turn to usability testing to aid them in further implementation decisions.
From Availability to Access / Information Technology and Libraries
This paper reports on student perspectives on access to online information resources when conducting an initial search for a school project. Through thematic analysis and user vignettes based on data from 175 students in elementary through graduate school, this paper explores how students determine whether they have access to online information resources, the barriers and enablers they attend to when pursuing access, and the characteristics that influence this process. Results reveal that resource previews, university and library branding, and the word download are generally viewed as enablers of access, while payment cues, learned heuristics around brands and formats, and the need to take extra steps to obtain the full text were barriers that often prevent students from trying to get access even when resources were available to them. Potential influences on individual capacity are also revealed, including experience in high- or low-availability information environments, ability to manage the complex cognitive load of determining access alongside other types of point-of-selection evaluation, a variety of dispositions related to information seeking, and situational factors related to the importance of the information need to the individual. While library staff work diligently to make online resources available, this does not automatically result in students’ ability to access those resources. This paper provides evidence to better equip library professionals for constructing their online information systems, collaborating with information providers about their online information systems, and teaching students about converting availability to access.

























