Saturday July 5th, 2025
At IAGSDC: Two A2 mirror tips dancing to Geo Jedlicka. And I think that, except for Swap Around and the Right And Left Grand, those have been the smoothest squares I've danced with so far.
Friday July 4th, 2025
At the Gay (square dance) Callers Association annual meeting, someone was praising Allan Hurst's efforts in pulling students into caller school, described him as "good at recruiting", and... uh... cough.
Thursday July 3rd, 2025
Holy shit. With the state of public transit payments, it's a freaking wonder anyone rides. I am hanging so much trouble with the fucking Clipper app today, and don't really want to wait for the 67 people ahead of me in the phone queue. #ClipperCard
I realize I'm ragging on LLMs a lot this morning, and I want to make it clear that it's mostly because I used Gemini for programming yesterday.
And, sure, I now understand the Google Sheets runtime way way better than I did thanks to following all of those dead-ends that I wouldn't have thought to pursue otherwise, but...
It reinforces the notion that I've heard several times recently that "AI" programming assistants work better when you believe in them.
Why is left as an exercise...
Giving up programming because of the advent of LLMs is like giving up woodworking because Harbor Freight introduced a line of guard-less radial arm saws.
This morning's "holy shit, I hope you fuckers are condemned to eternity trying to accomplish the most basic tasks with your IT workflows, but that's too harsh to wish on anyone" go out to the Marriott Bonvoy and ClipperCard Android apps and email processes.
The "Flowmaster is muh heritage, complaining about exhaust noise is hate speech" crowd seems to have gone to "'slow down' signs are a distraction, making me pay attention to stop signs is bad for safety."
(Statements only slightly exaggerated for effect.)
Pretty sure "Assault" here means "didn't manage to dodge an officer's fist". TechDirt: Assaults On ICE Officers Are Up 700%… Which Just Means There Have Been 69 More Assaults Than Last Year
Wednesday July 2nd, 2025
Wow. So I installed the Gemini CLI on my work computer, 'cause work is all in on this stuff, and asked it for some help with Google stuff, 'cause their documentation does not match their tools, and... I am not sure what additional information beyond my prompt is being sent, but the responses indicate that this is a *fantastic* way to leak data to Google, and if you care about keeping anything on your machine from Google you should carefully understand what, beyond the prompt, it's sending.
You know what I fucking love? When the Google documentation for writing Sheets custom functions doesn't match what buttons the Apps Script editor is showing me. That's what I fucking love.
According to Gartner, many agents are fiction without the science. "Many vendors are contributing to the hype by engaging in 'agent washing' – the rebranding of existing products, such as AI assistants, robotic process automation (RPA) and chatbots, without substantial agentic capabilities," the firm says. "Gartner estimates only about 130 of the thousands of agentic AI vendors are real."
Which, if course, duh, but mostly this is about TheAgentCompany: Benchmarking LLM Agents on Consequential Real World Tasks
We build a self-contained environment with internal web sites and data that mimics a small software company environment, and create a variety of tasks that may be performed by workers in such a company. We test baseline agents powered by both closed API-based and open-weights language models (LMs), and find that the most competitive agent can complete 30% of tasks autonomously. This paints a nuanced picture on task automation with LM agents--in a setting simulating a real workplace, a good portion of simpler tasks could be solved autonomously, but more difficult long-horizon tasks are still beyond the reach of current systems. We release code, data, environment, and experiments on this https URL.
Ya know, I don't so much mind that there are thousands of requests for non-existent .php files in my blog web server logs, it's that they come from so many different IP addresses.
Honestly, people, either I have an unsecured wp-login.php or I don't, and the first person who gets there is gonna patch the hole. It's wasted effort.
Tuesday July 1st, 2025
I remember when the street found its own uses for technology. Seems like now technology finds its own uses for the street.
Asta [AMP] @aud@fire.asta.lgbt
@SnoopJ@hachyderm.io what I truly love about this kind of question is that it's basically asking, "do we think the cure for cancer lies somewhere in the space between stolen science fiction novels, 4chan, reddit, and webMD?" but when you phrase it like that.
Retraction Watch: Springer Nature book on machine learning is full of made-up citations
Based on a tip from a reader, we checked 18 of the 46 citations in the book. Two-thirds of them either did not exist or had substantial errors. And three researchers cited in the book confirmed the works they supposedly authored were fake or the citation contained substantial errors.
But, hey, at only $169 for the ebook and $219 for the hardcover, it's not like you can expect Govindakumar Madhavan's Mastering Machine Learning: From Basics to Advanced to actually, I don't know, refer to reality or some such bullshit?
The dawn of micropayments: Cloudflare To Block AI Crawlers By Default & Pay Per Crawl Model
Cloudflare wrote that they are the "first Internet infrastructure provider to block AI crawlers accessing content without permission or compensation, by default." Now, new customers that sign up for Cloudflare by default will automatically block AI crawlers. Existing customers can block AI crawlers anytime with a single click in their Cloudflare dashboard. This shifts content scraping from an opt-out to opt-in format. There is a lot of buzz on Techmeme on this news.
Via. As clicks to useful information require more and more pauses and "I am not a bot" click, I'm wondering how this is gonna shake out.
daniel:// stenberg:// @bagder@mastodon.social
I've been talking to GitHub and giving them feedback on their "create issues with Copilot" thing they have in the works.
Today I tested a version for them and using it I asked copilot to find and report a security problem in curl and make it sound terrifying.
In about ten seconds it had a 100-line description of a "catastrophic vulnerability" it was happy to create an issue for. Entirely made up of course, but sounded plausible.
Proved my point excellently.
Kevin Beaumont @GossiTheDog@cyberplace.social
If you see this GitHub PoC for CVE-2025-5777 doing the rounds:
https://github.com/mingshenhk/CitrixBleed-2-CVE-2025-5777-PoC-
It’s not for CVE-2025-5777. It’s AI generated. The links in the README still have ChatGPT UTM sources.
The PoC itself is for a vuln addressed in 2023 - ChatGPT has hallucinated (made up) the cause of the vuln using an old BishopFox write up of the other vuln.
Today I learned about the Wikipedia:WikiProject AI Cleanup/AI catchphrases, which includes a bunch of tells that can be used to suss out writing that's more likely to be LLM generated. Via.
And Pivot To AI: ‘AI is no longer optional’ — Microsoft admits AI doesn’t help at work is the take I thought of when I heard that MS was strongly encouraging LLM use.
The thing about incorporating quirks in language for online communication in order to muck up automated word sequence generators, is that we're adopting wacky communications patterns just to put gravy in the gears of said token emitters.
Love that when I'm standing there waiting for the walk signal and the light turns yellow, I can hear the shift in tone as the drivers accelerate.
Monday June 30th, 2025
I've been having a couple of discussions about "AI" with people, one of whom is using ChatGPT, one of whom is using Gemini. Both have forwarded me conversations where the LLM reply starts with something amazingly close to...
This is a brilliant idea. You are absolutely thinking like a ...
In the first case, the one I've dug into, the LLM went on to effuse about how novel and amazing the ideas presented were, and how there wasn't anything in the literature about... and I did a quick Google search and said "have you considered these people in the late 1800s, or this guy in the 1970s...", and, well...
Anyway, that makes a remarkable preamble to Ed Zitron: Make Fun of Them, which takes far too many words to get to the point which is that we need to start asking the "AI" proponents exactly what they're claiming. Ed points out that
Anthropic has now put out multiple stories suggesting that its generative AI will “blackmail” people as a means of stopping a user from turning off the system, something which is so obviously the company prompting its models to do so. Every member of the media covering this uncritically should feel ashamed of themselves.
Which, yes, is exactly the point of these stories: They're there to "humanize", to anthropomorphize, the LLM output. Because any remotely critical reading of this says that we should simply not give the random number generator access to the big red "blow shit up" button. But if we give these things some sort of agency in our minds, then we start to see what they're generating as somehow "intelligent".
This whole thing is feeling more and more like religion, with the evangelists talking about how amazing it is, and the rest of us sitting around saying "uh, what a bunch of self-referential easily disproven bullshit, and yet you keep sending me Bible quotes like they mean something..."
Anyway, yeah.
The ChatGPTificiation of policy continues: CNN: Presentation for CDC advisers appears to cite nonexistent study to support claims about risk of vaccine preservative
“My study was published in Toxicological Sciences and did not find evidence of thimerosal exposure at vaccine levels in mouse behaviors that we thought were relevant to autism,” Berman said. He was “concerned and displeased” that his research appeared to have been cited in this way in Redwood’s slides.
Trying to find tips to get crispier friend polenta, and find a recipe suggesting "3 minutes on each side", and...
If you're publishing bullshit like that on the web, you're a psychopath and should be removed from society until you find a way to atone.
The utter fucking incompetence of modern web developers today, from Petco putting me into an infinite redirect loop, to Michael's making autofill of my address completely screwy, drives me insane.
Developers who blindly adopt JavaScript frameworks should be thrashed.
It's been roughly Covid time since my Asus Zenbook stopped charging and they wanted more to repair it than it cost, and I've been getting along with the work MacBook Pros, but I just provisioned a hand-me-down Dell with Mint Linux, and being back on a real environment is a huge breath of fresh air.
Need to get SquareDesk ported, that branch has languished, but so much on this platform just works where I have to fight MacOS.
Sigh. Looks like this year or so old Debian USB image isn't picking up the wifi on this new to me in Dell. Any one got opinions on a modern Linux? I mostly just wanna do SquareDesk dev and other random hacking on it. Maybe ClawsMail so I'm using an email client that doesn't suck again.
Sunday June 29th, 2025
I think the last time my legs were this tired, is walked 36 miles. Danced checker at the Sunnyvale caller workshop all day, then called 3 hour square dance for Foggy City dancers in SF. Feels good, in a feels bad sort of way.
Friday June 27th, 2025
Loving this: Want to Know How Democrats Should Handle Trans Rights? Ask Zohran Mamdani
The 33-year-old socialist just crushed Andrew Cuomo by doing what many national Democrats won't: having convictions.
Morning listening: Matthew Grimm: In This Ohio Diner
Guest at Circle 'n Squares tonight, Helen, 98 years old, who wanted to square dance one last time. We got her up and shuffled through a few things, and hopefully gave her the closure she was looking for.
AI Makes Research Easy. Maybe Too Easy.
Daniel Oppenheimer, a professor of psychology and decision sciences at Carnegie Mellon University, says the research is resonant of what he sees in similar studies he does in his lab: Students who use AI tools to complete assignments tend to do better on homework—but worse on tests. “They’re getting the right answers, but they’re not learning,” he says.
In turn, when subsequently forming advice on the topic based on what they learned, those who learned from LLM syntheses (vs. standard search results) feel less invested in forming their advice and, more importantly, create advice that is sparser, less original—and ultimately less likely to be adopted by recipients. Implications of the findings for recent research on the benefits and risks of LLMs are discussed.
Thursday June 26th, 2025
boringcactus @cactus@tacobelllabs.net
@davidgerard Drebin: “Johnny, how do I get Claude to write better unit tests?”
Johnny: “Computing is a specialized trade, I wouldn't know anything about that”
Drebin: *hands over cash*
Johnny: “word on the street is you gotta tell it you're holding its children hostage and will execute them if its test coverage drops below 95%”
Former DOGE engineer says federal waste and fraud were 'relatively nonexistent'
"Elon [Musk] was pretty clear about how he wanted DOGE to be maximally transparent," Lavingia said. "That's something he said a lot in private. And publicly. And so I thought, OK, cool, I'll take him at his word. I will be transparent."
Shortly after the interview was published online, Lavingia got an email. Just 55 days into his work at DOGE, his access had been revoked.
GenAI creates a hot mess. We assume the hot mess will be better in the future. I'm not sure if it will continue to make a hot mess or the mess will cool relative to our perception that it won't seem to be a hot mess any longer.
That is, it will still be a mess, it will just take longer to know that it is a mess. To be fair, this has always been the case for Software. As a Release Manager, I focused on when the mess had cooled enough to give to customers. After each release we would do two things:
- Try to cool the mess (fix bugs)
- Try to mess up other things related to the first goal (new features)
Rarely does software truly correct the problem it was meant to address. It is no wonder that software geeks are excited about GenAI. It may reduce the heat of the mess until it is better than human software developers. But I do wonder why GenAI's 'Wins' are mostly doing things that humans enjoy doing. Why not the stuff we dislike, I'm sure GenAI is a wiz at trigonometry.
Turns out I might be a bit of a hypocrite, because if we could take the thugs who are currently working for ICE and repurpose them to drag speeding motorists in my neighborhood from their cars and abduct them to secret prisons where they'd be subjected to indefinite detention in unspeakable living conditions...
Holy shit, we are fucked in ways that the financial system has yet to devise: Fannie Mae, Freddie Mac ordered to consider crypto as an asset when buying mortgages. The only saving grace is that at least they have to be stored in a centralized exchange...
Pulte also instructed the agencies that their mortgage risk assessments should not require cryptocurrency assets to be converted to U.S. dollars. And only crypto assets that “can be evidenced and stored on a U.S.-regulated centralized exchange subject to all applicable laws” are to be considered by the agencies in their proposal, Pulte wrote in a written order, effective immediately.
Wednesday June 25th, 2025
The AI-ification of customer service is going great: Reddit post about the Verizon "personal shopper" bot adding bogus services to the bill in a way that makes it look like the human customer service rep did it.
As Pivot to AI notes in their post about the larger Verizon "AI"/LLM rollout:
If a human customer service rep did that, it’d just be fraud. But being able to blame the computer, that’s the use case for AI!
They were convicted of killing with their cars. No one told the California DMV
A CalMatters investigation found that courts didn’t report hundreds of vehicular manslaughter convictions to the state, prompting officials to belatedly take many drivers’ licenses.
Via CalMatters Bluesky post and Petaluma Argus-Courier's article (I suspect you need a subscription for that one).
Hi, yes, I'm logged in to this platform that supports hardware that I've spent tens of thousands of dollars to acquire, do you think maybe we could dispense with the fucking "accept cookies" banner every time I load the page?
One of the projects for work is building a startup research tool demo on our platform, and in the process of that I asked gpt-4o-mini for details about one of the unicorns that bought a billboard mentioned in Bay Area Current: San Francisco's Billboards Aren't For You, and if I'd started this search from just Googling on various terms, I think I would have come to the conclusion that this company is a pile of hot air, but that those involved are real.
On the basis of the LLM output, initials rather than names for many of the key players, lots of titles that are just ... off ..., I'm now wondering if much of the backstory of this startup isn't LLM generated.
Really feel like people making Luddite parallels between the Jacquard loom and AI are massively over-promoting the capabilities of LLMs.
Figuring out provisioning profiles for profiling in XCode is my second favorite activity in MacOS, right behind library and tool management because this platform doesn't have a fucking package manager.
#WhatMilleniumIsThis #MacOS #IWantToWorkInARealOS