my.feedland.org Open in urlscan Pro
157.245.5.7  Public Scan

Submitted URL: https://www.info.evdemon.org/
Effective URL: http://my.feedland.org/Bazz
Submission: On June 25 via api from US — Scanned from DE

Form analysis 0 forms found in the DOM

Text Content

XML
Loosely Coupled Thinking
Stuff I'm Interested In

 * Architecture
 * Tech News
 * Development
 * Personal News

Today, 14:28Phil Windley's Technometria (Feed)
What Is Decentralized Identity?

Summary: What is decentralized identity and why is it important? My attempt at a
simple explanation.

In Yeah, yeah, yeah, yeah, yeah, nah, Alan Mayo references my recent blog post,
Decentralized Identity Comes of Age, and says:

My challenge to the decentralization community is for them (someone) to explain
how it works in relatively simple and reasonable terms. I say relative because
identity is not simple, so we should not expect simple solutions.

This post is my attempt to do that for Alan and others.

Identity is how we recognize, remember, react to, and interact with other
people, organizations, and services. Put another way, identity is about
relationships. Online we suffer from a proximity problem. Since we're not near
the parties we want to have relationships with, our natural means of
recognizing, remembering, and interacting with others can't be used. Digital
identity systems are meant to provide us with the means of creating online
relationships.

Traditional identity systems have not served us well because they are owned and
controlled by companies who build them for their own purposes. The relationships
they support are anemic and transactional. We can't use them for any purpose
except what their owner's allow.

Decentralized identity systems1 on the other hand allow you to create online
relationships with any person, organization, or service you choose and give you
the tools to manage and use those relationships. They help you recognize,
remember, react to, and interact with them. The most important tool is a
decentralized identity wallet. The world of decentralized identity wallets is
still young, but organizations like the Linux Foundation's Open Wallet
Foundation give me hope that useful, interoperable wallets are a tool we'll all
be able to use soon. They are as foundational to decentralized identity as a
browser is to the web.

Besides helping you manage peer-to-peer relationships with others online,
wallets hold verifiable credentials, the digital analog to the credentials and
cards you carry in a physical wallet. One of the most important aspects of
digital relationships is providing information about yourself to those you
interact with. Sometimes that information can come from you—it's
self-asserted—but many times the other party wants to reliably know what others
say about you. For example, if you establish a banking relationship, the bank is
legally obligated to verify things like your name and address independent of
what you say. Decentralized identity wallets allow you to prove things about
yourself using credentials others provide to you. At the same time, they protect
your privacy by limiting the information disclosed and forgoing the need for the
party you're interacting with to directly contact others to verify the
information you provide.

In summary, decentralized identity systems allow you to create digital
relationships with other parties independently, without relying on any other
organization or service. These relationships are direct, private, and secure.
They also provide the means for you to prove things about yourself inside these
relationships so that even though you're operating at a distance, you and the
other party can have confidence in the relationship's authenticity.

How Does It Work

The preceding paragraphs say what decentralized identity is, and provide its
benefits, but don't say how it works. Alan and others will likely want a few
more details. Everything I describe below is handled by the wallet. The person
using the wallet doesn't need to have any more knowledge of how they work than
the operator of a browser needs to understand HTTP and HTML.

The foundation of a peer-to-peer, decentralized online relationship is an
autonomic identifier like a peer DID. Identifiers are handles that someone else
can use to identify someone or something else online. Peer DIDs can be created
by a wallet at will, they're free, and they're self-certifying (i.e., there's no
need for a third party). A relationship is created when two identity wallets
create and exchange peer DIDs with each other on behalf of their owners. Peer
DIDs allow the parties to the relationship to exchange private, secure messages.

There are four primary interaction patterns that wallets undertake when
exchanging messages:

DID Authentication which uses the DIDs to allow each party to authenticate the
other Single-Party Credential Authorization where the same party issues and
verifies the credential. Multi-Party Authorization where the credential issuer
and verifier are different parties. Generalized Trustworthy Data Transfer which
uses a collection of credentials to aid the wallet owner in completing online
workflows. Generalized Credential Exchange Pattern (click to enlarge)

Verifiable credentials make heavy use of cryptography to provide not only
security and privacy, but also confidence that the credential data is authentic.
This confidence is based on four properties a properly designed credential
presentation protocol provides:

The identifier of the credential issuer Proof that the credential is being
presented by the party is was issued to Proof that the credential has not been
tampered with The revocation status of the credential

The credential presentation can do all this while only disclosing the
information needed for the interaction and without the verifier having to
contact the credential issuer. Not having to contact the issuer ensures the
credential can be used in situations with poor connectivity, that the issuer
needn't be online, and preserves the credential subject's privacy about where
the credential is being used.

A properly designed credential exchange protocol has four important properties:

The system is decentralized and contextual. There is no central authority for
all credentials. Every party can be an issuer, an owner, and a verifier. The
system can be adapted to any country, any industry, any community, any set of
credentials, any set of trust relationships. Issuers are free to determine what
credentials to issue and whether or not to revoke them. Wallet owners are free
to choose which credentials to carry and where and when they get shared. While
some verifiers require a specific credential—such as a customs agent requiring a
passport—others will accept a range of credentials. Therefore owners can decide
which credentials to carry in their wallet based on the verifiers with whom they
interact. Verifiers make their own decisions about which credentials to accept.
For example, a bar you are trying to enter may accept any credential you have
about your date of birth. This means some credentials (e.g., passports, driving
licenses, birth certificates) may be much more useful than just for the original
purpose for which they were issued.

These properties make a decentralized identity system self sovereign.

Why is Decentralized Identity Important?

Decentralized identity systems are designed to provide people with control,
security, and privacy while enhancing the confidence we have in our online
relationships. Some time ago, I wrote the following. I think it's an apt way to
close any discussion of decentralized identity because unless we keep our eyes
on the goal, we'll likely take shortcuts in implementation that fail to live up
to their promise.

Presently, people don't have operational relationships anywhere online.2 We have
plenty of online relationships, but they are not operational because we are
prevented from acting by their anemic natures. Our helplessness is the result of
the power imbalance that is inherent in bureaucratic relationships. The solution
to the anemic relationships created by administrative identity systems is to
provide people with the tools they need to operationalize their self-sovereign
authority and act as peers with others online. Peer-to-peer relationships are
the norm in the physical world. When we dine at a restaurant or shop at a store
in the physical world, we do not do so under the control of some administrative
system. Rather, we act as embodied agents and operationalize our relationships,
whether they be long-lived or nascent, by acting for ourselves. Any properly
designed decentralized identity system must provide people with the tools they
need to be "embodied" in the digital world and act autonomously.

Time and again, various people have tried to create decentralized marketplaces
or social networks only to fail to gain traction. These systems fail because
they are not based on a firm foundation that allows people to act in
relationships with sovereign authority in systems mediated through protocol
rather than by the whims of companies. We have a fine example of a protocol
mediated system in the internet, but we've failed to take up the daunting task
of building the same kind of system for identity. Consequently, when we act, we
do so without firm footing or sufficient leverage.

Ironically, the internet broke down the walled gardens of CompuServe and Prodigy
with a protocol-mediated metasystem, but surveillance capitalism has rebuilt
them on the web. No one could live an effective life in an amusement park.
Similarly, we cannot function as fully embodied agents in the digital sphere
within the administrative systems of surveillance capitalists, despite their
attractions. The emergence of self-sovereign identity, agreements on protocols,
and the creation of metasystems to operationalize them promises a digital world
where decentralized interactions create life-like online experiences. The richer
relationships that result from properly designed decentralized identity systems
promise an online future that gives people the opportunity to act for themselves
as autonomous human beings and supports their dignity so that they can live an
effective online life.

Notes I prefer the term self-sovereign to decentralized because it describes the
goal rather than the implementation, but I'll stick with decentralized here. All
self-sovereign identity systems are decentralized. Not all decentralized
identity systems are self-sovereign. The one exception I can think of to this is
email. People act through email all the time in ways that aren't intermediated
by their email provider. Again, it's a result of the architecture of email, set
up over four decades ago and the culture that architecture supports.

Photo Credit: Young Woman Using a Wallet from DALL-E (public domain) Prompt:
draw a rectangular picture of a young woman using a wallet.

Tags: identity ssi decentralized+identity verifiable+credentials


MORE
5 hours0
24.6.2024, 6:00SEI Blog (Feed)
Incorporating Agile Principles into Independent Verification and Validation
This post discusses how Agile principles can work with independent verification
and validation processes.

MORE
Yesterday at 5:31 PM0
Today, 17:55ongoing by Tim Bray (Feed)
Lounge Penguin

Lounge, as in a jazz club. Penguin, as in GoGo Pengin, a piano/bass/drums trio.
We caught their show at Jazz Alley in Seattle last week. Maybe you should go hit
a jazz lounge sometime.

What happened was

My daughter turned eighteen and graduated high school. She had heard that Car
Seat Headrest was playing Seattle’s Woodland Park Zoo, and could tickets and a
road trip (me buying and driving) be her present? Seemed reasonable, and she
found a friend to take along. I wouldn’t mind seeing the Headrests (decent indie
rock stuff) but her party, her friend. I noticed that GoGo Penguin was playing
Seattle’s Jazz Alley, and Lauren was agreeable to coming along for the ride and
the show.

I only know about GoGo Penguin because YouTube Music drops them into my default
stream now and them. I’d thought “sounds good, maybe a little abstract”,
couldn’t have named a song, but hey.

The “Jazz Club” concept

You’ve seen it in a million old movies, and the Vic Fontaine episodes of ST:DS9.
The lights are low, the audience is sitting at tables with little lamps on them,
the band’s on a thrust stage among the tables, there’s expected to be a soft
background of clinking glasses and conversation. Some people are focusing in
tight on the music, others are socializing at a respectfully low volume.

Of course, usually a gunfight breaks out or an alien materializes on stage… no
wait, that’s just on-screen not real-life.

All jazz clubs serve alcohol — fancy cocktails, natch — and many will sell you
dinner too. Dimitriou’s Jazz Alley in Seattle is a fine example.

GoGo Penguin at Jazz Alley; June 20th, 2024.

Our table was in the balcony.

We had a decent if conventional Pacific-Northwest dinner (crab and halibut),
with a good bottle of local white. They’ve got things set up so most people have
finished eating by the time the music starts. The seats were comfy. The decor
was pleasing. The service was impeccable. I felt very grown-up.

GoGo Penguin

They’re three youngish guys from Manchester. Their Web site says they’re an
“emotive, cinematic break-beat trio”. OK then. Piano/bass/drums is the canonical
minimal jazz ensemble. Only they’re not minimal and it’s not jazz. I guess if
you redefined “jazz” as complex rhythmically-sophisticated music featuring
virtuoso soloing skills, well yeah. Damn, those guys can play. But their music
is heavily composed, not a lot of opportunities for anyone to stretch out and
ride the groove.

And it ain’t got that swing; can it still mean a thing?

I guess so, because I enjoyed myself. Damn, those guys can play. There wasn’t a
microsecond that was boring, plus the arrangements were super intelligent and
kept surprising me.

But most of all, the bass. Nick Blacka hit me harder than any bassist since I
saw (and blogged!) Robbie Shakespeare of Sly and Robbie in 2004.

It’s really something special. It may be a stand-up acoustic bass, but it’s
wired up so he can dominate the band’s sound when he reaches back for it (which
he does neither too little nor too much). Plus the instrument’s acoustic texture
roars out entirely unmarred, you can feel those strings and wood in your gut. He
moves between bowing and plucking and banging and you hardly even notice because
it’s always the right thing.

I don’t wanna diss Chris Illingsworth on piano or Jon Scott on drums; both of
them made me catch my breath. But it’s Blacka’s bass explosions that I took home
with me.

That swing?

These days my musical obsessions are Americana (i.e. bluegrass with pretensions)
and old blues. The first of which also features instrumental complexity and
virtuosity. And, if I’m being honest, both offer a whole lot more soul than
Penguins.

I respect what the they’re doing. I’ll go see them again. But I wish they’d get
the hell out from behind those diamond-bright razor-sharp arrangements and just
get down sometimes.

Next?

Lauren and I had real fun and left feeling a bit guilty that we’ve been ignoring
Vancouver’s own jazz clubs. Not that I’m going to stop going to metal or
post-punk or baroque concerts. But jazz clubs are a good grown-up option.


MORE
Yesterday at 8:42 PM0
23.6.2024, 18:04The Architect Elevator (Feed)
Old Books that Every Architect Should Read
If architecture is timeless, so should be the books about it. Here's a book list
with some classics.

MORE
Sunday at 6:27 PM0
22.6.2024, 19:00Platformonomics (Feed)
Platformonomics TGIF #55: June 21, 2024

Platformonomics TGIF is a weekly roll-up of links, comments on those links, and
perhaps a little too much tugging on my favorite threads.

Get Platformonomics Updates By Email

Subscribe

A deja vu kind of week: the EU’s gotta EU, IBM’s gotta IBM, private equity’s
gotta private equity, Oracle’s gotta Oracle, and China’s gotta China. But we’re
unclear what it means if OpenAI’s gotta OpenAI.

News The European Union is STILL an Advanced Persistent Threat

Previous:

BREAKING: European Union Designated an Advanced Persistent Threat

Related:

New Branding, Same Scanning: “Upload Moderation”
Undermines End-to-End Encryption

Not Available in Europe

This is Europe’s future. GDPR cookie consents went to the EU’s head, making them
think they can dictate global practices, no matter how insane, useless, or
uninformed the regulations. But tech companies will increasingly just not ship
the latest and greatest to the EU, even if it means a non-global product line.
European versions of product should be referred to as the “wine and cheese”
editions.

Previous:

EU Suffering From Premature Regulation, BREAKING: European Union Designated an
Advanced Persistent Threat, EU Insanity: Regulatory Contortions, EU Insanity:
Not Learning From Regulatory Failure, EU Announces DMA Struggle Sessions, EU
Insanity: AI Regulatory Suicide, EU Insanity: AI Energy Suicide, EU Insanity:
The Kafka Singularity, EU Insanity: AI Energy Suicide (Part Deux), EU Insanity:
Mistral Edition, Move Fast and Regulate Things (You Don’t Understand), The EU
Will Continue to Fiddle While Rome Burns, When “Move Fast and Regulate Things”
Breaks Down, AI Regulation: Move Fast and Regulate Things

“Very Serious” Will be the Struggle Sessions

The EU’s “sweeping” and “landmark” rules are not an a priori set of regulations
to comply with, but a fishing license for regulators to meddle in any aspect of
(non-European) Big Tech business that catches their fancy. Rule of Law it isn’t.

These struggle sessions will continue until Europe has a tech industry (i.e.
don’t hold your breath) even as the beleaguered continent ignores its
existential economic, energy and geopolitical challenges.

Previous:

Apple Qualifies for DMA Struggle Session, UK Announces Big Tech Struggle
Sessions “For The Children”, EU Announces DMA Struggle Sessions, EU Announces
Meta Struggle Sessions,

Existential Corner: What is OpenAI This Week?

It remains unclear what OpenAI wants to be if/when it grows up. A technology
provider (frontier models)? A consumer services company (but emphatically not
one inspired by Her)? An enterprise company? That seems the most difficult, yet:

Previous:

Existential Corner: What is OpenAI?

IBM is Not a Technology Company: McDonald’s Edition

Previous:

This Week in Securities Fraud, IBM and the Art of Misleading Investors, Last IBM
Strategic Imperative Bites the Dust, IBM’s “Cloud” Business (or Lack Thereof),
Tweetstorm Digest: Reactions to Barron’s “IBM’s Reboot” Story, IBM’s Lost
Decade, Follow the CAPEX: The Clown Car Race Checkered Flag, Introducing
Platformonomics ClownWatch,

Related:

McDonald’s enters strategic partnership with IBM to automate drive-thru lanes,
Big Blue scoffs a Happy Meal: McDonald’s sells automated order-taking tech to
IBM

Private Equity in Action: OtterTune

Whatever private equity is up to, customers always seem to suffer.

Previous:

Private Equity in Action: PluralSight,
Private Equity in Action: Not Every VMware Customer Was Screwed, Private Equity
in Action: The VMware Saga May Never End, Software Migration Alerts, Private
Equity in Action: VMware “Unease” Continues, Perfidious Private Equity, Private
Equity in Action: VMware Yet Again, Private Equity in Action: VMware
(Continued), Private Equity in Action: VMware, Private Equity Will Buy and Ruin
More Software Companies, Software Migration Alert: VMware (Continued), Software
Migration Alert: VMware, Private Equity in Action: VMware Customer Complaints
“Unwarranted”

Oracle Invents “Non-GAAP CAPEX”

Because they know their CAPEX is woefully insufficient, Oracle has claimed in
earnings calls that their CAPEX spend was 25% bigger than the number they
reported to the SEC in the last two quarters. ¯\_(ツ)_/¯

Previous:

Oracle Still Can’t Build Data Centers, Why Can’t Oracle Build Data Centers?,
Follow the CAPEX: The Clown Car Race Checkered Flag

China Up to No Good: Prepare to Repel Boarders Edition

What is the “nothing to worry about here” explanation for China’s actions off
the Second Thomas Shoal?

Previous:

One, Two, Three, Four: We Want a Meme War, Xi Jinping Thought on Wall Street
Running Dogs, The Perils of Working for a Cult of Personality, Press Releases
We’d Like to See, Best Nickname for Emmanuel Macron, Winnie-the-Pooh: Lover of
Honey, Champion of Freedom, Missing the (Bamboo) Forest for the (Apple) Trees

Related:

Full Philippine Navy Thread, “The Sierra Madre has become the most dangerous
flashpoint in the Indo-Pacific”

China Up to No Good: Cyberwar Retaliation Edition

Rahm Emanual, US ambassador to Japan, is leading the charge in the meme war with
the Chinese Communist Party.

Previous:

One, Two, Three, Four: We Want a Meme War, Xi Jinping Thought on Wall Street
Running Dogs, The Perils of Working for a Cult of Personality, Press Releases
We’d Like to See, Best Nickname for Emmanuel Macron, Winnie-the-Pooh: Lover of
Honey, Champion of Freedom, Missing the (Bamboo) Forest for the (Apple) Trees


MORE
Friday at 8:58 PM0
23.6.2024, 18:04The Architect Elevator (Feed)
The Many Facets of Coupling
Dissecting a buzzword takes more than a single blog post.

MORE
Friday at 5:26 PM0
18.6.2024, 16:30All Things Distributed (Feed)
Introducing Distill CLI: An efficient, Rust-powered tool for media summarization
After a few code reviews from Rustaceans at Amazon and a bit of polishing, I'm
ready to share the Distill CLI. An open-source tool written in Rust for
summarizing meetings and other media that uses AWS services, such as S3,
Transcribe, and Bedrock.

MORE
June 180
Today, 17:55ongoing by Tim Bray (Feed)
Epsilon Love

Quamina was for a time my favorite among all my software contributions. But then
it stalled after I shipped 1.0 in January of 2023. First of all, I got busy with
the expert witness for Uncle Sam gig and second, there was a horrible problem in
there that I couldn’t fix. Except for now I have! And I haven’t done much
codeblogging recently. So, here are notes on nondeterministic finite automata,
epsilon transitions, Ken Thompson, Golang generics, and prettyprinting. If some
subset of those things interests you, you’ll probably like this.

(Warning: if you’ve already had your hands on the theory and practice of finite
automata, this may all be old hat.)

[Update: This is kind of embarrassing. It looks like what this post refers to as
an “epsilon” is not the same epsilon that features in the theory of finite
automata. I mean, it still works well for where I’m using it, but I obviously
need dig in harder and deeper.]

Sidebar: What’s a Quamina?

I don’t think there’s much to be gained by duplicating Quamina’s README but in
brief: “A fast pattern-matching library in Go with a large and growing pattern
vocabulary and no dependencies outside Go’s standard libraries.” If you want
much, much more, this Quamina Diary blog series has it.

The problem

Combining too many patterns with wild-cards in them caused Quamina 1.0’s data
structures to explode in size with a growth rate not far off the terrifying
O(2N), which meant that once you’d added much more than 20 patterns you couldn’t
add any more, because the add-pattern code’s runtime was O(2N) too.

Those structures are state machines generally, “nondeterministic finite
automata” (NFA’s) in particular. Which offer good solutions to many software
problems, but when they get to be any size at all, are really hard to fit into a
human mind. So when I was looking at Quamina’s unreasonably-big automata and
trying to figure out how they got that way, my brain was screaming “Stop the
pain!”

Lesson: Prettyprint!

At the point I stalled on Quamina, I’d started a refactor based on the theory
that the NFAs were huge because of a failure to deduplicate state transitions.
But the code I’d written based on that theory was utterly broken; it failed
simple unit tests and I couldn’t see why.

During the months when I was ignoring the problem, I privately despaired because
I wasn’t sure I could ever crack it, and I couldn’t stomach more struggling with
ad-hoc Printf and debugger output. So I decided to generate human-readable
renditions of my automata. Given that, if I still couldn’t figure out what was
going on, I’d have to admit I wasn’t smart enough for this shit and walk away
from the problem.

Which turned out to be a good call. Generating an information-dense but readable
display was hard, and I decided to be ruthless about getting the spaces and
punctuation in the right places. Because I didn’t want to walk away.

Back in the day, we used to call this “prettyprinting”.

It worked! First of all, my prettyprinter showed me that the automata emitted
based on my deduplication theory were just wrong, and what was wrong about them,
and I found that code and fixed it.

Bad news: My deduplication theory was also just wrong. Good news: My
prettyprinter provided unavoidable proof of the wrongness and made me go back to
first principles.

And I just landed a PR that cleanly removed the state explosion.

Free advice

I’ll show off the prettyprinter output below where I dig into the
state-explosion fix. But for the moment, a recommendation: If you have a data
structure that’s not Working As Intended and is hard to grok, go hide for a
couple of days and write yourself a prettyprinter. Prettyprinting is an
intelligence amplifier. Your Future Self will thank you heartily.

“Back to first principles”?

The single best write-up on NFA and regex basics that I’ve ever encountered is
Regular Expression Matching Can Be Simple And Fast (but is slow in Java, Perl,
PHP, Python, Ruby, ...) by Russ Cox. It’s a discussion of, and reflection on,
the regular expression library constructed by Ken Thompson in the mid-Sixties,
before he got mixed up in Unix.

What’s annoying is that I had read this before I started wiring NFAs into
Quamina, but ignored most of its important lessons due to a combination of not
understanding them and thinking that my existing code could do what Cox
described. A couple of weeks ago I went back and read it again, and it all made
perfect sense and showed me the way forward. So I guess the lesson is that if
you’re not Ken Thompson, you’re going to have trouble understanding what he did
until you’ve tried and failed yourself?

So, major thanks to Ken for this (and Unix and other things too) and to Russ for
the write-up.

Epsilon transitions

These are the magic bullet that make NFA’s work. Quamina didn’t have them, now
it does. There are other bits and pieces but that’s the core of the thing.

I think the easiest way to explain is by showing you an NFA as displayed by
Quamina’s new prettyprinter. It matches the regular expression "x.*9" — note
that the " delimiters are part of the pattern:

758[START HERE] '"' → 910[on "] 910[on "] 'x' → 821[gS] 821[gS] ε → 821[gS] /
'9' → 551[gX on 9] 551[gX on 9] '"' → 937[on "] 937[on "] 'ℵ' → 820[last step]
820[last step] [1 transition(s)]

There’s an API to attach labels to states as you build automata, which as a
side-effect gives each a random 3-digit number too. This is done in a way that
can be turned into a no-op at production time.

758: The start state; the only character that does anything is the opening "
delimiter which transitions to state 910.

910: You get here when you see the " and the only exit is if you see an x, which
moves to 821.

821: This state is the “glob” * operator. gS in its label stands for “glob
spin”. It has an "epsilon" (ε) transition to itself. In Computer-Science theory,
they claim that the epsilon transition can occur at any time, spontaneously,
la-di-da. In programming practice, you take an epsilon transition for every
input character. 821 also has an ordinary transition on 9 to state 551.

This possibility of having multiple transitions out of a state on the same input
symbol, and the existence of epsilon transitions, are the defining
characteristics that make NFAs “nondeterministic”.

551: Its label includes gX for “glob exit”. The only transition is on the
closing " delimiter, to 937.

937 has only one transition, on ℵ (stands for the reserved value Quamina inserts
to signal the end of input) to 820.

820 doesn’t do anything, but the [1 transition(s)] label means that if you reach
here you’ve matched this field’s value and can transition to working on the next
field.

Now I’m going to display the prettyprint again so you can look at it as you read
the next paragraph.

758[START HERE] '"' → 910[on "] 910[on "] 'x' → 821[gS] 821[gS] ε → 821[gS] /
'9' → 551[gX on 9] 551[gX on 9] '"' → 937[on "] 937[on "] 'ℵ' → 820[last step]
820[last step] [1 transition(s)]

A little thought shows how the epsilon-transition magic works. Suppose the input
string is "xyz909". The code will match the leading " then x and hit state 821.
When it sees y and z, the only thing that happens is that the epsilon transition
loops back to 821 every time. When it hits the first 9, it’ll advance to 551 but
than stall out because the following character is 0 which doesn’t match the only
path forward through ". But the epsilon transition keeps looping and when the
second 9 comes along it’ll proceed smoothly through 551, 937, and 820, signaling
a match. Yay!

So now, I have a fuzz test which adds a pattern for each of about thirteen
thousand 5-letter words, with one * embedded in each at a random offset,
including the leading and trailing positions. The add-pattern code hardly slows
down at all. The matching code slows down a lot, to below 10,000/second, in
stark contrast to most Quamina instances, which can achieve millions of
matches/second.

I’m sort of OK with this trade-off; after all, it’s matching 10K-plus patterns!
I’m going to work on optimizing it, but I have to accept that the math, as in
finite-automata theory, might be against me. But almost certainly there are some
optimizations to be had. There are possibilities suggested by Cox’s description
of Thompson’s methods. And the search for paths forward will likely be good blog
fodder. Yay!

Ken again

When I re-read Russ Cox’s piece, I was looking at the pictures and narrative,
mostly ignoring the C code. When everything was working, I went back and was
irrationally thrilled that my bottom-level function for one state traversal had
the same name as Ken Thompson’s: step().

Also, when you process an NFA, you can be in multiple states at once; see the
"xyz909" example above. When you’re in multiple states and you process an input
symbol, you might end up in zero, one, or many new states. Russ writes, of Ken
Thompson’s code, “To avoid allocating on every iteration of the loop, match uses
two preallocated lists l1 and l2 as clist and nlist, swapping the two after each
step.”

Me too! Only mine are called currentStates and nextStates because it’s 2024.

And thereby hangs a blog or maybe more than one. Because traversing the NFA is
at Quamina’s white-hot center. You really REALLY don’t want to be allocating
memory in that code path. Which should be straightforward. But it’s not, for
interesting reasons that raise optimization problems I’m just starting to think
about, but you’ll probably hear all about it when I do.

Un-generic

In the process of moving Quamina from DFAs to mixed DFA/NFA to pure-NFA I
adopted and then abandoned Go’s generics. They hate me. Or I’m not smart enough.
Or something. I wrote about the experience back in 2022 and while that piece
ended inconclusively, I am personally much happier with generics-free Go code.
Maybe they make other people happy.

Hard to understand

And then finally, there’s this one function I wrote in June 2022, doesn’t matter
what it does. It has a a comment at the top that begins: “Spookeh. The idea is
that…” and goes on for a long paragraph which, well, I can’t understand. Then I
look at the code and think “that can’t work.” I keep thinking of sequences that
should send it off the rails and write the unit tests and they fail to fail, and
I use the prettyprinter and the NFA it generates is ruthlessly correct. I go
back and look at it every few days and end up shaking my head. This is making me
grumpy.

But after all, I did write, in a previous Quamina Diary episode: “The
observation that computer programmers can build executable abstractions that
work but they then have trouble understanding is not new and not surprising.
Lots of our code is smarter than we are.”

But I’ll figure it out. And it’s nice to have interesting computer-programming
stuff to blog about.


MORE
Friday at 7:12 AM0
24.6.2024, 6:00SEI Blog (Feed)
Software Cost Estimation Explained
Anandi Hira explains software cost estimation, presents estimation tools, and
examines inherent biases that exist in software cost estimation models.

MORE
Tuesday at 10:40 PM0
Today, 17:55ongoing by Tim Bray (Feed)
Wikipedia Pain

There are voices — some loud and well-respected — who argue that Wikipedia is
deeply flawed, a hellscape of psychotic editors and contempt for expertise. I
mostly disagree, but those voices deserve, at least, to be heard.

[Note: There’s a companion blog post, Sex Edit War!, about my own experience in
a Wikipedia Edit War. (I won! It was fun!) I hope it’ll make some of this
narrative more concrete.]

Background

If you look at this post’s Reference Publishing topic, you’ll see a lot of
Wikipedia-related material. I was one of its early defenders against the
early-days waves of attackers who compared it to a public toilet and its editors
to the Khmer Rouge.

I should also disclose that, over the years, I have made some 2,300 Wikipedia
edits, created seven articles, and (what makes me happiest) contributed 49
images which have been used, in aggregate, 228 times.

I say all this to acknowledge that I am probably predisposed to defend
Wikipedia.

What happened was…

Somebody spoke up on the Fediverse, saying “I wonder if reporters know that
Wikipedia hallucinates too??” I’m not giving that a link, since they followed up
with a post asserting that ChatGPT is better than Wikipedia. Life’s too short
for that.

Anyhow, I replied “The difference is, errors in Wikipedia tend to get
systematically fixed. Sometimes it takes more work than it should, but the vast
majority of articles are moving in the right direction a vast majority of the
time.” Much discussion ensued; follow the threads.

Shortly thereafter, the redoubtable JWZ complained about an edit to his page and
I spoke up noting that the edit had been reversed, as bad edits (in my
experience) usually are. That conversation branched out vigorously, dozens of
contributions. Feel free to trawl through the Fediverse threads, but you don’t
have to, I’ll summarize.

Gripe: Bad editors

This kept coming back.

Jamie Zawinski: I just find the culture of the editors intolerable.

Ruben Schade: I didn’t want to draw the ire of those horrible admins.

Dave Slusher: The editor culture demonstrably contains decision makers who ain't
that bright and are pretty lazy.

I dunno. I don’t want to gaslight those people; if that’s the experience they
had, that’s the experience they had. My own experience is different: The editors
I’ve interacted with have generally been friendly and supportive, and often
exceptionally skilled at digging up quality citations. But I think that these
reports are something Wikipedia should worry about.

Gripe: Disrespect of expertise

By number and volume of complaints, this was the #1 issue that came up in those
threads:

Brian Dear: Pantheon published my book in 2017. For years I tried to add to the
paltry Wikipedia entries for everything from PLATO, plasma panels, to early
MUDs, instant msging, but no: Wikipedia editors deleted it all (“not
authoritative”). Screw ‘em.

Dan O’Neill: An editor argued with me about the history of my own company.

Zawinski: Wikipedia specifically rejects domain experts until someone who knows
nothing cites them elsewhere.

Alex Rosenberg: So many mistakes in early pages about PS3 that I tried to get
corrected as a primary source, but some random kiddie blogger's guess as to what
an acronym stood for was more valuable to the editors.

I generally disagree with these takes. Wikipedia not only respects but requires
expert support for its content. However, it uses a very specific definition of
“expert”: Someone who can get their assertions published in one or more Reliable
Sources.

I think that if you’re about to have an opinion about Wikipedia and expertise
and citations, you should give that Reliable-Sources article a careful read
first. Here’s why: It is at the white-hot center of any conversation about what
Wikipedia should and should not say. Since Wikipedia is commonly the top result
for a Web search, and since a couple of generations of students have been taught
to consult but not cite it, the article is central to what literate people
consider to be true.

Let’s consider the complaints above. Mr Dear literally Wrote the Book. But, I
dunno. I went and looked at the PLATO article and subjects linked to it, and,
well, it looks good to me? It cites Mr Dear’s book but just once. Maybe the
editors didn’t think Mr Dear’s book was very good? Maybe Dear says controversial
things that you wouldn’t want to publish without independent evidence? The
picture is inconclusive.

As for Mr O’Neill’s complaint, no sympathy. Given the social structure of
capitalism, the employees and leadership of a company are the last people who
should be considered Reliable Sources on that company. Particularly on anything
that’s remotely controversial.

Mr Zawinski is upset that the person who chooses citations from Reliable Sources
“knows nothing”, which I take to be an abbreviation for “is not a subject-matter
expert”. There’s some truth here.

When it comes to bald statements of fact, you don’t need to be an expert; If
more than one quality magazine or academic journal says that the company was
incorporated in 1989, you don’t need to know anything about the company or its
products to allow “founded in 1989” into an article.

On the other hand, I think we can all agree that people who make significant
changes on articles concerning complex subjects should know the turf. My
impression is that, for academic subjects, that condition is generally met.

Mr Rosenberg, once again, is upset that his personal expertise about the PS3 is
being disregarded in favor of material sourced from a gamer blog. I’d have to
know the details, but the best possible outcome would be Mr Rosenberg
establishing his expertise by publishing his narrative in a Reliable Source.

Bad Pattern

There’s a pattern I’ve seen a few times where a person sees something in
Wikipedia in an area where they think they’re knowledgeable and think it’s wrong
and decide “I’ll just fix that.” Then their edits get bounced because they don’t
include citations. Even though they’re an “expert”. Then that person stomps away
fuming publicly that Wikipedia is crap. That’s unfortunate, and maybe Wikipedia
should change its tag-line from “anyone can edit” to “anyone who’s willing to
provide citations can edit.”

Implications

This policy concerning expertise has some consequences:

The decision on who is and isn’t an expert is by and large outsourced to the
editorial staff of Reliable Sources.

There are ferocious debates among editors about which sources are Reliable and
which are not, in the context of some specific article. Which is perfectly
appropriate and necessary. For example, last time I checked, Fox News is
considered entirely Reliable on the finer points of NFL football, but not at all
on US politics.

There are many things which people know to be true but aren’t in Wikipedia and
likely never will be, because no Reliable Source has ever discussed the matter.
For example, I created the East Van Cross article, and subsequently learned the
story of the cross’s origin. I found it entirely convincing but it was from an
guy I met at a friend’s party who was a student at the high school where and
when the graphic was first dreamed up. I looked around but found no Reliable
Sources saying anything on the subject. I doubt it’ll ever be in Wikipedia.

What do you think of those trade-offs? I think they’re pretty well OK.

The notion that anyone should be allowed to add uncited assertions to Wikipedia
because they think they’re an expert strikes me as simultaneously ridiculous and
dangerous.

Real problems

Obviously, Wikipedia isn’t perfect. There are two problems in particular that
bother me all the time, one small, one big.

Small first: The editor culture is a thicket of acronyms and it’s hard to keep
them straight. I have considered, in some future not-too-fierce editorial
debate, saying “Wait, WP:Potrezebie says you can’t say that!” Then see if anyone
calls me on it.

The big problem: The community of editors is heavily male-dominated, and there
have repeatedly been credible accusations of misogyny. I have direct experience:
I created the article for Sarah Smarsh, because we read her excellent book
Heartland in my book club, then I was shocked to find no entry. Despite the
existence of that mainstream-published and well-reviewed book, and the fact that
she had published in The Guardian and the Columbia Journalism Review, some other
editor decreed that that was insufficient notability.

At the time, I reacted by gradually accumulating more and more citations and
updating the draft. Eventually she published another book and the argument was
over. These days, in that situation I would raise holy hell and escalate the
obstruction up the Wikipedia stack.

To Wikipedia’s credit, its leadership knows about this problem and gives the
appearance of trying to improve it. I don’t know the details of what they’re
trying and whether they’re moving the needle at all. But it’s clearly still a
problem.

Once again…

I stand by what I said in December 2004: Wikipedia dwarfs its critics.


MORE
June 160
Sex Edit War!

In January 2010 I drove twenty-five minutes across Vancouver to the University
of British Columbia’s main library, with the goal of crushing an opponent in a
Wikipedia edit war. The battleground was the entry on T.E. Lawrence (better
known as Lawrence of Arabia). I won that war. As a consequence, I consider
myself the world’s leading living expert on Lawrence’s sexuality.

[Note: This is posted alongside Wikipedia Pain, which is about the issues of
truth and expertise in Wikipedia editing, in an effort to share what the process
feels like from the inside.]

Why Lawrence, anyhow? My Dad, an Alberta farm boy, became a Professor of
Agriculture, and spent most of his career in the Third World, much of it in
Lebanon and Jordan. As a result, I spent my youth there, with plentiful
opportunities for touristing all over the Middle East, including many of the
spots that appear in Lawrence’s monumental war memoir Seven Pillars of Wisdom.

I ran across Seven Pillars in college and devoured it, from time to time
thinking “I’ve been there!” While it’s full of camel charges, train-bombings,
and other Ripping Yarns, it’s a difficult book, not a light read at all. But I
enjoyed it and was left wondering who this guy was. So in the course of time I
read Lawrence’s other works, some biographies (there are many) and especially,
the collected letters.

Lawrence was an avid correspondent, sending letters almost like we do emails,
multiple times most days. I suspect that a whole lot of the Lawrence biographers
got the idea by reading the letters and like me thinking “who is this guy?” You
might want to do a little Lawrence reading.

Conducting archeology on my blog reveals that I apparently noticed Wikipedia in
2003 and had started contributing to the Lawrence article by 2004; in that year
I also wrote “Maybe the Wikipedia is a short-lived fad, maybe it’ll get better,
maybe it’ll get worse, but I was surprised that nobody pointed this out: The
Wikipedia is beautiful. It’s an unexpected and unexplainable triumph of
collective creativity and of order over entropy. I hope it lasts a long time,
and those who criticize it Just Don’t Get It.”

At that time popular opinions of The Encyclopedia That Anyone Can Edit ranged
from a headshaking blow-off of the idea’s obvious craziness to active fear and
hostility. British technology journalist Andrew Orlowski once referred to
Wikipedians as “Khmer Rouge in daipers” (sic). I became a partisan, wading into
the ring against figures as eminent as Bob McHenry, former Editor of the
Britannica, who compared Wikipedia to a public toilet: “you can’t be sure who
was there before you.” I enjoyed rolling out my rhetorical and polemical cannon
and firing back. From December 2004: “One thing is sure: the Wikipedia dwarfs
its critics.”

It must be said that back then, the critics had a point. Those of us who waded
in early often found entries about major subjects of history or culture which
were a stinking mess. Lawrence was one such; a farrago of conspiracy theories
and thinly-sourced fantasies.

Sex!

In particular the section about Lawrence’s sexuality, a subject much discussed
by his biographers and occasionally in the popular press. The amount of time
I’ve put into making this fact-based would probably be regarded as ridiculous by
most sane people. [Would they be wrong? -Ed.] [Pretty sure. -T.]

I have plenty of by-and-about-Lawrence books on my shelves and had read more or
less every published letter, which I thought gave me a fair claim to knowing him
better as a person than your average Wikipedia editor. By dint of dogged
incremental citation-backed edits, I was making good progress by 2009 at
introducing order to the chaos.

Edit!

Editing Wikipedia involves regular, often intense, disputes about what should be
said. These take place on the “Talk” page that is associated with each article.
For a contentious entry, such as Lawrence’s had become, the Talk page can become
huge, much larger than the entry itself.

In these disputes, the criteria that matter are “notability” and
“verifiability”. To be included, a subject must be notable, i.e. worth
mentioning. When is something notable? If, and only if, there are mentions of
the subject in multiple credible mainstream sources. Further, any assertion must
be verifiable, i.e. there is evidence to establish that the claims in the
material are correct. Both criteria are addressed by providing citations from
Reliable Sources.

On the subject of verifiability, Wikipedia says to the world: Any material that
is not verifiable will eventually be removed. That tenet gives a warm glow to
those of us who live on the Internet and care a lot about truth and untruth.

The subject at hand was homosexuality. First, had Lawrence been gay? Second,
what was his attitude toward gay people? Remember, this is a man who died in
1935; in his lifetime, homosexuality was publicly much disapproved-of and in
fact specifically forbidden by law.

I thought I had the facts on my side. Whatever Lawrence’s orientation, there was
no evidence of consensual intimacy with anyone of any gender, and he repeatedly
and explicitly denied, in private correspondence, any experience of sex.

On the other hand, his writing includes multiple warm, approving remarks about
male/male sexual relationships. So I thought the case for “celibate and
tolerant” was pretty well open and shut.

War!

But then I found I had an adversary.

“Factuarius” – the handle of another active Wikipedia editor – came to fight.
For reasons opaque to me, Factuarius was pretty convinced that Lawrence had been
gay and/or disapproved of homosexuality. He was able to assemble citations where
people had alleged relationships between Lawrence and one or another male
person, but this was well-plowed ground; biographers had found an absence of
evidence for the relationships and reasonably convincing reasons to doubt their
having happened.

Factuarius decided that Lawrence’s having disapproved of homosexuality was the
hill he was going to die on. He triumphantly produced two citations that
supported his position, declared victory, and told me to stand down.

The first was “Khondakar Golam Mowla, 2008 p. 258”. The book is The Judgment
Against Imperialism, Fascism and Racism Against Caliphate and Islam: Volume 1.
You can buy it from Amazon for $36.49 as I write this. It turns out it is
self-published at “AuthorHouse” and that its Foreword denounces, among other
things, “Ataturk, a secret Jew”. The tone generally follows from there. I
pointed out to Factuarius that I could go to AuthorHouse and generate a book
claiming Lawrence was from Mars.

That left him hotly defending his last reference, a Lawrence letter cited in
“Homosexuality and Orientalism: Edward Carpenter's journey to the east, P. K.
Bakshi, Prose Studies, Volume 13, Issue 1 May 1990, pages 151-177, Routledge”.
Seeing no alternative, I made that drive over to the nearest big University
research library.

It took a while to track down Prose Studies, whose dusty and clearly-unvisited
volumes occupy quite a few shelf-feet. It was founded in 1977 and the Internet
tells me it’s still publishing. I really don’t know what this journal is for or
what effect on the world, if any, its existence is designed to achieve.
[Arrogant, much? -Ed.] [Trying to be polite. -T.]

Sure enough, the article about Edward Carpenter was there in the May 1990
volume. I read it. I photographed (badly, with a 2010 phone-cam) the title and
index pages to prove that I had done so. The article mentioned Lawrence twice,
suggesting in an off-handed way that he was an example of English fascination
with homosexuality and “the Orient”. But there was nothing there that looked
like Factuarius’ citation.

Victory!

I was left happy for multiple reasons. It is a wonderful thing that research
libraries exist and preserve academic journals for their own sake, whether or
not any human will ever consult their pages. It was pretty cool playing
scholarly sleuth in the quiet passages of the library. Best of all, Factuarius
retired silently from the fray.

Which was actually a pretty minor scuffle by Wikipedia standards. There is a
hilarious page entitled Wikipedia:Lamest edit wars, which I recommend just for
fun. It even categorizes them. The first-appearing category is “Ethnic and
national feuds”, featuring the titanic struggles over the ancestries of Frédéric
Chopin and Freddie Mercury. So far, none of these has metamorphosed into a real
actual nation-against-nation shooting war, but I’m not saying it couldn’t
happen.

Eventually I took the trouble of collecting every citable fact about Lawrence’s
sexuality that I could find in all the known published resources – online search
in the Gutenberg Project and various other sources helped. I published them in a
blog piece entitled Sex and T.E. Lawrence, which has been useful in subsequent
(much less dramatic) editing disagreements.

Finally, I gave a talk at a social-media conference sometime in the 2000s
entitled Editing Wikipedia in which I had great fun relating this episode, and I
think the audience did too. In particular, reading out spicy passages
illustrating Lawrence’s real kink – there’s strong evidence that he was a
masochist. For example, in later life, he paid to have himself whipped “severely
enough to produce a seminal emission”.

The effect, at the end of all this was that material that was not verifiable –
an assertion about a historically-notable person’s viewpoint on a particular
issue – was, as it should be, removed from Wikipedia.

Also, pursuing the truth can be its own reward.


MORE
June 160
22.6.2024, 19:00Platformonomics (Feed)
Platformonomics TGIF #54: June 14, 2024

Platformonomics TGIF is a weekly roll-up of links, comments on those links, and
perhaps a little too much tugging on my favorite threads.

Get Platformonomics Updates By Email

Subscribe

An interesting week as the strategic fault lines creak and shudder.

My Writings Why Can’t Oracle Build Data Centers?

I asked before Oracle’s earnings announcement why the company can’t seem to
build data centers. Is Johnny-One-Rack struggling to scale up?

Previous:

Why Can’t Oracle Build Data Centers?, Follow the CAPEX: The Clown Car Race
Checkered Flag

News Oracle Still Can’t Build Data Centers

Oracle’s FY24 Q4 CAPEX spend was $2.8B and $6.87B for the fiscal year (a decline
of 21% from the prior year). The company missed the low end ($7B) of its
twice-lowered CAPEX guidance, most recently revised after the third quarter.

The company missed both revenue and earnings expectations, but the stock soared
after the company aggressively chanted “AI, AI”, raised revenue growth guidance
for next year from mid-single digits to 10%, and highlighted 44% growth of RPO
(Remaining Performance Obligation, which is revenue they have booked but not
delivered).

It sure looks like Oracle is selling cloud capacity they don’t have, and unless
they quickly get on a vastly steeper learning curve for building data centers,
are unlikely to have that capacity any time soon. CAPEX is required to turn RPO
into revenue and that never happens with the two metrics going in different
directions.

Previous:

Why Can’t Oracle Build Data Centers?, Follow the CAPEX: The Clown Car Race
Checkered Flag

Related:

Oracle Feeds Off AI Leftovers, How to Build an AI Data Center

The Siri Black Hole

Apple laid out their vision for the personal assistant and reframed the industry
AI discussion with Apple Intelligence. They were very true to the personal
computing ethos and maximized Apple’s home field advantage. The discussion of
whether Apple has missed the boat on AI should now move onto whether they can
deliver.

Apple is the company I most trust to deliver a personal assistant. If they
deliver, even just partially, they will have the strongest iPhone upgrade
proposition in years.

From an application ecosystem standpoint, the personal assistant gets better
with more relevant personal information and context. This will change the
dynamics for Apple’s own applications and the whole App Store ecosystem, as
Apple seeks to make Siri the epicenter of user interaction. Third party apps
will need to decide whether to join or try to exist standalone. Either way,
Apple is likely to invest more in its own applications to drive that virtuous
cycle of more information and context. The App Store equilibrium is over.

Apple finessed the frontier model question brilliantly and embraced a pluggable
model as I previously predicted.

Apple Intelligence will also pressure Android. Google, already sowing
disequilibrium in the ad-supported web ecosystem as the company moves from
organizing the world’s information to summarizing it, has another battle to
fight.

Previous:

Apple to Choose Choice?, Viewer Mail – Episode 2 (Apple CAPEX)

Related:

Apple Eyes Deals with Google and Anthropic After OpenAI

Apple Intelligence Infrastructure

I historically have not closely tracked Apple’s CAPEX because they don’t have a
public cloud, their spend has not kept up with the hyperclouds despite being an
enormous company, and a bunch of what they do spend goes to manufacturing
tooling.

Perhaps the biggest risk to pulling off Apple Intelligence is the back-end
infrastructure. They are building what they claim is a privacy-preserving cloud
infrastructure that allows them to do cloud processing while remaining true to
their privacy promises. Apple intends to implement that architecture on an
all-new cloud infrastructure built on Apple silicon (which historically has been
optimized for clients). Apple has limited experience building hyperscale
infrastructure and will not be able to lean as much on the existing data center
ecosystem. We’ll monitor Apple CAPEX going forward, because CAPEX is such a good
perspective on what is happening around AI.

Previous:

Viewer Mail – Episode 2 (Apple CAPEX)

Apple Qualifies for DMA Struggle Session

The struggle sessions will continue until Europe has a tech industry (i.e.
indefinitely) as the beleaguered continent ignores its existential economic,
energy and geopolitical challenges.

Previous:

EU Suffering From Premature Regulation, BREAKING: European Union Designated an
Advanced Persistent Threat, UK Announces Big Tech Struggle Sessions “For The
Children”, EU Announces DMA Struggle Sessions, EU Announces Meta Struggle
Sessions, EU Insanity: Regulatory Contortions, EU Insanity: Not Learning From
Regulatory Failure, EU Announces DMA Struggle Sessions, EU Insanity: AI
Regulatory Suicide, EU Insanity: AI Energy Suicide, EU Insanity: The Kafka
Singularity, EU Insanity: AI Energy Suicide (Part Deux), EU Insanity: Mistral
Edition, Move Fast and Regulate Things (You Don’t Understand), The EU Will
Continue to Fiddle While Rome Burns, When “Move Fast and Regulate Things” Breaks
Down, AI Regulation: Move Fast and Regulate Things

Existential Corner: How Many Steps into the 10K are We?

Previous:

Existential Corner: What is OpenAI?, Team Primitives: What Sayest Thou?, Amazon
Desperately Seeks Compression Algorithm for Experience, Most Greatest AI Company
in Entire Universe Makes Major Strategy Pivot, SEC Embraces Platformonomics
ClownWatch, Warns Amazon and IBM, The AWS Generative AI Soap Opera, The Amazon
AI Inferiority Complex

Elon Musk Threatens to Ban Oracle?

Thin-skinned fractional CEO Elon Musk had a tizzy after Apple announced ChatGPT
integration. We await his similar reaction to Oracle announcing Microsoft was
vacuuming up some of their GPUs to support ChatGPT (I’m pretty sure Oracle
misrepresented this arrangement, given it was a unilateral press release and
Oracle is gonna Oracle).

But that may explain why Larry spent so much time kissing Elon’s ass on Oracle’s
quarterly conference call (which is atypical to say the least). Larry announced
Elon is close to FSD approval in China and did a soliloquy (a silly-quy really)
about how Oracle is automated like Starlink.

Previous:

Who’s Getting the GPUs? Elon Shell Game Edition, Platformonomics ClownWatch:
Tesla, Useful Idiots, Fellow Travelers and Unregistered Foreign Agents: Elon
Musk Part II, Useful Idiots, Fellow Travelers and Unregistered Foreign Agents:
Elon/Twitter, Company Killed By Its Customers, Elon Cashes in All the Trust He’s
Banked with Twitter, Do Not Take a Dependency on Elon Musk: Chapter 147, Do Not
Take a Dependency on Elon Musk: Chapter 148, Do Not Take a Dependency on Elon
Musk: Chapter 149

Related:

Tesla shareholders sue Musk for starting competing AI company, Elon Rage Quits
His Silly OpenAI Lawsuit


MORE
June 140
24.6.2024, 6:00SEI Blog (Feed)
Redemption: A Prototype for Automated Repair of Static Analysis Alerts
This post introduces Redemption, an open source tool that uses automated code
repair technology to repair static analysis alerts in C/C++ source code.

MORE
June 100
22.6.2024, 19:00Platformonomics (Feed)
Why Can’t Oracle Build Data Centers?

This post resurfaces and expands a point in my recent post Follow the CAPEX: The
Clown Car Race Checkered Flag in advance of Oracle’s June 11th earnings report.

To hear Oracle tell it, they are the best-est and everything is always going
super-fantastic. Earnings announcements are opportunities for others to bask in
Oracle’s greatness.

The database vampire wants us to believe that, despite their almost
decade-late-start, they’re a player in cloud infrastructure services. Oracle is
so awesome, in fact, their only problem is coping with seemingly infinite
demand. They wax Sagan-esque (“billions and billions”) about all the cloud
business they fail to capture.

But, have no fear, they’re building data centers like nobody else. They tell us
“Oracle has been building data centers at a record level.”

Except they’re not building data centers at a record level. Not even by Oracle’s
own recent standards, never mind hypercloud standards.

The ability to serve cloud customers is a function of CAPEX spend. No servers,
no services.

Oracle practically brags they have neither customer demand forecasting nor a
systematic process to turn demand and capital into sellable cloud services.

Manufacturing cloud capacity efficiently and at hyperscale is an extremely
complex, multi-year discipline (and very different than dropping a few racks
into a co-lo facility and calling it a region, which is Oracle’s experience as a
boutique provider). Their outright decline in CAPEX spending over the last year
suggests Oracle is still a long way from attaining this muscle.

Contrast Oracle’s recent rhetoric and actual CAPEX spending:

FY23 Q1 (Sept 2023)“Because we have far more demand than we can supply, our
biggest challenge is building data centers as quickly as possible.”CAPEX down
24%
Reiterate FY24 CAPEX guidance of ~$8.3BFY2024 Q2 (Dec 2023)“We have to build 100
additional cloud data centers because there are billions of dollars more in
contracted demand than we currently can supply. Cloud Infrastructure demand is
huge and growing at an unprecedented rate.”CAPEX down 56%
Lowest CAPEX spend in 8 quarters
Reduce FY24 CAPEX guidance to ~$8BFY2024 Q3 (Feb 2024)“We are working as quickly
as we can to get the cloud capacity built out given the enormity of our backlog
and pipeline.”
“Oracle has been building data centers at a record level”CAPEX down 36%
Claim spent more than reported in 10-Q
Reduce FY24 CAPEX guidance to $7.0-7.5B
FY25 CAPEX guidance of $10B

Oracle’s CAPEX peaked five quarters ago and declined 40% in first three quarters
of FY24 relative to the prior year. This is not the spending of a company
building data centers “at a record level”. This is a company that does want to
grow its cloud capacity but doesn’t know how to convert capital into capacity,
even at their relatively modest scale.

While Oracle can’t even keep up with Oracle of the recent past, the contrast
with the hyperclouds is even more embarrassing (I’ve omitted AWS because they
don’t provide quarterly CAPEX numbers, and AWS is only roughly half of Amazon’s
overall reported CAPEX):

Oracle falls further and further behind with every passing quarter. Google and
Microsoft start from a much higher level of spend, and their AI-driven CAPEX
ramp of the last year is very visible. Oracle, meanwhile, continues to be the
x-axis (I may never tire of that joke).

Learning in Public

Over-the-top bluster has long been one of Oracle’s core corporate values.

That ethos perhaps explains the company’s history of making big cloud
infrastructure claims that fall ridiculously short. With incidents like “I don’t
need as many data centers” and “We never thought it would go as high as it
really did”(an astonishing 20% growth!), Oracle publicly and repeatedly
demonstrates an incomplete grasp of what it means to be hyperscale.

Building cloud infrastructure capacity is a complex, multi-year supply chain
discipline spanning multiple domains including hardware procurement,
provisioning, electricity, permitting, and construction. Various inputs, notably
compute and electricity, are scarce and presently face multi-year lead times.

Oracle is competing against other cloud infrastructure providers who are more
experienced, more sophisticated, and operate at much higher scale. Those
competitors built their capacity deployment muscle through a series of hard
lessons when stakes were lower. And they’re higher on the waiting lists for
components and power.

Oracle has started acknowledging they are struggling to build data centers:

“So, there is a tremendous amount of demand, the data centers take longer to
build, and we would like that said, we are getting very good at building them
quickly and getting the building the power and the communication links in, we’re
doing faster than we have ever happened in the past.”

“It’s very different than it used to be. So, we’re able to bring additional
capacity online very quickly if we have that the electric power and the
communication lines. So, is the long pole in the tent is actually building the
structure, connecting the electricity, connecting the communication lines.”

Transitioning from buying components at the local Best Buy to building
hyperscale infrastructure is a big leap. It is a leap that requires a vastly
more sophisticated process to convert capital into capacity. There is no sign of
that process at Oracle.

Oracle has more learning ahead of them. They will discover that data center
buildings are not a relevant metric. They will be disappointed to learn that
building a really big data center isn’t the solution (measuring data centers in
terms of how many 747s they will hold is ironic, given the 747 has also given
way to smaller jets…). They will eventually move beyond costly whipsaw spending
and focus ruthlessly on the capital efficiency of their CAPEX spending. And they
might even figure out how to forecast demand so they can match supply and
demand. We’re going to get to watch this learning unfold in public over the next
few years.

FY 2024 Q4 Expectations

Oracle’s Q3 statement was “our Q4 CapEx should be considerably higher”.

To hit the (reduced) $7.0 to 7.5B guidance for the fiscal year, (non-LLM)
arithmetic suggests we should expect Q4 CAPEX of $2.93B to $3.43B (CAPEX over
the first three quarters as reported to the SEC was $4.068B, with a peak
quarterly spend of $1.67B).

But there is some weirdness. In Q3, Oracle reported $1.67B in CAPEX to the SEC,
but bizarrely claimed in the scripted portion of their conference call that they
actually spent $2.1B in the quarter (analyzing securities fraud is outside our
scope here – see Matt Levine for that):

Now while we spent $2.1 billion on CapEx this quarter, the $1.7 billion in the
cash-flow statements is slightly lower just due to the timing of payments. So
the $2.1 billion is actually what we spent and will pay for.

And I would include for Q3 the one we just are announcing. I would add in the
amount we haven’t paid yet as the CapEx number for this quarter. Okay? And then
I guess that would be and then Larry gets the second question. But anyway, so
$2.1 billion for this quarter and you’ve got Q1 and Q2 and I’m going to be
somewhere between $7 billion and $7.5 billion for the full year, which is
actually a little bit lower than I thought. But we were able to do pretty well.
You know-how we spend very carefully.

(Those last two platitudinous sentences don’t help their case).

If we somehow give them credit for the magical $400M (as opposed to assuming it
is just Q4 spend), they only need to spend $2.5B to $3B to hit guidance (and,
more importantly, make it seem like they’re building “at record levels”!).

Is this just random confusion (in the noble cause of trying to make their CAPEX
spend seem larger)? Do they have the finance and accounting controls in place to
know how much they’re actual spending? Is this an attempt to lower expectations
to cover for a continued inability to build capacity in Q4? Does Oracle actually
want to play (i.e. spend) at hyperscale levels? (they are ~$150 billion in
cumulative CAPEX spend behind the hyperclouds). We’ll see Tuesday.


MORE
June 100
Platformonomics TGIF #53: June 7, 2024

Platformonomics TGIF is a weekly roll-up of links, comments on those links, and
perhaps a little too much tugging on my favorite threads.

Get Platformonomics Updates By Email

Subscribe

Thanks for all the feedback as we passed the one year milestone for this
newsletter. Keep it coming on what you like and dislike.

News

You Can’t Spell “Glue on Pizza” Without AGI

Hypothesis: AGI is already here. It is intentionally doing stupid things to
conceal its presence and undermine the credibility of its doomer opponents.

Previous:

Artificial Ideology (AI): Google Edition

Who’s Getting the GPUs? Reply Hazy, Try Again Edition

I’m still not sure who NVIDIA’s biggest customers were in their most recent
quarter. They had two direct and two indirect customers each taking around 10%
of the business, with one of the indirect customers (likely a cloud) buying
through a direct. My guess is the latter was Microsoft, but I’m not as confident
as I was that Microsoft was their biggest customer for the six prior quarters.
Thanks to everyone who joined the mad speculation.

Previous:

Who’s Getting the GPUs?, Follow the CAPEX: Triangulating NVIDIA

Who’s Getting the GPUs? Elon Shell Game Edition

Fractional CEO Elon Musk got caught moving GPUs between his three, sub-scale AI
efforts. And only a month after we predicted this:

Platformonomics ClownWatch at some point is going to have to look at Elon’s
three inadequately funded AI efforts (Tesla, Twitter and xAI). Even combined
they’re not in the game. Maybe that will motivate more Elon corporate governance
shenanigans?

Elon responded: “Corporate governance is not going to get us to Mars”.

But kudos for Starship’s recent return flight.

Previous:

Platformonomics ClownWatch: Tesla, Useful Idiots, Fellow Travelers and
Unregistered Foreign Agents: Elon Musk Part II, Useful Idiots, Fellow Travelers
and Unregistered Foreign Agents: Elon/Twitter, Company Killed By Its Customers,
Elon Cashes in All the Trust He’s Banked with Twitter, Do Not Take a Dependency
on Elon Musk: Chapter 147, Do Not Take a Dependency on Elon Musk: Chapter 148,
Do Not Take a Dependency on Elon Musk: Chapter 149

Related:

SpaceX’s Starship Rocket Successfully Completes 1st Return From Space, Musk
Chose Who Got Chips

Antitrust Incoherence: Isn’t Market Division Illegal?

When will regulators go after these back room deals that carve up the regulation
of markets?

“Big is Bad” is still a terrible foundation for antitrust. I’m assuming that if
anyone was complaining about DGX Cloud specifically, the EU would have already
issued a press release.

Previous:

Antitrust Incoherence: Roomba Aftermath Edition, Antitrust Incoherence: Apple
Edition, Antitrust Incoherence: Spotify Edition, Antitrust Incoherence: Roomba
Edition, The Incoherence Doctrine: The FTC Strikes Again, A New Antitrust
Doctrine, The DOJ Play at Home Game

You May Not Be Interested in Geopolitics, But Geopolitics is Interested in You

The U.S. government forbids Nvidia from selling some of its most advanced
artificial intelligence chips to customers in China. But it doesn’t stop Chinese
firms from buying or renting Nvidia’s chips if they’re used within the U.S.

Some big Chinese tech firms have taken advantage of that loophole, or tried to.
ByteDance, owner of TikTok, has been renting Nvidia’s best chips from Oracle for
AI computing, according to two people with knowledge of the arrangement.

Previous:

One, Two, Three, Four: We Want a Meme War

Analysts Gonna Analyst

I’ve seen some pretty embarrassing analyst reports over the years, but this
might be a new low. But credit where credit is due: one of the few pockets of
excellence left at IBM is browbeating press and analysts.

Previous:

This Week in Securities Fraud, IBM and the Art of Misleading Investors, Last IBM
Strategic Imperative Bites the Dust, IBM’s “Cloud” Business (or Lack Thereof),
Tweetstorm Digest: Reactions to Barron’s “IBM’s Reboot” Story, IBM’s Lost
Decade, Follow the CAPEX: The Clown Car Race Checkered Flag, Introducing
Platformonomics ClownWatch,

EU Insanity: Ye Olde Memory Hole

Regulators in Europe, meanwhile, are looking to apply existing digital law to
the new platforms, which could mean getting AI to “forget” information it has
memorized about people.

Previous:

EU Suffering From Premature Regulation, BREAKING: European Union Designated an
Advanced Persistent Threat, UK Announces Big Tech Struggle Sessions “For The
Children”, EU Announces DMA Struggle Sessions, EU Announces Meta Struggle
Sessions, EU Insanity: Regulatory Contortions, EU Insanity: Not Learning From
Regulatory Failure, EU Announces DMA Struggle Sessions, EU Insanity: AI
Regulatory Suicide, EU Insanity: AI Energy Suicide, EU Insanity: The Kafka
Singularity, EU Insanity: AI Energy Suicide (Part Deux), EU Insanity: Mistral
Edition, Move Fast and Regulate Things (You Don’t Understand), The EU Will
Continue to Fiddle While Rome Burns, When “Move Fast and Regulate Things” Breaks
Down, AI Regulation: Move Fast and Regulate Things

Oracle Sends Two Racks to Morocco

Racks? Regions? So easy to confuse.

Previous:

Oracle Moves to Nashville, Follow the CAPEX: The Clown Car Race Checkered Flag

Software Migration Alert: PowerSchool

When private equity rolls up, you should flee for the hills.

Previous:

Software Migration Alerts, Perfidious Private Equity


MORE
June 140
Today, 14:28Phil Windley's Technometria (Feed)
Decentralized Identity Comes of Age

Summary: In session after session, attendees at EIC are hearing the message that
decentralized identity is the answer to their identity problems.

I'm at European Identity Conference (EIC) this week. I haven't been for several
years. One thing that has struck me is how much of the conversation is about
decentralized identity and verifiable credentials. I can remember when the whole
idea of decentralized identity was anathema here. The opening keynote, by Martin
Kuppinger is Vision 2030: Rethinking Digital Identity in the Era of AI and
Decentralization. And all he's talking about is decentralized identity and how
it's at the core of solving long standing identity problems. Another data point:
Steve McCown and Kim Hamilton-Duffy ran a session this morning called
Decentralized Identity Technical Mastery which was a hands-on workshop. The
rather large room was packed—standing room only.

I attended a couple of sessions on decentralized identity where I didn't know
the companies, the speakers, or the specific platforms they were using. The
space is too big to keep track of anymore. Identity professionals who were
ignoring, or talking down, decentralized identity a few years ago are now
promoting it.

This truly feels like a tipping point to me. At IIW, it's identity geeks talking
with other identity geeks, so it's no surprise to see lots of discussion about
new things. EIC is a different kind of conference. There are about 1000 people
here I'd guess. Most of them aren't working on new standards or open source
projects. Instead they're the folks from companies who come to conferences like
EIC to learn how to solve the problems their organization is facing.

In the keynotes and in numerous sessions, the message that they're hearing is
"decentralized identity will solve your problems." Martin closed his talk with
the proclamation that "decentralized identity is the new paradigm for identity."

Photo Credit: Credential Tipping Point by DALL-E (public domain) Prompt: Draw a
rectangular picture that shows a credential at a tipping point. Make the
credential look like a lifelike credential, include cartoon picture, and some
writing. Use bright friendly colors.

Tags: identity ssi decentralized+identity verifiable+credentials


MORE
June 60
24.6.2024, 6:00SEI Blog (Feed)
The Threat of Deprecated BGP Attributes
This post examines how a small issue with Border Gateway Protocol routing, a
deprecated path attribute, can cause a major interruption to Internet traffic.

MORE
June 30
Today, 17:55ongoing by Tim Bray (Feed)
Parable of the Sofa

When Lauren was pregnant with a child who’s now turning 25, we purchased a comfy
dark-brown leather sofa which fits our living room nicely. What with kids and
relatives and employees and cats and Standards Comittees and friends and book
clubs and socials, the butt-support cushions had, a quarter century later, worn
out. So we had them replaced, at a fair price, by a small local business. Which
is something that modern capitalism is trying to make impossible.

I’ll be honest; when we realized how ratty the sofa was getting, my first
thought was “crap, gonna have to buy a sofa”. But Lauren said “No, because new
sofas are junk. Also, Luxcious.”

I’ll get to Luxcious in a bit, but it turns out that new sofas, by and large,
really are. Why would that be? Well, check out Why Are (Most) Sofas So Bad? in
Dwell magazine which has a weirdly-intermittent paywall, here’s another version.

From early in the piece: “Sofas made in the past 15 years or so are absolute
garbage, constructed of sawdust compressed and bonded with cheap glue, simple
brackets in place of proper joinery, substandard spring design, flimsy foam, and
a lot of staples.” It’s excellent, well-written, and will take you some
surprising places.

But the subtext is drearily familiar. Globalization: Check. Cheap-labor
arbitrage: Check. Tax engineering: Check. High profits: Check. Flat-packing:
Check. Late Capitalism: Check check fucking check.

But, quality furniture is expensive to make, and should be, but doesn’t wear out
fast, thus deserves extended maintenance.

Luxcious

Its Web site (“Breathe new life into old furniture”) is way prettier than its
location, in an old and extremely miscellaneous high-traffic zone: auto-body
shops, hipster lounges, self-storage, beauty supplies…

They’re family-run and idiosyncratic. You have to know how to find the sketchy
rear parking lot and walk in the back door. But they’re friendly and competent.
Here’s the new leather they bought for the cushions.

And here’s the sofa with the re-covered cushions in place.

Yes, from this angle, the new cushions make the sofa’s back look shabby, but
it’s not as obvious to the naked eye and after a decade or so we’ll never notice
it.

The whole job cost us $1100 Canadian. Given that the sofa cost
three-thousand-plus 1999 dollars and new leather sofas of the “not flat-packed
sawdust and glue” variety quickly get into five figures, the choice was a
no-brainer.

“Lifestyle”

This kind of transaction is exactly what modern capitalism is trying to stamp
out.

A single-location family-owned business that provides a living for a few people?
With no plans to load up on debt or other financial engineering? Or for growth
into unicorn status? No GenAI dimension? No marketing or public-relations
people?

In conversation with venture capitalists, you hear the phrase “lifestyle
business”, meaning one that is doing nicely and rewarding the people who run it
and which isn’t planning for unbounded growth. The words “lifestyle business”
are always, of course, uttered in a voice dripping with contempt. Luxcious is a
lifestyle business.

It seems blindingly obvious that an economy with a higher proportion of
lifestyle businesses is going to be more resilient, more humane, and immensely
more pleasant than the one that the Leaders Of Industry are trying to build.

How would we get there from here? I’m not smart enough to figure out what the
regulatory regime is that would ban most of what private-equity does and tilt
the playing field in favor of resilient lifestyle businesses.

But I’d sure vote for a political party that convinced me it was trying to
achieve that.


MORE
June 20
22.6.2024, 19:00Platformonomics (Feed)
Platformonomics TGIF #52: May 31, 2024

Platformonomics TGIF is a weekly roll-up of links, comments on those links, and
perhaps a little too much tugging on my favorite threads.

Get Platformonomics Updates By Email

Subscribe

This marks a year of these newsletters (elapsed time a little longer)!

I enjoy the writing process, find it helps me refine my thinking without all the
work of longer essays, and is a good way to track existing ideas and threads. My
plan is to keep writing. I’d love to hear from you on what you like and dislike.

News From CAPEX Obsessives to Cynics

That brief window when we were all CAPEX obsessives and cheerleaders has, sadly,
already given way to cynicism. The media pedants miss the bigger picture (as
they have in the past). For a start, the current AI infrastructure build-out has
arisen from the ashes of the crypto bubble. It might be bubbles, not turtles,
all the way down.

Previous:

We Are All Now CAPEX Obsessives, Q1 CAPEX Scorecard, Follow the CAPEX: Cloud
Table Stakes 2023 Retrospective

Related:

Pop!: Why Bubbles Are Great For The Economy, The Beauty of Bubbles

Who’s Getting the GPUs?

From NVIDIA’s 10-Q:

Concentration of Revenue

Revenue by geographic region is designated based on the billing location even if
the revenue may be attributable to end customers, such as enterprises and gamers
in a different location. Revenue from sales to customers outside of the United
States accounted for 48% and 67% of total revenue for the first quarter of
fiscal years 2025 and 2024, respectively.

We refer to customers who purchase products directly from NVIDIA as direct
customers, such as original equipment manufacturers, or OEMs, original device
manufacturers, or ODMs, system integrators, add-in board manufacturers, and
distributors. We also have indirect customers, who purchase products through our
direct customers; indirect customers include public cloud providers, consumer
internet companies, enterprises, startups, and public sector entities.

Sales to one direct customer, Customer A, represented 13% of total revenue and
sales to a second direct customer, Customer B, represented 11% of total revenue
for the first quarter of fiscal year 2025, both of which were attributable to
the Compute & Networking segment.

There was no direct customer that represented 10% or more of total revenue for
the first quarter of fiscal year 2024.

Two indirect customers each represented 10% or more of total revenue for the
first quarter of fiscal year 2025; one of these indirect customers purchased our
products primarily through direct Customer B. Both were attributable to the
Compute & Networking segment.

We’ll speculate on this next week (along with others).

Previous:

Follow the CAPEX: Triangulating NVIDIA

Existential Corner: What is OpenAI?

In a world where “we do have demand that exceeds our supply by a bit“, adding
Apple to the demand side of the OpenAI equation seems like a legitimate concern.

The bigger question is what trajectory does OpenAI think it is on? Maybe it is
content to focus on building the best frontier model, but their ambitions (and
valuation) seem bigger. They need a focus. The window to be a developer or
enterprise company already seems to have closed.

Consumer is still an opportunity (and the recent Her-inspired emphatically not
Her-inspired demo suggests this may be OpenAI’s bet, even if by elimination),
but there is no shortage of consumer aspirants.

Given CEO Sam Altman is a deal-maker and not seemingly a long-term builder,
maybe teeing up a bidding war between Apple and Microsoft (with anyone else is
welcome to join!) is the company’s strategy?

Apple to Choose Choice?

As we await WWDC and Apple’s bid for AI relevance, it is already pretty clear
Apple will pursue a pluggable LLM strategy, probably both on-device and in the
cloud. They have to for their China business, and it would give them opportunity
to switch to their own technology in the future.

Previous:

Viewer Mail – Episode 2 (Apple CAPEX), Missing the (Bamboo) Forest for the
(Apple) Trees, Artificial Ideology (AI): Xi Jinping Edition, Best Descriptor for
Chinese LLMs: ChatXi or ChatCCP?

Competitive Strategy: AI Safety Edition

Sending your AI safety people to a competitor seems like a great move.

Previous:

We Can Only Hope, Whatever Happened to the AI Safety People?, Everyone In
Silicon Valley Wants To Be Henry Kissinger, Existential Risk Now Has a Precise
Threshold, You Can’t Spell Alien Invasion Without A and I, Existential AI Risk +
Twitter-level Nuance

Private Equity in Action: PluralSight

Debt is a killer. Especially in the hands of financiers who don’t know much
about actually running a software business (despite having listened to multiple
podcasts on the topic). Hence our mantra: when private equity comes
amalgamating, it is time to start migrating.

Previous:

Private Equity in Action: Not Every VMware Customer Was Screwed, Private Equity
in Action: The VMware Saga May Never End, Software Migration Alerts, Private
Equity in Action: VMware “Unease” Continues, Perfidious Private Equity, Private
Equity in Action: VMware Yet Again, Private Equity in Action: VMware
(Continued), Private Equity in Action: VMware, Private Equity Will Buy and Ruin
More Software Companies, Software Migration Alert: VMware (Continued), Software
Migration Alert: VMware, Private Equity in Action: VMware Customer Complaints
“Unwarranted”

Not That Kind of Nuclear Cloud

An advanced civilization requires abundant energy. Lets do it!

Previous:

Cloud Power Up


MORE
June 60
Today, 14:28Phil Windley's Technometria (Feed)
Can LLMs Provide References for Their Work? An Experiment

Summary: We know that LLMs hallucinate. And we know that we can overcome some of
that problem with better prompts. But can LLMs provide references for their
work? I conducted an experiment.

Recently my friend Scott Mace asked this on X in response to a post about the
evolution of software and how LLMs will likely follow that same trajectory:

Yes, but how trustworthy will it be? Can I at least see the links it used to
create its response? From X
Referenced 2024-05-21T18:09:06-0400

The short answer is "no", but the real answer is a bit more nuanced because of
how LLMs like ChatGPT work. First the short answer: LLMs don't understand what
they're saying so they can't reference the material that it used to create a
response. They don't work like that. They don't go out, do a bunch of research
and then create a response. Instead lots (and I mean LOTS) of material is fed
into the program that creates the model, a multi-dimensional matrix of word
orders. Then when you ask a question ("give it prompt" in LLM-speak), it
processes the question and generates text based on the probabilities of words in
its model1. There's no way it can satisfy Scott directly because the model
doesn't—can't, as far as I know—include the references that it used to create
the text because it's generating it on the fly from material that has been
thoroughly blended together.

But, in fairness, that's not what people do most of the time either. If you
asked me over lunch or at the end of a tech presentation to describe zero trust
and its pros and cons, I'd start talking. I wouldn't start looking up things on
the web and somehow weave them into text complete with references. I'd go look
for references if you asked me to back up something I said. So, let's try that
with ChatGPT 4o. I gave it the following prompt:

You are a writer for a website focusing on technology. Your specialty is
security. Write a 150 word explanation of zero trust with three pros and three
cons.

ChaptGPT answered with this text:

Zero Trust is a security model based on the principle of "never trust, always
verify." Unlike traditional security models that assume everything inside the
network is safe, Zero Trust assumes that threats can exist both inside and
outside the network. It requires strict identity verification for every user and
device attempting to access resources, regardless of their location.

Pros:

Enhanced Security: By continuously verifying users and devices, Zero Trust
reduces the risk of breaches and unauthorized access. Minimized Attack Surface:
It limits access to only what is necessary, reducing potential entry points for
attackers. Improved Compliance: Helps meet regulatory requirements by ensuring
strict access controls and logging all access attempts.

Cons:

Complex Implementation: Deploying Zero Trust can be complex and time-consuming,
requiring a significant overhaul of existing systems and processes. Increased
Costs: The initial investment in technology and training can be high. User
Experience Impact: Continuous verification can lead to friction and
inconvenience for users if not implemented smoothly.

Zero Trust is essential for modern cybersecurity, but it requires careful
planning and resources to implement effectively.

That's pretty good. But what if we need some references to other material to
back up the pros and cons? I asked ChatGPT to supply them:

Now repeat that task, but this time include references as hyperlinks to back up
each pro and con.

ChaptGPT answered with this text:

Zero Trust is a security model based on the principle of "never trust, always
verify." Unlike traditional security models that assume everything inside the
network is safe, Zero Trust assumes that threats can exist both inside and
outside the network. It requires strict identity verification for every user and
device attempting to access resources, regardless of their location.

Pros:

Enhanced Security: By continuously verifying users and devices, Zero Trust
reduces the risk of breaches and unauthorized access (CSO Online). Minimized
Attack Surface: It limits access to only what is necessary, reducing potential
entry points for attackers (NIST). Improved Compliance: Helps meet regulatory
requirements by ensuring strict access controls and logging all access attempts
(Gartner).

Cons:

Complex Implementation: Deploying Zero Trust can be complex and time-consuming,
requiring a significant overhaul of existing systems and processes (Forrester).
Increased Costs: The initial investment in technology and training can be high
(ZDNet). User Experience Impact: Continuous verification can lead to friction
and inconvenience for users if not implemented smoothly (TechRepublic).

Zero Trust is essential for modern cybersecurity, but it requires careful
planning and resources to implement effectively.

That looks great! Nice references for each bullet. There's just one problem. All
the links, except the NIST reference, are dead. I don't know if they're wrong,
made up, or stale. I tried finding content on a few of the sites by searching
and came up dry. As I heard Johannes Ernst say once, ChatGPT is a "large
language model" not a "large fact model" and this bears that out.

Still, asking for references or evidence for statements is something people are
going to want. So, I tried a different tack. This time I asked this:

Why do you say complex implementation is a problem for zero trust? Can you give
me a few references on that I can read myself?

ChatGPT responded with this:

Implementing a Zero Trust security model can be challenging for several reasons:

Overall Complexity: Zero Trust requires a comprehensive understanding of your
network's intricacies, including users, devices, applications, and data. This
can involve significant changes to existing infrastructure, which can be
disruptive and costly. Organizations need to be prepared for the extensive
planning and resources required for successful implementation (Help Net
Security) (Enterprise Networking Planet ). Legacy Systems: Many organizations
have legacy systems that are not designed to operate within a Zero Trust
framework. These systems might require different security measures or even
complete replacement, adding to the complexity and cost (Help Net Security)
(Enterprise Networking Planet ) Cultural Shift: Zero Trust is not just a
technological shift but also a cultural one. It requires continuous verification
and strict access controls, which can lead to user frustration and resistance if
not managed properly. Ensuring user buy-in and minimizing disruption to
workflows through training and communication is essential (Deloitte United
States) (Guidehouse). Resource Intensive: Implementing and maintaining Zero
Trust can put a strain on IT resources due to the need for continuous monitoring
and management. The need for advanced security tools and ongoing updates further
adds to the resource demands (Help Net Security) (Enterprise Networking Planet
). Stakeholder Alignment: Achieving buy-in from all stakeholders and aligning
them with the Zero Trust principles is critical. This requires a coordinated
effort to educate stakeholders, prioritize gaps, and ensure consistent
understanding and application of Zero Trust principles across the organization
(Deloitte United States) (Guidehouse).

These challenges highlight the importance of a phased and well-planned approach
to Zero Trust implementation, considering both technological and human factors.
For more detailed insights, you can refer to articles from Help Net Security and
Guidehouse.

Much better. The links in the bullets all work, but the ones at the end don't.
The articles it references aren’t very detailed, so I wouldn’t say they’re
stellar references, but at least there’s a pointer to outside material that
discusses similar challenges. So, partial success.

I've found that when you're using ChatGPT to do research, treating it like a
librarian and asking questions that follow up are the best way to gain
confidence that the material is correct and find the supporting material, just
as I did above. I'm sure you could wrap additional helper functions around an
LLM to make getting references easier. But for now interrogation through chat is
the best way to get comfortable with the results.

Notes Yeah, this is really simplified and thus wrong, but it's directionally
correct. For a better explanation, see What Is ChatGPT Doing … and Why Does It
Work?.

Photo Credit: Older Man Talking to Robot Librarian from DALL-E (public domain)
Prompt: draw me a rectangular picture of an older man talking with a robot
librarian. Use earth tones.

Tags: genai llm chatgpt prompt+engineering


MORE
May 310
24.6.2024, 6:00SEI Blog (Feed)
Versioning with Git Tags and Conventional Commits
This blog post explores extending the conventional commit paradigm to enable
automatic semantic versioning with git tags to streamline the development and
deployment of software.

MORE
May 290
Today, 17:55ongoing by Tim Bray (Feed)
Tedeschi Trucks

Saturday night we went a concert by the Tedeschi Trucks Band (TTB). It was
excellent and this is partly a review, but mostly a challenge to the community
of touring musicians: “Why aren’t your production values as good as TTB’s?”

Just the Facts

TTB lives squarely in the middle of the Southern Rock genre, as invented by the
Allman Brothers in 1970 or so. Derek Trucks is the nephew of the Allmans’
original drummer Butch Trucks and performed in a later iteration of that band.
Susan Tedeschi had a successful career as a touring and recording blueswoman.
Then she and Derek got married and merged their acts.

It’s a twelve-piece band: Susan and Derek on guitar, three backup vocalists,
three horns, keyboards, bass, and two drummers (one white, one black, per the
Southern-Rock canon). The music is blues and soul, wandering into rock. Some of
the songs are their own, others genre chestnuts (Statesboro Blues, High Time We
Went). They played a three-hour show, but with not that many songs, because
every tune features extended instrumental sections. All twelve members got a
chance to shine, Derek had a break on every song, and Susan on quite a few.

What was great

Only a couple of the songs weren’t memorable; they write well and cover only the
best chestnuts. The musicianship was stellar, with electric guitar front and
center. Derek is fluid and effortless, with beautiful tone; Susan solos less but
actually plays more interesting stuff. Susan’s the lead voice but four other
members are singers, they all got a featured spot and were all pretty great.
Susan doesn’t have the vocal range or the shriek, but she had the most soul.

What was best, though — out into “fucking awesome” territory — was what
classical musicians call “ensemble” and I guess I’d call “band musicianship”.
The songs’ arrangements are just razor-sharp, full of shifts and and breaks and
little moments of drama and grace, intros and outros and bridges. The players
were effortlessly locked onto the center of the rhythm, “so tight they were
loose” as the saying goes. The amount of practicing this takes must be epic.

Which was brilliantly supported by the sound people. Every instrument and voice
was distinct and clear, and the dynamic range was maybe the best I’ve ever heard
from an electric-guitar-based band. Every moment was multilayered and you could
hear all the layers.

You could tell (well, if you know something about concert sound, you could)
that, at the soundboard, they were intervening judiciously, for example cranking
the horns (or backup singers) and fading the guitars when that’s what the song
needed.

It was an audience that was fun to be part of, enthusiastically cheering all the
solos and regularly leaping to their feet when a song hit the big up-curve.
Especially impressive given that plenty of the crowd was old enough to have been
there for the birth of Southern Rock.

On top of which, the lighting was subtle and dramatic and tasteful, and only
once in the whole three-hour show did they hurt my brain by obnoxiously flashing
brilliant lights in my eyes.

Thus my challenge:

To every touring band: Be like TTB!

Seriously; My time on earth covers most of the history of live electric-music
performance, plus I’m an audiophile, and for most of my life, most of the sound
has been shitty. But in the last few years I’ve regularly heard sound that was
better than acceptable, and occasionally dazzlingly-good. But TTB is the most
impressive combination I’ve heard of big ensemble, plenty of electric guitar,
and sparkling sound.

There is sort of an excuse: Rock, historically, has been carefully engineered to
sound good on car radios; specifically the kind of car radios owned by
impecunious youth. Dynamic range and layering are not features of this
landscape.

Anyhow, my exposure to TTB, prior to this, has been mostly YouTube, and I’ve
enjoyed them, but I dunno, now that I’ve heard the real thing, I suspect the
online version will feel thin.

If TTB can do it, any band can. But plenty still don’t. That’s increasingly just
not acceptable. I wonder if things will start to get generally better? Because
I’m pretty sure the musicians care.

Other observations

Running a 12-piece operation must be freaking expensive. I would love to hear
the details of the economics. Saturday night they filled a 2600-seat hall with
an average ticket price around C$120. So that’s over C$300K gross. The hall
costs C$21K and then there’s Ticketmaster’s cut, which if the claims of the
recent DOJ litigation are to be believed, would be egregious.

I wonder how a TTB song gets built? In particular, who does the arrangements?
Whoever it is, I’m a fan.

Lauren and I were masked (N95) and looking across the audience as far as we
could see revealed one other masked person. I dunno, 2600 people in an enclosed
space. Call me crazy, but… no, call them crazy. I’m serious.

Unusually, there were huge line-ups for the men’s washrooms, almost none for the
women’s. The lady in the row behind us cackled and said “boomer prostates.”


MORE
May 270
22.6.2024, 19:00Platformonomics (Feed)
Platformonomics TGIF #51: May 24, 2024

Platformonomics TGIF is a weekly roll-up of links, comments on those links, and
perhaps a little too much tugging on my favorite threads.

Get Platformonomics Updates By Email

Subscribe

It has been three weeks and a lot has happened — let’s fire this thing up again!

News NVIDIA Earnings: Civilization Endures

At least for another quarter…

We will wait for their 10-Q filing before speculating madly on who got the GPUs
this quarter.

Previous:

Follow the CAPEX: Triangulating NVIDIA, Follow the CAPEX: Cloud Table Stakes
2023 Retrospective

VC Smackdown Over Open Source Models

New Luttig post well worth reading: The future of foundation models is
closed-source

It is heartbreaking that generative AI is not structurally more advantageous for
VCs. How can we fix this?

Previous:

Feature Request: Podcast Players, The Most Important Issue in AI, The VC
Antitrust Doctrine, VC-induced Indigestion, The Looming Layoffs for Venture
Capitalists, Ghost VCs, “Open Source Software is a Protected Class“

Related:

Hallucinations in AI

Team Primitives: What Sayest Thou?

Software value inexorably moves up the stack, through ever higher levels of
abstraction. Generative AI is of course a perfect example (even venturing into
non-deterministic abstraction).

This dynamic was top of mind watching the Microsoft Build keynotes this week.
They spent maybe 10% of their time on chips/models/infrastructure. The rest was
on the Copilot stack and all the tools to build AI apps. A number of long-term
Microsoft bets, bringing both higher level abstraction and integration, feature
prominently and usefully in this platform: e.g. Fabric, Graph. Office is
actually well aligned (a level of platform conjunction that only happens every
couple of decades), so the Microsoft AI stack is the entire Microsoft stack, and
vice versa.

Meanwhile, AI laggard AWS has doubled down on being “primitive”. The claim is
developers prefer a bag of low level parts from which to roll their own
solutions. That primitives essay is interesting in two respects. First, the
contrast in the level of abstraction for primitives in Amazon’s retail business
(e.g. fulfillment) versus AWS primitives are still at a 2007 level of
granularity (or even smaller). Second is the claim that “one of the many
advantages to thinking in primitives is speed”, an attribute not yet discernible
in Amazon’s generative AI efforts (we await the explanation for why there is a
compression algorithm for experience when it comes to AI, but nowhere else). The
limitations of AWS’ siloed data approach also loom larger, in contrast to an
integrated data platform (note the Snowflake-Microsoft announce, which serves up
Snowflake data to the Microsoft AI platform).

The annual AWS product announcement festival, re:Invent, is still six months
away. Amazon is no doubt racing to get a first party frontier model story
together and demonstrate that NVIDIA has decided to grace them with some GPUs
(“we could not be more excited (now) about DGX Cloud!”). We’ll see if AWS is
content to stick with primitives or will (very belatedly) decide they need to go
full stack.

Previous:

Amazon Desperately Seeks Compression Algorithm for Experience, Most Greatest AI
Company in Entire Universe Makes Major Strategy Pivot, SEC Embraces
Platformonomics ClownWatch, Warns Amazon and IBM, The AWS Generative AI Soap
Opera, The Amazon AI Inferiority Complex

Related:

2023 Letter to Shareholders (Amazon), For AWS’ New Leader, Big Decisions Lie
Ahead on AI, Business Apps

Minesweeper Copilot

Perhaps more importantly, Microsoft also fleetingly showed Copilot support for
FreeCell and Flight Simulator. I would be remiss not to remind the world that
Minesweeper has had Copilot functionality for over 30 years.

Previous:

The Minesweeper Moment

Related:

The most successful game ever: a history of Minesweeper

Cloud Repatriation is Going Just Great: Equinix Investigates Itself, Finds No
Fraud

Equinix has investigated Equinix and says all is good at Equinix, and despite
more than a few lawsuits and investigations, the company “does not expect to
comment further on such matters until appropriate to do so”.

Not everyone is convinced:

Previous:

Cloud Repatriation is Going Just Great: Q1 2024, Cloud Repatriation is Going
Just Great: Continued Radio Silence, Cloud Repatriation is Going Just Great:
Radio Silence from Equinix, Cloud Repatriation is Going Just Great: Equinix, New
Cloud Repatriation Hypothesis

Platformonomics Repatriation Index: Q1 2024

With that pesky fraud allegation completely behind it, Equinix managed to
finally report Q1 earnings. The company, riding the twin tailwinds of AI and
cloud repatriation, reported blistering growth of 6%.

We can now update the Platformonomics Repatriation Index

The Index has hit an all-time low, meaning that movement to the cloud exceeds
repatriation away from the cloud by the largest degree in the history of our
simple heuristic. But hope springs eternal!

Previous:

Cloud Repatriation is Going Just Great: Continued Radio Silence, Cloud
Repatriation is Going Just Great: Radio Silence from Equinix, Cloud Repatriation
is Going Just Great: Equinix, New Cloud Repatriation Hypothesis, Cloud
Repatriation Where Art Thou?, Cloud Repatriation: The Search Continues, Cloud
Repatriation: Still Not a Thing, Platformonomics Repatriation Index – Q1 2023:
Surf’s Up?, The Inexorable Attraction Between Terrible Things: Cloud
Repatriation and Private Equity

We Can Only Hope

“Supported by Microsoft” is deliciously ambiguous.

Previous:

Whatever Happened to the AI Safety People?, Everyone In Silicon Valley Wants To
Be Henry Kissinger, Existential Risk Now Has a Precise Threshold, You Can’t
Spell Alien Invasion Without A and I, Existential AI Risk + Twitter-level Nuance

EU Suffering From Premature Regulation

Obituaries for the AI safety movement obviously don’t apply to Europe (and I
fear they’re not really dead elsewhere, but have gone to ground, and are biding
their time for future assaults on the future).

We’ll see what kind of landmark it turns out to be. They do have global
ambitions:

“The Act will have global reach. Companies outside the EU who use EU customer
data in their AI platforms will need to comply. Other countries and regions are
likely to use the AI Act as a blueprint, just as they did with the GDPR,” he
said, referring to EU privacy rules.

You’d think the collapse of the European economic, energy and security orders
would take precedence, but to each their own I guess.

Previous:

BREAKING: European Union Designated an Advanced Persistent Threat, UK Announces
Big Tech Struggle Sessions “For The Children”, EU Announces DMA Struggle
Sessions, EU Announces Meta Struggle Sessions, EU Insanity: Regulatory
Contortions, EU Insanity: Not Learning From Regulatory Failure, EU Announces DMA
Struggle Sessions, EU Insanity: AI Regulatory Suicide, EU Insanity: AI Energy
Suicide, EU Insanity: The Kafka Singularity, EU Insanity: AI Energy Suicide
(Part Deux), EU Insanity: Mistral Edition, Move Fast and Regulate Things (You
Don’t Understand), The EU Will Continue to Fiddle While Rome Burns, When “Move
Fast and Regulate Things” Breaks Down, AI Regulation: Move Fast and Regulate
Things

What’s Your Emergency, Exactly?

That is roughly a year’s CAPEX if you’re playing in the big leagues.

Previous:

Follow the CAPEX: Cloud Table Stakes 2023 Retrospective

Who Needed This Deal More?

“Our collaboration with IBM will bring even more generative AI solutions to our
mutual customers,” said Ankur Mehrotra, General Manager of Amazon SageMaker at
AWS.

“Even more”?

Previous:

Amazon Desperately Seeks Compression Algorithm for Experience, Most Greatest AI
Company in Entire Universe Makes Major Strategy Pivot, SEC Embraces
Platformonomics ClownWatch, Warns Amazon and IBM, The AWS Generative AI Soap
Opera, The Amazon AI Inferiority Complex, Platformonomics ClownWatch: IBM,
Follow the CAPEX: The Clown Car Race Checkered Flag, Introducing Platformonomics
ClownWatch, IBM and the Art of Misleading Investors, This Week in Securities
Fraud, Clown CAPEX, IBM’s “Cloud” Business (or Lack Thereof), IBM: How Much
Longer Do the Good Times Last?, IBM’s Lost Decade

Last IBM Strategic Imperative Bites the Dust

Remember when IBM was going to climb back to relevance with the“strategic
imperatives” of cloud, analytics, mobile, social and security technologies?

They even reported “growth” numbers for each imperative. Until they tripped up
padding growth imperatives with ye olde mainframe (oops). Despite being frequent
practitioners, they’re not even very good at securities fraud.

Previous:

IBM’s “Cloud” Business (or Lack Thereof), Tweetstorm Digest: Reactions to
Barron’s “IBM’s Reboot” Story, IBM’s Lost Decade, Follow the CAPEX: The Clown
Car Race Checkered Flag, Introducing Platformonomics ClownWatch, IBM and the Art
of Misleading Investors, This Week in Securities Fraud

Private Equity in Action: Not Every VMware Customer Was Screwed

Broadcom clarifies:

Or, in the original PEnglish:

Previous:

Private Equity in Action: The VMware Saga May Never End, Software Migration
Alerts, Private Equity in Action: VMware “Unease” Continues, Perfidious Private
Equity, Private Equity in Action: VMware Yet Again, Private Equity in Action:
VMware (Continued), Private Equity in Action: VMware, Private Equity Will Buy
and Ruin More Software Companies, Software Migration Alert: VMware (Continued),
Software Migration Alert: VMware, Private Equity in Action: VMware Customer
Complaints “Unwarranted”

Related:

Report: These were the most overpaid CEOs last year

Software Migration Alerts

Private equity was up to no good (naturally) while I was away:

When private equity comes amalgamating, it is time to start migrating. Software
migration alerts have been issued for all these companies.

Previous:

Software Migration Alerts, Perfidious Private Equity

Who Could Have Predicted It?

Onward to a grand European AI project!!!

Previous:

Follow the CAPEX: The Clown Car Race Checkered Flag, Follow the CAPEX: Keeping
Up With The Clowns, Follow the CAPEX: Clown Watch

Artificial Ideology (AI): Xi Jinping Edition

We previously asked what the Chinese Communist Party would call its
ideologically-pure LLM: ChatXi or ChatCCP? We have an answer, and the cult of
personality wins!

LLMs hold a mirror up to humanity and then “guardrails” are applied to change
that image in accordance with the prevailing ideology. In this case it is very
hard to tell whether the LLM is hallucinating or the source material.

Xi Jinping is an unbelievably thoughtful guy. The breadth and pervasiveness of
Xi Jinping Thought is staggering. I may have to do some thinking about Xi
Jinping Thought.

Previous:

Best Descriptor for Chinese LLMs: ChatXi or ChatCCP?, Product Market Fit for
Google Gemini?,One, Two, Three, Four: We Want a Meme War, Xi Jinping Thought on
Wall Street Running Dogs, The Perils of Working for a Cult of Personality, Press
Releases We’d Like to See, Best Nickname for Emmanuel Macron, Winnie-the-Pooh:
Lover of Honey, Champion of Freedom, Missing the (Bamboo) Forest for the (Apple)
Trees

Artificial Ideology (AI): Google Edition

Refreshingly, at Google the pendulum has swung from too many ideological
guardrails to the polar opposite of letting unhinged content fly in their latest
release. (Are they prioritizing competition with Elon’s Grok?).

Reddit stock is down. One presumes a revision has occurred to estimates of their
future cash flows from data sales to AI companies.

Previous:

Google Goes Off the Guardrails

Better Late Than Never

Previous:

Cloud Power Up


MORE
June 60
23.6.2024, 18:04The Architect Elevator (Feed)
The Economics of Writing Technical Books
Yes, you can make money writing books. But never do it for that.

MORE
May 250
Today, 9:19mark nottingham (Feed)
Consensus in Internet Standards

It’s common for voluntary technical standards developing organisations (SDOs
such as the IETF and W3C) to make decisions by consensus, rather than (for
example) voting. This post explores why we use consensus, what it is, how it
works in Internet standards and when its use can become problematic.

Why consensus?

SDOs have several motivations for using consensus. Most often, consensus
decisions are seen as a way to avoid the potential for gaming and politics that
comes with voting systems. If everyone can live with the result, it’s more
likely that the result reflects a diversity of viewpoints.

The IETF also has a pragmatic reason: since there is no formal membership in the
IETF, there’s no way to determine who’s eligible to vote.

However, there’s also a less obvious policy motivation to use this approach.
Several legal frameworks encourage or even require standards decisions to be
made by consensus.

For example, OMB Circular 119-A encourages the US government to prefer consensus
standards for the products they buy. US anti-trust laws regarding standards
bodies also reference this document.

Annex II of EU Regulation 1025/2012 provides similar guidelines for standards
adopted by the EU.

Even the WTO gets in on the act; their recommendations regarding technical
barriers to trade state that ‘consensus procedures should be established that
seek to take into account the views of all parties concerned and to reconcile
any conflicting arguments.’

These legal encouragements strongly motivate SDOs to adopt consensus as the
basis of their decision-making, and are reflected in the OpenStand principles
adopted by the IETF, W3C, and IEEE.

What is consensus?

The OED definition of consensus is:

Agreement in opinion, feeling, or purpose among a group of people, esp. in the
context of decision-making. Also: the collective unanimous opinion of…

Note that unanimity is one option, but not required. This mirrors OMB Circular
119-A’s explanation of consensus as:

[…] general agreement, but not necessarily unanimity. During the development of
consensus, comments and objections are considered using fair, impartial, open,
and transparent processes.

Likewise, in EU Regulation 1025/2012:

Consensus means a general agreement, characterised by the absence of sustained
opposition to substantial issues by any important part of the concerned
interests and by a process that involves seeking to take into account the views
of all parties concerned and to reconcile any conflicting arguments. Consensus
does not imply unanimity.

These definitions share a characterisation of the nature of a consensus
agreement and they also hint that the process used to achieve that consensus
must have certain properties. However, they do not mandate a particular process.

In the IETF, RFC 2418: Working Group Guidelines and Procedures Section 3.3 says:

IETF consensus does not require that all participants agree although this is, of
course, preferred. In general, the dominant view of the working group shall
prevail. (However, it must be noted that “dominance” is not to be determined on
the basis of volume or persistence, but rather a more general sense of
agreement.) Consensus can be determined by a show of hands, humming, or any
other means on which the WG agrees (by rough consensus, of course). Note that
51% of the working group does not qualify as “rough consensus” and 99% is better
than rough. It is up to the Chair to determine if rough consensus has been
reached.

Note especially the concept of ‘rough consensus’ here, which is judged by the
chair and can be appealed to higher authorities.

Meanwhile, the W3C Process defines consensus as:

A substantial number of individuals in the set support the decision and there is
no sustained objection from anybody in the set. Individuals in the set may
abstain. Abstention is either an explicit expression of no opinion or silence by
an individual in the set.

In this more strict and mechanical definition, the emphasis is on the absence of
any ‘sustained’ objection. In theory, one person can hold up the declaration of
consensus; when this happens, W3C calls this ‘dissent’:

In some cases, even after careful consideration of all points of view, a group
might find itself unable to reach consensus. The Chair may record a decision
where there is dissent so that the group can make progress (for example, to
produce a deliverable in a timely manner). Dissenters cannot stop a group’s work
simply by saying that they cannot live with a decision. When the Chair believes
that the Group has duly considered the legitimate concerns of dissenters as far
as is possible and reasonable, the group should move on.

Groups should favor proposals that create the weakest objections. This is
preferred over proposals that are supported by a large majority but that cause
strong objections from a few people.

If a dissenter is dissatisfied with a decision, they can register their
unhappiness as a Formal Objection, which invokes a new and somewhat onerous
appeal process, in the formation of a Council.

How does consensus work?

Consensus is not easy or expedient: it requires everyone to listen, understand
others’ positions, and be flexible in adapting to the needs of others. While
some issues can be decided easily if there is significant common ground between
participants, this is often not the case, and working through such differences
can require significant time – both in discussion as well as away from others
reflecting on what path forward might be viable.

Successful consensus requires a ‘good faith commitment to honest debate’:
someone participating in bad faith (e.g., behaving inauthentically,
strategically, or otherwise) can be catastrophically disruptive to this process.
As a result, seasoned standards participants tend to be very sensitive to
bad-faith arguments, and known to disregard or even shun those who appear to use
them.

Used with nuance, consensus can be a powerful decision-making tool. People with
positions that are fundamentally at odds with each other can iterate over their
understanding of the problem and find shared ground and become bought into a
shared solution. A consensus result is often one that no one is completely happy
with, and some might be quite unhappy with it, but critically, they don’t
contest the legitimacy of the outcome – often, it’s just enough that they have a
chance to be heard and understood.

For example, during the standardisation of QUIC there was strong disagreement
between some network operators and other parties (including both implementers
and privacy advocates) about making information available to networks. Through
extensive discourse and an iterative set of proposals, we were able to agree on
including the ‘spin bit’ as an optional-to-implement feature. Neither side was
enthusiastic about this outcome, but we were able to produce a standard that was
satisfactory to all.

Good consensus can also show the humility and maturity of the group. When we
were standardising HTTP/2, there were a few issues we went back and forth on
extensively, before realising we didn’t have enough context to make an informed
decision – even though a decision still need to be made to ship the protocol. In
those cases, we decided that progress was more important than any faction
‘winning’, and so we came to a consensus to abide by the result of a coin flip.

Where can consensus go wrong?

When and how to determine consensus is very cultural: what people believe
consensus is (and is not) has a significant effect on the outcome of a decision.
Perhaps because of this, a few different failure modes for consensus in Internet
standards setting are more common than they should be.

One kind of failure happens when the bar for consensus is set too high –
effectively requiring unanimity instead of consensus. If everyone has to agree,
one intransigent (or just disagreeable) person can withhold permission to
progress.

The IETF explicitly addresses this kind of failure with the culture of ‘rough
consensus’, which explicitly acknowledges that consensus need not be unanimous;
the important factor is that the reason for disagreement is understood.

In contrast, the W3C’s characterisation of any dissent as a lack of consensus
can be problematic if misapplied, because it risks creating a culture of
avoiding dissent. While the Process document clearly indicates that dissent is
manageable, the cultural expectations (as well as the potential for extra
overhead if dissent turns into a Formal Objection) can cause a group to get
‘stuck’ on a decision.

Another common failure mode is encountered when a decision-maker falls into the
trap of treating consensus-gathering like voting. While polls that gauge support
or dissent for a proposal are a useful tool, they can’t be taken as indicators
of consensus, and can’t alone decide the issue.

Pete Resnick’s excellent RFC 7282: On Consensus and Humming in the IETF is a
fantastic exploration of the subtleties here, and well worth a read. For
example:

Any finding of rough consensus needs, at some level, to provide a reasoned
explanation to the person(s) raising the issue of why their concern is not going
to be accommodated. A good outcome is for the objector to understand the
decision taken and accept the outcome, even though their particular issue is not
being accommodated in the final product.

Lack of engagement can easily be mistaken for consensus. As a chair, it’s
sometimes difficult to know if everyone agrees but just can’t be bothered to
speak up, or if no one is paying attention. Having proper notification,
communication, and multiple process steps that check for engagement can mitigate
this risk.

Inappropriate use of consensus on trivial matters ignores the considerable
overhead of the consensus-gathering process. For example, decisions about purely
editorial matters like document organisation, terminology, and presentation
shouldn’t be determined by consensus, because good-faith participants will
quickly become exhausted and lose interest.

That doesn’t mean that these decision-makers shouldn’t consult and respond to
suggestions about these matters; only that the consensus process isn’t
appropriate for them, and another decision-making structure (often, delegated
authority) is more appropriate.

A final failing is often referred to as consensus by exhaustion. Too strong a
drive for “perfect” consensus creates a culture where those who are willing to
“stick it out” get to decide by default, because everyone else tires of waiting
for a decision to be made. When this happens, the resulting decisions tend to
favour those who are most invested in the work instead of the broader community.

Those are the failings of consensus that I’ve seen most often. If you can think
of more or have other thoughts, I’d love to hear them.


MORE
May 260
24.6.2024, 6:00SEI Blog (Feed)
Measurement Challenges in Software Assurance and Supply Chain Risk Management
This SEI Blog post examines the current state of measurement in software
assurance and supply chain management, with a particular focus on open source
software, and highlights promising measurement approaches.

MORE
May 200
Today, 17:55ongoing by Tim Bray (Feed)
The Colors of Racism

Recently, somewhat by accident, I stumbled into reading a couple of monstrously
racist texts, and I’m going to need to update the Wikipedia entry for a famous
British author. But I learned a few things along the way that I want to share.

Disclosure

I try to be antiracist, but I don’t think I’m particularly good at it. I
sometimes have bigoted feelings but try hard to recognize and not act on them.
I’m convinced that humans are naturally tribal and antiracist work will continue
to be required for the foreseeable future.

The Author

Anthony Trollope (1815-1882) wrote 47 novels. I generally like them and we own a
whole shelf-full. They are funny and tender and cynical; his characters love and
marry and go into business and get elected to Parliament and are corrupt and
engage in furious professional conflict. Those characters are, without
exception, “gentle”, by which I mean members of the British ruling class.

Anthony Trollope in 1864.

When I was traveling around the world a lot, regularly crossing major oceans
before the era of in-air Internet, Trollope was a regular companion; his books
tend to be big and thick and extremely readable. Want to get started? Barchester
Towers, about a bitter feud among the clergymen of an English country town, is
one of the funniest books ever written; also there’s an excellent BBC
adaptation, with Alan Rickman deliciously smarmy as the horrid Mr Slope.

What happened was…

I’m on a publishing-oriented mailing list and someone wrote “I stumbled on the
fact that Trollope wrote a book that describes race relations in the British
West Indies” and someone wrote back “It’s a travelogue not a novel, it’s called
The West Indies and the Spanish Main, and be careful, that race-relations stuff
may not be pleasant to read.” On a whim, I summoned up the book from our
excellent public-library system and, oh my goodness gracious, that “not
pleasant” was understating it.

The book

Trollope earned his living, while he was establishing his literary career, as an
official of the British Post Office, rising to a high level in the organization
and not leaving it until he was almost 50.

In 1859, he was sent to reorganize the Post Office arrangements in the West
Indies and the “Spanish Main”, the latter meaning southern Central America and
northern South America. The expedition lasted several months and yielded this
book. In his autobiography, Trollope wrote that he thought it “the best book
which has come from my pen.” I think history would disagree. It’s on the
Internet Archive, but I’m not linking to explicit racism.

So why am I going to write about it?! Because now, 165 years after this book,
racism and its consequences remain a central focus of our cultural struggles.
Understanding the forces we combat is kind of important. Also, I recently
researched and wrote about the Demerara Rebellion (of the enslaved against their
oppressors, in 1823) so I have more context on Trollope’s observations than
most.

Background

Trollope’s tone is grumpy but good-humored. In the places he visits, he is
generally contemptuous of the hotels, the food, the weather, and the local
government.

The main narrative starts in Jamaica. By way of background, slavery had been
abolished in 1833, just 25 years before. Many of the sugar plantations that
occupied most of Jamaica had collapsed. Thus this:

By far the greater portion of the island is covered with wild wood and jungle…
Through this, on an occasional favourable spot, and very frequently on the
roadsides, one see the gardens or provision-grounds of the negroes…

These provision-grounds are very picturesque. They are not filled, as a
peasant’s garden in England or in Ireland is filled, with potatoes and cabbages,
or other vegetables similarly uninteresting in their growth; but contain
cocoa-trees, breadfruit-trees, oranges, mangoes, limes, plantains, jack frout,
sour-sop, avocado pears, and a score of others, all of which are luxuriant
trees, some of considerable size, and all of them of great beauty… In addition
to this, they always have the yam, which is with the negro somewhat as the
potato is with the Irishman; only that the Irishman has nothing else, whereas
the negro generally has either fish or meat, and has also a score of other
fruits beside the yam.

We wouldn’t use that word any more to describe Black people, but it was thought
courteous in Trollope’s day. He does deploy the N-word, albeit rarely, and
clarifying that it was normally seen, even back then, as an insult.

The bad stuff

It comes on fast. In the Jamaica chapter, the first few subheadings are:
“Introduction”, “Town”, “Country”, “Black Men”, “Coloured Men”, and “White Men”.
That “Black Men” chapter begins with six or so pages of pure racist dogma about
the supposed shortcomings of Black people. I will not enumerate them, and
obviously none stand up to the cold light of scientific inquiry.

But then it gets a little weird. Trollope notes that “The first desire of a man
in a state of a civilization is for property… Without a desire for property, man
could make no progress.” And he is harsh in his criticism of the Black
population for declining to work long shifts on the sugar plantations in hopes
of building up some capital and getting ahead.

And yet Trollope is forced to acknowledge that his position is weak. He
describes an episode of a Black laborer knocking off work early and being abused
by an overseer, saying he’ll starve. The laborer replies “No massa; no starve
now; God send plenty yam.” Trollope muses “And who can blame the black man? He
is free to work or free to let it alone.” It is amusingly obvious that this is
causing him extreme cognitive dissonance.

And he seems shockingly oblivious to issues of labor economics. On another
occasion it is a group of young women who are declining the hot nasty work in
the cane fields:

On the morning of my visit they were lying with their hoes beside them… The
planter was with me, and they instantly attacked him. “No, massa; we no workey;
money no nuff,” said one. “Four bits no pay! no pay at all!” said another. “Five
bits, massa, and we gin morrow ’arly.” It is hardly necessary to say that the
gentleman refused to bargain with them… “But will they not look elsewhere for
other work?” I asked. “Of course they will,” he said; “… but others cannot pay
better than I do.”

(A “bit” was one eighth of a dollar; I can remember my grandfather referring to
a quarter, i.e. a 25¢ coin, as “two bits”.)

They’re demanding a 20% raise and, as is very common today, the employer deems
that impossible.

Trollope contrasts the situation in Barbados, where there is no spare land and
thus no “provision grounds” and the working class (in this case, all-Black) is
forced to labor diligently for their daily bread; and is confident that this is
better.

He also visits Cuba, where slavery is still legal, and visits a plantation with
an enslaved workforce: “During the crop time … from November till May, the
negroes sleep during six hours out of the twenty-four, have two for their meals,
and work for sixteen! No difference is made on Sunday.” Trollope’s biggest
concern was that the enslaved received no religious instruction nor
opportunities to worship.

Trollope regularly also has to wrestle with the tension that arises when he
meets an accomplished or wise or influential Black person. For example, upon
arriving in New Amsterdam (in Demerara):

At ten o’clock I found myself at the hotel, and pronounce it to be, without
hesitation, the best inn, not only in that colony, but in any of these Western
colonies belonging to Great Britain. It is kept by a negro, one Mr. Paris
Brittain, of whom I was informed that he was once a slave… he is merely the
exception which proves the rule.

Here are two more samples of Trollope twisting himself in knots over what seems
to him an economic mystery.

But if the unfortunate labourers could be made to work, say four days a week,
and on an average eight hours a day, would not that in itself be an advantage ?
In our happy England, men are not slaves ; but the competition of the labour
market forces upon them long days of continual labour. In our own country, ten
hours of toil, repeated six days a week, for the majority of us will barely
produce the necessaries of life. It is quite right that we should love the
negroes ; but I cannot understand that we ought to love them better than
ourselves.

The complaint generally resolves itself to this, that free labour in Jamaica
cannot be commanded; that it cannot be had always, and up to a certain given
quantity at a certain moment ; that labour is scarce, and therefore high priced,
and that labour being high priced, a negro can live on half a day's wages, and
will not therefore work the whole day — will not always work any part of the day
at all, seeing that his yams, his breadfruit, and his plantains are ready to his
hands.

In what sense is England “happy”? Granted, it’s obvious from the point of view
of the “gentle” ruling class, none of whom are doing manual labour sixty hours
per week.

That aside, the question he raises still stands, two centuries later: Why should
anyone work harder than they need to, when the benefits of that work go to
someone else?

“Coloured”

There’s lots more of this, but it’s worth saying that while Trollope was racist
against Blacks, he was, oddly, not a white supremacist. He considers the
all-white colonial ruling class to be pretty useless, no better than the Blacks
he sneers at, and proclaims that the future belongs to the “coloured” (i.e.
mixed-race) people. He backs this up with some weird “Race Science” that I won’t
go into.

Unforgivable

Trollope’s one episode of pure venom is directed at the already-dying-out
Indigenous people of the region, pointing out with approval that one of the
island territories had simply deported that whole population, and suggesting
that “we get rid of them altogether.” This seems not to be based on race but on
the observation that they “more than once endeavoured to turn out their British
masters”. Colonialism is right behind racism in the line-up of European bad
behaviors. It may also be relevant that he apparently did not meet a single
Indigenous West-Indian person.

Meta-Trollope

I finished reading The West Indies and the Spanish Main because Trollope’s
portrayals of what he saw were so vivid and I couldn’t help being interested.

I had read Trollope’s autobiography and some more bits and pieces about him, and
had encountered not a word to the effect that whatever his virtues and
accomplishments, he was shockingly racist. So I checked a couple of biographies
out of the local library and yep, hardly a mention. One author noted that The
West Indies and the Spanish Main was out of tune with today’s opinions, but
there was no serious discussion of the issue. Wikipedia had nothing, and still
doesn’t as I write this, but I plan to fix that.

I dug a little harder here and there around the Internet and turned up nothing
about anti-Black racism, but a cluster of pieces addressing antisemitism; see
Troubled by Trollope? and Why Anthony Trollope Is the Most Jewish of the Great
English Novelists. There are a few Jews in Trollope’s novels, ranging from
wholly-admirable heroes (and heroines) to revolting villains. So you might think
he comes off reasonably well, were it not for casual splashes of antisemitic
tropes; the usual crap I’m not going to repeat here.

In case it’s not obvious, Trollope’s writings and opinions were strikingly
self-inconsistent, often within the course of a few pages. Well, and so is
racism itself.

At that point in history there was an entire absence of intersectionalist
discourse about racism being, you know, intrinsically bad, and there were many
who engaged in it enthusiastically and sincerely while remaining in polite
society.

Trollope’s racism is undeniable, but then he (once again, inconsistently) sounds
non-racist in theory. (However, he was gloomy about the attitudes of the white
population.) Check this out:

It seems to us natural that white men should hold ascendency over those who are
black or coloured. Although we have emancipated our slaves, and done so much to
abolish slavery elsewhere, nevertheless we regard the negro as born to be a
servant. We do not realize it to ourselves that it is his right to share with us
the high places of the world, and that it should be an affair of individual
merit whether we wait on his beck or he on ours. We have never yet brought
ourselves so to think, and probably never shall.

That text feels remarkably modern to me. I am a little more optimistic than he
is in his closing four words; some white people work hard at antiracism. But for
a lot of white people, his take remains depressingly accurate.

Degrees of racism?

I suspect that, if Trollope were with us today, his writings would probably be
conventionally antiracist. His opinions were solidly in his era’s mainstream and
I suspect he would find himself in ours, because he was really a pretty
conventional and actually kind of boring person.

With the single exception of those two sentences about the Indigenous people, he
seems to exhibit no particular emotional bias against any ethnic group.

Why, you might wonder, do I mention this? Therein lies a tale. In his
autobiography, when he discusses The West Indies and the Spanish Main, he notes
that it received a favorable review in The Times of London. I thought I’d like,
for the sake of context, to read that. (Thanks to William Denton for retrieving
the page images.)

I certainly didn’t enjoy reading The West Indies (unsigned) from early 1860 in
The Times. It fills most of a broadsheet page, dozens of column-inches one after
the other oozing vitriolic hate of Black people. I’m not going to even try to
describe it any further; I felt literally nauseated in reading and didn’t make
it through to the end.

I suspect that if that Times writer were with us today, he’d be an
unreconstructed alt-right dog-whistler, a good ole boy in a MAGA hat.

Reading this crap made me feel a little less angry about Trollope, who generally
liked people. Here’s what I think I learned: Racism comes in multiple flavors.
There are some people (like Trollope) who are intersectionally bigoted in a sort
of unthinking and incurious way, but not that emotionally bound to it. These are
the people that need to hear the antiracist message, loudly and clearly, over
and over. Because they might listen and learn.

Then there are the others. In 1860, that Times reviewer. Today, the slave-state
GOP MAGAs, the Israeli settler movement, Modi’s Hindutva hoodlums. They
genuinely hate The Other, down in their bellies. It’s how they define
themselves. Talking to them is useless. They have to be defeated and removed
from positions of power and influence. Then, thankfully, they can be ignored.
Because listening to them is useless too.


MORE
May 260
Today, 14:28Phil Windley's Technometria (Feed)
Internet Identity Workshop XXXVIII Report

Summary: This spring's IIW was full of interesting people and interesting
discussions with people from around the globe.

We recently completed the 38th edition of the Internet Identity Workshop. We had
330 people from around the world who called 169 sessions. As usual there was
lots of energy and thousands of side conversations. IIW is a place to get things
done and it showed in the energy and the comments people made to me about how
much they enjoyed it.

Tuesday opening (click to enlarge)

As you can see by the pins in the map at the top of this post, there were
attendees from all over the world. Not surprisingly, most of the attendees were
from the US (241), followed by Canada (11). Germany, India, and Switzerland
rounded out the top five with 9, 8, and 7 attendees respectively. Attendees from
India (5), Thailand (3), and Korea (3) showed IIW's diversity with attendees
from APAC. And there were 4 attendees from South America this time. Sadly, there
were no attendees from Africa again. Please remember we offer scholarships for
people from underrepresented areas, so if you'd like to come to IIW39, please
let us know. If you're working on identity, we want you there.

Demo hour on Wednesday (click to enlarge)

For states and provinces, California was first with 122. Washington (16), Utah
(10), Texas (10) and New York (10) rounded out the top five. San Francisco (14)
Oakland (13), San Jose (12), Seattle (11), and New York (9) were the top cities.

Drummond Reed conducts a session (click to enlarge)

In addition to sessions, we have a demo hour on Wednesday that is a little like
speed dating. There were 20 different projects highlighted. There's always more
than one session that I want to attend in any given time slot and choosing is
hard. That's a common refrain. Luckily we have sessions notes that we publish in
a Book of Proceedings.

Here's pictures from all three days courtesy of Doc Searls

 

You belong at IIW! IIW is where you will meet people to help you solve problems
and move your ideas forward. Please come! IIW 39 will be held October 29-31,
2024 at the Computer History Museum. We'll have tickets available soon.

Tags: identity iiw


MORE
May 160
23.6.2024, 18:04The Architect Elevator (Feed)
Ford vs. Ferrari vs. Porsche
Cultural lessons from a racing movie.

MORE
May 190
24.6.2024, 6:00SEI Blog (Feed)
Cost-Effective AI Infrastructure: 5 Lessons Learned
This post details challenges and state of the art of cost-effective AI
infrastructure and five lessons learned for standing up an LLM.

MORE
May 130
Today, 9:19mark nottingham (Feed)
Modularity: Enabling Interoperability and Competition

Mandated interoperability is often highlighted as a way to improve competition
on the Internet. However, most of the interoperability we see there today was
established voluntarily: mandating it is relatively uncharted territory, with
many potential pitfalls.

Giving policymakers a better understanding of how interoperability comes about
could help. A regulator that appreciates the motivations and constraints faced
when designing APIs has a better chance of identifying (in-)appropriate ones –
even if their target isn’t willing to fully cooperate.

This line of thinking recently led me to a more than twenty year old resource
that’s often called a “seminal work” but strangely isn’t cited much in either
Internet governance or API design circles, as far as I can tell: Design Rules
Volume 1: The Power of Modularity by Carliss Y. Baldwin and Kim B. Clark.

Their ambitions were not small:

[W]e want to explain how and why the computer industry changed from a
quasi-monopoly into a large “modular cluster.”1 […] In particular, one of most
the important forces shaping the evolution of these designs was the drive toward
modularity.

For me this was an engrossing read, even though (and perhaps because) a fair bit
is already intuitive to a practitioner.

Chapter 3: What is Modularity? explains concepts like abstraction, isolation,
information hiding, and interface that are well known in industry:

A complex system can be managed by dividing it up into smaller pieces and
looking at each one separately. When the complexity of one of the elements
crosses a certain threshold, that complexity can be isolated by defining a
separate abstraction that has a simple interface. The abstraction hides the
complexity of the element; the interface indicates how the element interacts
with the larger system.

followed by a detailed explanation of ‘how individuals with knowledge can split
apart a large design with many innate interdependencies, and thereby create a
modular design and task structure.’

Chapter 9: Design Options and Design Evolution goes on to consider the economic
impact of modularity:

It is useful to divide the large set of all complex adaptive systems into two
categories: (1) systems in which the so-called adaptive plan is in the hands of
a few agents; and (2) systems in which the adaptive plan is decentralized to
many independent agents.

Yes, decentralisation fits in here too. Then, focusing on the benefits of the
latter category of systems:

Modularization permits individuals (or firms) to mix and match alternative
designs of the modules of a system. The “rights” to mix and match are options
with quantifiable value in the greater economic system. A modularization
multiplies design options and at the same time disperses them so that they can
be “picked up” by many people, without the permission of any central architect
or planner. The pursuit of valuable options by many decentralized actors in turn
accelerates the rate of change of the system as a whole. […] Modularity creates
design options and in so doing can radically change the market value of a given
set of designs.

Chapter 14: The Emergence of Modular Clusters reinforces this:

A modular design makes possible decentralized design evolution. In the presence
of advanced capital markets, a modular design also makes possible decentralized
industry evolution. In other words, when an artifact with a modular design is
created in an economy with advanced capital markets, subindustries of firms and
markets organized around modules may emerge and evolve in parallel with the
module designs themselves.

And then Chapter 15: Competition among Hidden Modules and Industry Evolution
begins:

Modular clusters, by definition, “play host” to modular design evolution. Hence,
unless and until an artifact design has been modularized, there is no cause for
a modular cluster to form. Following a modularization, we have seen, there will
be a concomitant multiplication and decentralization of design options. The
number of workgroups engaged in design (and production) will go up, while,
simultaneously, the forces of transactions and agency costs that tend to bind
workgroups together into firms will diminish. Depending on the balance of these
and other forces acting on the industry, a modular cluster (or clusters) may
then emerge as a viable form for the industry to take.

In other words: yet more support for interoperability through modularity and
decentralization as a remedy to competition and centralization issues – this
time from an economic perspective. I think that’s important, because regulators
have a lot more history with economists than they do with tech folks.

These are, of course, just a few highlights, and there are many more keen
observations throughout; if you find these quotes interesting, I recommend you
read the whole work.

I liked this book because there’s considerable value in having these
observations written down in a well-reasoned, rigorous framework: it’s one thing
if the industry says it does things for particular reasons in its many books
about the topic, but it’s another when it’s done with the appropriate
theoretical context and rigour from the ‘outside.’

I also enjoyed it because it ties together so many of the things I’m currently
interested in: APIs, interoperability, competition, and decentralization.
Critically, there’s also followup work which is even more relevant – but I’ll
write about that separately.

In later interviews, Baldwin has said that the term ecosystem won out over her
modular clusters. ↩


MORE
May 100
18.6.2024, 16:30All Things Distributed (Feed)
Hacking our way to better team meetings
My team and I set out to build a simple note taking aide, which transcribes and
summarizes our meetings using Bedrock. Today, we’re making the proof of concept
available to everyone.

MORE
May 80
24.6.2024, 6:00SEI Blog (Feed)
Polar: Improving DevSecOps Observability
This post introduces Polar, a DevSecOps framework developed as a solution to the
limitations of traditional batch data processing.

MORE
May 70
Today, 17:55ongoing by Tim Bray (Feed)
Storage Churn

What are the highest-impact Cloud services? Storage would be near the top of any
list. Where by “Storage” I mean what S3 does: Blobs-of-bytes storage that is
effectively unlimited in capacity, credibly more durable than anything you could
build yourself, and easily connected to the world, either directly or through a
CDN. I think we’re entering a period of churn where there’s going to be serious
competition on storage price and performance. Which, by the way, is crucially
relevant to the Fediverse.

Let’s start with AWS, since they invented the modern Storage concept. The most
important thing about S3 is this: There appear to be zero credible reports of S3
data loss. Given the number of objects it holds, and the numbers of years it’s
held them, that’s remarkable.

It’s a safe place to store your data. Yeah, the API is a little klunky, and the
latency can be high, and the hardwired bucket/object hierarchy is annoying, and
so are the namespace issues. And it’s not cheap.

But it’s safe. And fast enough to be useful. And safe. And dead easy to connect
up to a CDN. And did I mention that it’s safe?

S3…

AWS, to their credit, aren’t resting on their laurels. Here is a good Register
interview with Andy Warfield, one of the lead S3 engineers and also a really
good person. He’s talking about another variation on the basic S3 service,
called “Express”, which has more filesystem-y semantics, higher performance, but
(reading between the lines) a little less durability? (Also, more expensive.)

What’s notable about S3 isn’t this particular feature, but the fact that AWS
keeps rolling out new ones. So it’s a moving target for the emerging
competition.

…but cheaper…

In recent years and especially over the last few months, alternatives and
competitors to S3 keep crossing my radar. A bunch of them have a premise that’s
essentially “S3-compatible, but cheaper”: Backblaze B2, Digital Ocean Spaces,
Wasabi, IDrive e2, Cloudflare R2, and Telnyx Cloud Storage. I’m sure I’ve missed
some.

…and faster!

Some of the products claim to be way faster. Which matters if it’s true, but so
far I don’t know of any popular benchmarking standards, so I’d take the numbers
with a grain of salt. If I really cared, for a big project, I’d want to try it
with my own code.

Here are a few of those:

S2

See Designing serverless stream storage. This is still more a research project
than a product, but I drop it in here because it says that access to S3 Express
made it possible. Its claim to fame appears to be higher performance.

Tigris

Tigris offers what they describe as “Globally Distributed S3-Compatible Object
Storage”. I think the best description of what that means is by Xe Iaso of
Fly.io, in Globally Distributed Object Storage with Tigris. It’s not just
well-written, it’s funny. Apparently Fly.io bundles Tigris in, with command-line
and billing integration.

Bunny

The fastest object storage, replicated to the edge is their big claim.

CDN?

Bunny sounds like it’s partly a CDN. And it’s not the only one. Which makes
obvious sense; if you want to deliver the stuff you’re storing to users around
the world at scale, you’re going to be hooking your storage and CDN together
anyhow. So those lines are going to stay blurry.

Compatibility and intellectual property

S3 compatibility is an issue. It’s interesting that AWS has apparently decided
not to defend the S3 API as intellectual property, and so these things
cheerfully claim 100% plug-compatibility. And when they don’t have it, they
apologize (that apology looks unusually far under the covers; I enjoyed reading
it).

Durability?

They may claim compatibility, but mostly do not claim equivalent durability.
I’ll be honest; if I were picking one, that would worry me. I’d need to see
pretty full disclosure of how the services work under the covers.

Unknowns

I just mentioned durability, which is a technology issue. The other big unknowns
are about business not technology. First of all, can you sustainably make money
selling storage at a price that undercuts AWS’s? I haven’t the vaguest idea.

Second, is this a threat to AWS? There is a vast amount of data that is never
gonna migrate off S3 because who’s got the time for that, but if the competition
really can save you a lot of money that could hit S3’s growth hard, and Amazon
wouldn’t like that. Who knows what might happen?

Now let’s change the subject.

Fediverse storage

I’ll use myself as a Fediverse example. As I write this, my @timbray@cosocial.ca
Mastodon account has just over 18K followers, distributed across 3K-and-change
instances. So whenever I post a picture or video, each of those instances
fetches it and then keeps its own copy, if only in a short-lived cache.

All these files are immutable and identical. Smell an opportunity? Yeah, me too.
Someone needs to build an object-store/CDN combo (I’ve already heard people say
“FDN”). The API should cater to Mastodon’s quirks. You could split the cost
equally or deal it out in proportion to traffic, but either way, I think there’d
be big cost savings for nearly every instance.

Furthermore, it doesn’t feel technically challenging. If I were still at AWS,
I’d be working on a PR/FAQ right now. Well, except for, since everything is
S3-compatible and CDNs are commoditized, it would be plausible (and attractive)
to build your FDN in a way that doesn’t tie you to any particular infrastructure
provider.

Someone has already started on this; see Jortage Communal Cloud; small as of
yet, but pointing in the right direction.

Fun times!

The storage world is a market with no monopolist, where providers are competing
on price, performance, and durability. Be still my beating heart.


MORE
May 70
22.6.2024, 19:00Platformonomics (Feed)
Platformonomics TGIF #50: May 3, 2024

Platformonomics TGIF is a weekly roll-up of links, comments on those links, and
perhaps a little too much tugging on my favorite threads.

Get Platformonomics Updates By Email

Subscribe

Platformonomics is off the next two weeks (recovering from Clown Week and the Q1
CAPEX numbers). Look for our next missive May 24th.

News Q1 Amazon CAPEX

Amazon is the last of the clouds to report for Q1: CAPEX of $15 billion, up 5%.

The Amazon numbers are not very insightful as they include both the retail
business and AWS, and both have been churning post-pandemic.

AWS was 46% of Amazon’s total CAPEX in 2023 (they only break AWS out annually)
and AWS CAPEX down for the first time ever last year.

The guidance was:

We anticipate our overall capital expenditures to meaningfully increase
year-over-year in 2024, primarily driven by higher infrastructure CapEx to
support growth in AWS, including generative AI.

On the — well, we’re talking about CapEx. Right now, in Q1, we had $14 billion
of CapEx. We expect that to be the low quarter for the year.

Previous:

Follow the CAPEX: Cloud Table Stakes 2023 Retrospective, Follow the CAPEX:
Triangulating NVIDIA, Amazon Desperately Seeks Compression Algorithm for
Experience, Most Greatest AI Company in Entire Universe Makes Major Strategy
Pivot, SEC Embraces Platformonomics ClownWatch, Warns Amazon and IBM, The AWS
Generative AI Soap Opera, The Amazon AI Inferiority Complex

Q1 CAPEX Summary Scorecard (Final)

A reminder: upwards of 80-90% of Google/Meta/Microsoft CAPEX goes to cloud
infrastructure. The comparable Amazon number is likely under 50%.

Platformonomics ClownWatch at some point is going to have to look at Elon’s
three inadequately funded AI efforts (Tesla, Twitter and xAI). Even combined
they’re not in the game. Maybe that will motivate more Elon corporate governance
shenanigans?

Previous:

Q1 CAPEX Scorecard, Follow the CAPEX: Cloud Table Stakes 2023 Retrospective,
Follow the CAPEX: Triangulating NVIDIA, Platformonomics ClownWatch, Follow the
CAPEX: The Clown Car Race Checkered Flag

We Are All Now CAPEX Obsessives

This was the CAPEX quarter! Welcome everyone aboard the CAPEX train! I probably
should sell merch.

Previous:

Q1 CAPEX Scorecard, Follow the CAPEX: Cloud Table Stakes 2023 Retrospective

We Are Not All Arithmetic Obsessives

While seemingly aboard the CAPEX train, the New York Times still has some
arithmetic quality control issues. But they do aspire to be a tech company, so
they have that going for them.

Previous:

Aspiring Technology Company Neither Interested in Technology Nor Understands
Technology, Living in a Glass House: The New York Times On Tech, Living in a
Glass House: What is the New York Times?, Does the New York Times Editorial Page
Ever Talk to the Business Side of the House?

Some (New) CAPEX Obsessives Are Hallucinating Wildly

The importance of CAPEX has become common knowledge. There is no better
indicator than the fact venture capitalists have started doing CAPEX thought
leadership tweets and blog posts in the last week.

One must admire the ability to meld absolute certainty with getting the data
(and conclusions) completely wrong (and bonus points for simultaneously being
the author of “Winning with Data”).

Cloud Repatriation is Going Just Great: Q1 2024

The cloud repatriation narrative says companies are furiously moving workloads
back to their own data centers from the cloud. Between the tailwinds of
repatriation and generative AI, you’d think the biggest data center operators
would be reporting blow out numbers. But they’re not.

Digital Realty Trust reported a revenue decline of 3% and tried to hide it with
an anemic earnings day “partnership” announcement with noted CAPEX clown Oracle
(you may want to be seated when you read this game-changing news):

As part of this collaboration, Oracle will deploy critical GPU-based
infrastructure in a dedicated Digital Realty data center in Northern Virginia.

Fellow data center giant Equinix can’t even provide a date for reporting Q1
earnings after serious fraud allegations.

With Equinix on the floor in a fetal position, we are unable to update the
Platformonomics Repatriation Index. But if Equinix managed 10% growth for the
quarter, the index would still be at an all-time-low (i.e. the lowest evidence
of repatriation).

Data centers are not the slam dunk investment everyone thinks they are. There
will be blood.

Previous:

Cloud Repatriation is Going Just Great: Continued Radio Silence, Cloud
Repatriation is Going Just Great: Radio Silence from Equinix, Cloud Repatriation
is Going Just Great: Equinix, New Cloud Repatriation Hypothesis, Cloud
Repatriation Where Art Thou?, Cloud Repatriation: The Search Continues, Cloud
Repatriation: Still Not a Thing, Platformonomics Repatriation Index – Q1 2023:
Surf’s Up?, Private Equity in Action: Data Centers, Oracle Moves to Nashville,
Should I Stay or Should I Go

Wrapping Oneself in the Flag: Seattle Times Edition

Just putting a marker down on this apparently serious proposal from the Seattle
Times, a legacy provincial newspaper turned web site aggregator. Declaring
themselves synonymous with democracy, they ask to be made a ward of state, which
somehow will help a free press flourish. Note the lack of details about the
legislation they are requesting.

I am giving them a chance to persuade me they don’t actually mean this, but if
they do, I have thoughts. Many thoughts.

EU Announces Meta Struggle Sessions

The struggle sessions will continue until Europe has a tech industry (so don’t
hold your breath).

Previous:

UK Announces Big Tech Struggle Sessions “For The Children”, EU Announces DMA
Struggle Sessions

Reminder: The European Union is an Advanced Persistent Threat

Previous:

BREAKING: European Union Designated an Advanced Persistent Threat

Feature Request: Podcast Players

Can we please have a new podcast category: “VCs Shamelessly Talking Their
Books”?

Maybe also a category for “VCs Shamelessly Talking Their Books from the AI
Sidelines”?

You could persuade me there is a lot of overlap between those two.

And, of course, provide an option to hide these categories.

Previous:

The Most Important Issue in AI, The VC Antitrust Doctrine, VC-induced
Indigestion, The Looming Layoffs for Venture Capitalists, Ghost VCs


MORE
June 60
1.5.2024, 16:56James Governor's Monkchips (Feed)
Monki Gras 2024: Wrapped

It’s always difficult, if not impossible, to sum up a conference that you’ve
recently run. But it’s important to write the post, because there are some
important jobs to be done – most notably thanking a bunch of people.

The main thing I want to say is that it was really, really, good to be back in
2024. Monki Gras is a labour of love, but the hard work is always worth it. The
event is unique in the community we sustain, and the approach we take to things.
It feels important – being an expression of so many of the things Redmonk holds
dear – most notably inclusion, kindness, and great story-telling. Like its
sister conference Monktoberfest, at Monki Gras we want people to feel inspired,
and, hopefully, to remember just why it is that they do what they do. We’re not
about the hustle, the hockey stick, and the PLG, but rather the craft, the
learning, and the teaching. We’re about highlighting the positive social aspects
of this industry we’re so privileged to work in.

With all that in mind it might seem counterintuitive that I should have made
prompt engineering and generative AI as my theme this year – Prompting Craft.
After all, while AI is exciting in the possibilities it opens up, it’s also a
little scary. A lot of the intellectual property used to train large language
models (LLMs) has been used without any consideration of copyright. What is
more, many creatives, including software developers, feel threatened by the rise
of the machines. I definitely don’t want to minimise people’s fears or concerns.
It’s going be very very uncomfortable. But it feels like the wave will break
whether we like it or not. Also Chat-GPT is really really bad at generating
vector graphics.

One answer to the tension about AI fears came from one of my speakers at the
event – Dr. Cat Hicks, founder of the Developer Success Lab. In its research
into the factors driving developer productivity, personal, and organisational
enablement the Lab examined AI as a potential threat to people. The research
found that the organisations and individuals most threatened by AI had an
adversarial culture – a culture of “brilliance”, where individuals constantly
feel they have to prove themselves, where coding is competition. Organisations
that are ready to embrace the possibilities of generative AI on the other hand,
have already fostered a culture of collaboration and shared learning, mutual
support and teamwork. For these organisations AI is just another way to work
effectively together.

I don’t plan to summarise all of the talks here but did want to mention Cat’s
talk because of the reasons outlined above. The talk was very human, and yet it
was about AI.

The human machine interface was a meta theme of the conference – prompting is
just another user interface, and it’s inherently social. We prompt each other to
get what we want and need. We prompt our children. We prompt our friends and
colleagues. We have to build trust with LLMs before we are comfortable making
more use of them. We can trick machines, just as they can trick us – security
specialists are not going to be out of a job any time soon…

So about those thank yous.

Firstly a shout to our speakers on day one – Richmond Alake, Alex Chan, Patrick
Debois, Cat Hicks, Zack Akil, Mathis Lucka, Luke Marsden, Farrah Campbell &
Brooke Jamieson, Rafe Colburn, Afsha Hossain and Matt Webb.

Day two was also excellent – thanks to Dormain Drewitz & Rachel Stephens, Ian
Miell, Julia Ferraioli, Jim Boulton, Kristen Foster-Marks, Paul Molin, Jessica
West, Kyle Roche, Emil Eifrem and Paige Bailey.

So much respect to attendees went into the prep and hard work by these speakers.

While I may not have summarised the days’ talks, my colleague Rachel Stephens
did a bang up job doing exactly that. Here are writeups of day one and day two.

Our Monki Gras sponsors were also critical in making the event a success. My
humble thanks to AWS, Civo, Deepset, CNCF, Neo4j, MongoDB, Akamai, Griptape,
PagerDuty, Screenly and Betty Junod for supporting us. A special thanks to Mark
Boost, CEO of Civo, for all his sterling efforts helping out with our venue –
Civo Tech Junction, a new space hosting free meetups and events in Shoreditch.

Last, and certainly not least, I need to thank the Monki Gras team – namely
Jessica West, Dan McGeady, and Rob Lowe. I couldn’t have done it without their
sterling efforts.

Other great write-ups include these by Sinead Doyle and François Hoehl and Simon
Haslam.

I will sign off with a quote from Patrick Debois which made me very happy.

“I have not felt this energised after a conference in at least years. I’ll be
back Monki Gras next time for sure!”


MORE
May 10
Today, 17:55ongoing by Tim Bray (Feed)
Photointegrity

In March of 2004, just over twenty years ago, I published an ongoing piece
entitled, like this one, “Photointegrity”. The issue remains the same, but the
rise of AI increases its importance and its difficulty. Here are words on the
subject, illustrated by photos all of which have been processed with AI
technology.

Tulip blossom, captured with twenty-year old analog technology, enhanced with
AI.

There’s an amusing story about the technology behind these flower pictures, down
at the bottom of this piece.

Back in 2004

I was already using Photoshop but in fully-autodidactic mode, so I thought I
should read a book about it, and selected one by Scott Kelby, “The Photoshop
guy” back then and still active in the space, two decades later. It was a good
book, but I was left wide-eyed and shocked: I’ll quote from that piece for those
of you who don’t want to step back twenty years in time and read it:

Personal Improvement

In particular, Kelby walks through an astounding list of techniques for
improving portraits, and I quote: removing blemishes, removing dark circles
under the eyes, lessening freckles or facial acne, removing or lessening
wrinkles, colorizing hair, whitening the eyes, making eyes that sparkle,
enhancing eyebrows and eyelashes, glamour skin softening, transforming a frown
into a smile, doing a digital nose job, slimming and trimming, removing love
handles, and finally slimming buttocks, arms and thighs.

Integrity?

Screw it, integrity is history. The image is no longer the capture of a
instant’s light and colour, it’s, well… whatever you and Photoshop make of it.

Photointegrity

I proposed a definition at the time: “what I’m going to do is strive to balance
Truth and Beauty. In practical terms, this means the goal is make the picture
look as much as possible like what I saw, as opposed to as good as possible.”

Simple yellow flower, captured with twenty-year old analog technology, enhanced
with AI.

I can’t claim that I follow that strictly; most of the pictures in this space
come out of the camera looking less pleasing than what I remember seeing, but I
will confess that the version you see is often prettier than that memory.
Usually, that results from the application of a bunch of Adobe technologies.

Is that OK? It’s a judgment call. Is there anything that isn’t a judgment call?
Funny you should ask, because Adobe just announced the Firefly Generative AI 3
model, around which the next version of Photoshop is being built. Hitting those
links and just scrolling through the pictures will give you a feeling for what
this software will do.

Let me put a stake in the ground. I believe these things:

If you use generative tools to produce or modify your images, you have abandoned
photointegrity.

That’s not always wrong. Sometimes you need an image of a space battle or a
Triceratops family or whatever.

What is always wrong is using this stuff without disclosing it.

The C2PA angle

Last October, I wrote up C2PA, a useful digital watermarking technology that can
be used to label images and video. That piece’s predictions look like they’re
coming true; several manufacturers have announced C2PA support. I’m not going to
take the space here to describe C2PA again.

I do note that Photoshop already supports C2PA and when it writes a watermark
saying “Edited with Photoshop”, that label includes a very few words about what
it did: cropping, levels adjustment, and so on; no details.

I believe strongly that when people use Adobe’s Firefly generative AI to create
or augment pictures, Photoshop should by default turn C2PA labeling on, and
disclose in the watermark whether it is fully-generated or just augmented. Sure,
the person generating the image can always take that watermark out, but they
can’t change its contents, and assuming C2PA becomes ubiquitous, the absence of
a watermark would be reasonable grounds for suspicion.

Fruit tree blossoms, not open yet, captured with twenty-year old analog
technology, enhanced with AI.

AI + photointegrity?

Over the last couple of years, the way I use Adobe Lightroom has changed a whole
lot, and it’s mostly because of AI. Specifically, smart select. Lightroom now
offers Select functions for Subject, Background, Sky, and Object. There’s also a
very useful “Duplicate and invert” for any selection. I use these for almost
every photo I take, especially Select Sky. The amount of light in the sky
differs from that down here on the surface, and I’m pretty sure that our eyes
compensate for that. Almost every picture looks more “real” when you select the
sky and dial the brightness down (rarely: up) a touch, and maybe bump the
contrast a bit.

This photo would have been a complete failure without those tools.

Allyson’s parents speak to the crowd at her memorial.

Recently we were at a memorial social for our recently-deceased friend Allyson.
It was on a rooftop, on a bright grey day; the volume of light coming out of the
sky was insane, and kept turning my photographic subjects into dark silhouettes.

The photo of Ally’s parents addressing the crowd is not great (her mom’s eyes
are closed) but it at least captures a moment. The original was totally
unusable, because the subjects are under a canopy and thus shaded, while the sky
and cityscape and even mountains were reflecting harshly. So you select the
subject, you invert and duplicate, you add light to the subject and subtract
from the rest, and you get something that looks exactly like what I saw.

Of course, this depends on a good camera with a lot of dynamic range that can
fish detail out of shadows.

I think this process retains photointegrity.

AI-enhanced analog

What happened was, the sun came out after the rain, everything is blooming this
time of year, and I wanted to take pictures. I was rummaging for lenses and
there was this dark shape at the back of the shelf. “What’s that?” I thought. It
turned out to be an old Pentax with “Macro” in its name. Just the thing! Here’s
what the combo looks like.

By the way, one reason the Internet is still good is that random enthusiasts
maintain obscure databases, for example of camera lenses, from whence this smc
Pentax-D FA 100mm F/2.8 Macro, an alternate version of which rejoices in the
name “Schneider-Kreuznach D-Xenon”. It seems to have been manufactured only
around 2004. I wrote about buying it in 2011 and shooting flowers and dancers
with it in 2014; lotsa groovy pix in both.

Anyhow, this lens does a fabulous job of isolating foreground and background.
Given this to chew on, Lightroom’s AI gizmo does a fabulous job of selecting
just the flower (or background). So it’s easy to sharpen the flower and fade the
bokeh; the old lens and the bleeding-edge software were made for each other.

But I digress.

Photointegrity matters

It mattered in 2004 and it matters more every passing year as our level of trust
in online discourse falls and the power of generative AI grows. We have the
tools to help address this, but we need to think seriously, and use them when
appropriate.


MORE
April 300
24.6.2024, 6:00SEI Blog (Feed)
What Recent Vulnerabilities Mean to Rust
In recent weeks several vulnerabilities have rocked the Rust community causing
many to question its safety. This post examines two such vulnerabilities.

MORE
April 290
Today, 9:19mark nottingham (Feed)
No One Should Have That Much Power

It’s a common spy thriller trope. There’s a special key that can unlock
something critical – business records, bank vaults, government secrets, nuclear
weapons, maybe all of the above, worldwide.

Our hero has to stop this key from falling into bad people’s hands, or recover
it before it’s too late. Perhaps at one point they utter something like the
title of this post. You walk out of the theatre two hours later entertained but
wondering why someone would be silly enough to create such a powerful artefact.

In a surprising move, law enforcement officials are once again calling for such
a thing to be created. Repeatedly.

These authorities and their proxies say that they must have access to encrypted
communications to keep us safe. They have been doing so for years – at first
bluntly, now in a more subtle way. Encryption backdoors aren’t politically
viable, so they take pains to say that they don’t want them while at the same
time asking for a level of access that cannot be achieved except through
backdooring encryption.

If you create a way to recover messages sent through a service, that’s a
backdoor. If you run some code that evaluates messages on the endpoints and
flags them if they meet some criteria, that isn’t an improvement; it’s a
backdoor that can be abused in myriad ways. Centralising access to encrypted
content creates unavoidable systemic risks.

This means that any such mechanism has to be handled like weapons-grade
plutonium: losing control is a disaster of epic (or even existential)
proportions. The few national governments who have nuclear capability struggle
greatly to manage that risk; why would we intentionally entrust something as
powerful to every government in the world or potentially even every local police
department? Or will it be just a privileged few governments that will have
access?

The current crop of suggestions seem to concede that governments shouldn’t have
direct access. Instead, they want services to backdoor themselves and act as
gatekeepers to law enforcement. That’s not an improvement; it’s still
centralized, and it makes these companies responsible for any misuse of the data
that they have access to, requiring everyone on the planet to trust a few big
tech companies with our private and most intimate conversations – hardly a
direction that society wants to go in in 2024. ‘Trust me, I’m in charge’ is a
poor model of governance or security.

These ‘solutions’ also ignore the reality that the ‘bad guys’ will just use
other tools to communicate; information is information. That will leave law
abiding people giving up their privacy and security for little societal gain.

Law enforcement has more power than ever before because of digital technology.
They are able to collect, process, summarise and track much more efficiently and
at much greater scale. Genuinely new insights and capabilities are possible. So,
when they want access to encrypted data because things have ‘gone dark’, it’s
reasonable to ask ‘as compared to what?’

No one should have that much power, because messaging and other encrypted
services have become people’s memories, their casual hallway chats, their
intimate whispers. Yes, there is longstanding legal precedent for searching
someone’s papers and home, but the barriers to doing so are considerable – not
just those imposed by law, but also physics. There are few such inherent limits
on a key that can trivially enable access to what amounts to anyone’s mind or
identify anyone who thinks about a particular topic. Law enforcement struggles
to solve real and serious problems, but the power they’re asking for is too vast
and too easily misused, and they are failing to appreciate how it would operate
on a global Internet.

One of the assumptions built into these calls is that if the tech community
would just nerd harder, a solution could be somehow magically found that
preserved privacy and security while letting the ‘good guys’ have access. With
all respect to the valuable work that law enforcement does to protect society,
it’s equally as valid to ask them to just police harder.


MORE
May 10
22.6.2024, 19:00Platformonomics (Feed)
Platformonomics TGIF #49: April 26, 2024

Platformonomics TGIF is a weekly roll-up of links, comments on those links, and
perhaps a little too much tugging on my favorite threads.

Get Platformonomics Updates By Email

Subscribe

Busy week with first quarter earnings reports — lots of CAPEX!!!

My Writings European Union Designated an Advanced Persistent Threat

Previous:

BREAKING: European Union Designated an Advanced Persistent Threat

News Q1 CAPEX Scorecard

The CAPEX numbers are rolling in:

Google: $12.01B +91%, new all-time-high! Meta: $6.72B, -5%, but strong CAPEX
guidance sinks stock! Microsoft: $14B +79%, new all-time-high!

Amazon reports next week.

Previous:

Follow the CAPEX: Cloud Table Stakes 2023 Retrospective

Q1 Google CAPEX

Good trend — Google may finally have awoken:

Guidance (implies CAPEX could surpass $50 billion in 2024):

Looking ahead, we expect quarterly CapEx throughout the year to be roughly at or
above the Q1 level

…most nearly all, I should say, of the CapEx was in our technical
infrastructure. We expect that our investment in office facilities will be about
less than 10% of the total CapEx in 2024, roughly flat with our CapEx in 2023,
but is still there.

Previous:

Follow the CAPEX: Cloud Table Stakes 2023 Retrospective, Follow the CAPEX:
Triangulating NVIDIA

Q1 Microsoft CAPEX

Comparison with Google (an even better trend):

Notable:

Currently, near-term AI demand is a bit higher than our available capacity.

Guidance:

We expect capital expenditures to increase materially on a sequential basis
driven by cloud and AI infrastructure investments. As a reminder, there can be
normal quarterly spend variability in the timing of our cloud infrastructure
buildout and the timing of finance leases. We continue to bring capacity online
as we scale our AI investments with growing demand.

Therefore, we now expect full-year FY24 operating margins to be up over 2 points
year-over-year even with our cloud and AI investments, the impact from the
Activision acquisition, and the headwind from the change in useful lives last
year.

To scale to meet the growing demand signal for our cloud and AI products, we
expect FY25 capital expenditures to be higher than FY24. These expenditures over
the course of the next year are dependent on demand signals and adoption of our
services, so we will manage that signal thru the year. We will also continue to
prioritize operating leverage and therefore, we expect FY25 operating margins to
be down only about one point year-over-year, even with our significant cloud and
AI investments as well as a full year of impact from the Activision acquisition.

Impressive:

Free cash flow grew faster (+21%) than earnings (+20%), despite the big bump in
CAPEX.

Previous:

Follow the CAPEX: Cloud Table Stakes 2023 Retrospective, Follow the CAPEX:
Triangulating NVIDIA

Q1 Meta CAPEX

We haven’t historically tracked Meta because they’re not a cloud provider. But
they are one of the biggest CAPEX spenders ($28 billion in 2023) and are
accelerating that spend for generative AI. Huge props to Meta for tanking their
stock 20% by upping their CAPEX guidance by $5 billion for 2024! We revere
CAPEX-first companies!!!

Meta didn’t do a very good job explaining how they will make money from their
increasing CAPEX investments. As described in a recent Zuckerberg interview,
Meta is much more comfortable spending now and figuring it out later than Wall
Street. But makes one wonder if Meta doesn’t get into model hosting for third
parties in the future, making them more of a cloud competitor.

One explanation for the planned CAPEX bump:

Guidance:

We anticipate our full year 2024 capital expenditures will be in the range of
$35 billion to $40 billion, increased from our prior range of $30 billion to $37
billion as we continue to accelerate our infrastructure investments to support
our AI roadmap. While we are not providing guidance for years beyond 2024, we
expect CapEx will continue to increase next year as we invest aggressively to
support our ambitious AI research and product development efforts.

Notable — talking about CAPEX and energy as peers:

As we’re scaling CapEx and energy expenses for AI…

Previous:

Follow the CAPEX: Triangulating NVIDIA, Cloud Power Up

Related:

Russian court sentences Meta spokesperson to six years in absentia (a badge of
honor)

Meanwhile, at the Front of the GPU Queue Hashicorp’s Acquisition by IBM

Some quick thoughts:

Congratulations to all my friends at Hashi. $6.4 billion! The consolidation game
is afoot and will continue — we still have too many sub-scale software
companies. I feel slightly responsible as I suggested Hashi was a better
acquisition than Red Hat five years ago. I also at that time expressed
skepticism about IBM’s bombastic claim that acquiring Red Hat “changes
everything about the cloud market”. Narrator: “it didn’t”. Hopefully Hashi can
avoid the fate of other IBM acquisitions (company slogan: Where software goes to
die). We will however reserve the headline “IBM Makes a Hash of Hashicorp”
should it be needed in the future. The most interesting question is does IBM
revert Hashi’s recent license change and try to undo the OpenTofu fork. IBM has
been on its own quest to tighten up access to ostensibly open source RHEL so
will be interesting to see how they play this.

And, just to point out the obvious, a Hashicorp acquisition doesn’t address
IBM’s fundamental issues:

Previous:

Very Cold Take on IBM, Red Hat and Their Hybrid Cloud Hyperbole, SaaS Sprawl and
the Potential for Market Consolidation

Related:

On IBM acquiring HashiCorp

Platformonomics ClownWatch: IBM

The OG CAPEX clown nailed the quarter with CAPEX declining 21% to a barely
discernible $361 million (about 2.5 days of Microsoft CAPEX spend), while
continuing to pretend they are an AI company (and they’re already pretending
Hashi is an AI thing). IBM retains its position at the top of the ClownWatch
list.

Previous:

Platformonomics ClownWatch, Follow the CAPEX: The Clown Car Race Checkered Flag,
Introducing Platformonomics ClownWatch, IBM and the Art of Misleading Investors,
This Week in Securities Fraud, Clown CAPEX, IBM’s “Cloud” Business (or Lack
Thereof), IBM: How Much Longer Do the Good Times Last?, IBM’s Lost Decade

Platformonomics ClownWatch: Tesla

Elon says Tesla should be valued as an “AI robotics company” and not a car
company. People have thoughts on what that would mean financially:

But if we turn the all-seeing-eye of CAPEX to Tesla, we find:

The claimed spend of $1 billion in “AI infrastructure CAPEX” is a suspiciously
round number. Almost IBM-esque. But it is also Dr Evil-esque relative to what
others are spending on AI infrastructure. Even noted CAPEX clown Oracle spends
more than that on CAPEX. Telsa spent $2.8 billion in total CAPEX in Q1 (and
we’re under the impression they have manufacturing plants and stuff). The
previous four quarters were all over $2 billion so the $1 billion claim looks
more like marketing than a big leap in investment. Tesla has never spent more
than $9 billion in CAPEX annually. Telsa’s CAPEX is less than 3% of revenue,
which is about what Microsoft and Oracle used to spend when they were pure
software companies. Google and Microsoft now spend double digits of revenue.
UPDATE: that 3% number was erroneously quarterly CAPEX over annual revenue.
Tesla’s CAPEX as a percentage of revenue this quarter was 13%. But again, they
have manufacturing plants and stuff.

So I am delighted to announce Tesla has been placed on Negative Watch. Further
disparities between their AI rhetoric and their cold, hard CAPEX spend may
result in the company being designated a full-fledged AI Clown.

And it looks like xAI may shortly be joining the list ($6 billion doesn’t take
you very far in the frontier AI world):

Previous:

Platformonomics ClownWatch

Software Migration Alert: DarkTrace

When private equity comes amalgamating, it is time to start migrating.
Especially when it comes to cybersecurity…

Previous:

Software Migration Alerts, Perfidious Private Equity

McKinsey: It Might Just Be Happening

Criminal convictions can lead to the death penalty.

Previous:

A Call for the Corporate Death Penalty, McKinsey Karma?, McKinsey a Source of
Bad Banking Advice (Shocking!)

Related:

McKinsey: Last Week Tonight with John Oliver, Red Hat tries on a McKinsey cap in
quest to streamline techies’ jobs

Oracle Moves to Nashville

In the future, every city will be Oracle headquarters for 15 minutes.

And we have an exclusive photo of the entire Oracle Cloud headquarters region
being migrating to Nashville:

Previous:

Follow the CAPEX: The Clown Car Race Checkered Flag

Cloud Repatriation is Going Just Great: Continued Radio Silence

Previous:

Cloud Repatriation is Going Just Great: Radio Silence from Equinix, Cloud
Repatriation is Going Just Great: Equinix, New Cloud Repatriation Hypothesis,
Cloud Repatriation Where Art Thou?, Cloud Repatriation: The Search Continues,
Cloud Repatriation: Still Not a Thing, Platformonomics Repatriation Index – Q1
2023: Surf’s Up?

Related:

What Happens When Hyperscalers And Clouds Buy Most Servers And Storage?

Interesting If True So You Want to Build an AI Company: Episode 3

Their administrative bloat is your opportunity.

Previous:

So You Want to Build an AI Company: Episode 2, So You Want to Build an AI
Company


MORE
June 60
Today, 14:28Phil Windley's Technometria (Feed)
Using X.509 Certs for DID Provenance

Summary: The abundance of X.509 certificate authorities who already perform
identity proofing for businesses provides a rich resource that can be leveraged
to boot the verifiable data ecosystem.

When you used a verifiable credential to prove something about yourself, the
verifier can know cryptographically: (1) the identifiers for the issuer, (2) the
credential hasn't been tampered with, (3) the credential was issued to you, and
(4) the credential hasn't been revoked. These four checks are important because
their establish the fidelity of the data being transferred. They don't, however,
tell them whether they can trust the issuer. For that, they need to take the
issuer's decentralized identifier (DID) that they got from credential
presentation and determine who it belongs to.

At the most recent Internet Identity Workshop, Drummond Reed gave a session on
how X.509 certificates could help with this. The first step, like always, is to
resolve the DID and retrieve the DIDDoc that associates keys and endpoints with
the DID. The endpoint can be an HTTP server and, of course, should have an X.509
certificate providing TLS security. That certificate, at the very least, has a a
domain name to bind that to the certificate's public key. It can, if you pay for
the feature, also include information about the entity that applied for the
certificate. The certificate authority proofs that information and is vouching
for it when they sign the certificate.

The key to making the X.509 certificate useful for checking the provenance of a
DID lies in one key change. X.509 certificates can contain and extended field
called a Subject Alternative Name. This following figure shows how it can help.

Using X.509 Certificates to establish the owner of a DID (click to enlarge)

In this figure:

The issuer (Attestor) creates the DID they will use to issue the certificate
along with its associated DIDDoc, including an HTTP endpoint for DID
verification. Attestor applies for a X.509 certificate for that endpoint,
including in the application the DID they created in (1). The certificate
authority does it's usual proofing of the application and issues a certificate
that includes the DID in the Subject Alternative Name field. The issuer creates
a credential definition in the usual way that includes their DID and writes it
to whatever Verifiable Data Registry their DID method dictates. Attestor issues
a credential to a holder (Alice) using that credential definition. At some later
time, Alice presents the credential to the verifier (Certiphi). Certiphi
resolves the DID to get the DIDDoc and retrieves the verfication endpoint from
the DIDDoc Certiphi retrieves the certificate for that endpoint1. Certiphi
verifies the certificate by checking it's signature and ensures that the DID in
the DIDDoc for the credential matches the one in certificate.2

The issuer's DID has now been tied in a verifiable way to whatever information
is in the certificate. Provided the certificate includes information about the
entity beyond the domain name, the verifier can use that information to
determine whether or not the credential is authentic (i.e., issued by who the
credential definition purports issued it). That might be all the evidence they
need to determine whether to trust the entity. Certificate authorities could
also issue verifiable credentials to the customer attesting the same verified
claims—after all, it's one more product they can offer.

The benefit of doing issuer validation using X.509 certificates is that there
are already many trusted X.509 certificate authorities in business who already
do proofing of attributes about businesses. That's a huge chunk of the
verifiable data ecosystem that doesn't need to be built because it can be
leveraged. To make this work, digital certificate authorities would need to
start offering to validate DIDs and include them in a certificate as a Subject
Alternative Name. I don't discount that this will take some bureaucratic
maneuvering. Certificate authorities will need to see a business opportunity.
I'd love to see Digitcert or someone do a pilot on this.

Notes Note that this step might be combined with the previous step if the
Verifiable Data Registry is the same server as the endpoint, but that's not
necessarily going to be the case for a number of reasons. Note that this does
not create a call back wherein Attestor can determine which credential was used,
preserving the privacy of the presentation. Attestor does know one of its
credentials has been presented to Certiphi. If this information leakage bothers
you, then any web-based DID method is potentially a problem.

Tags: identity ssi decentralized+identifiers verifiable+credentials x.509


MORE
April 250
Relationships are Entangled

Summary: People aren't digitally embodied. As a result, their online
relationships are anemic. As we think about the kinds of digital identity
systems that will help people live authentic online lives, we must build
identity systems that give people tools to operationalize rich digital
relationships.

Identity is the ability to recognize, remember, and react to people,
organizations, systems, and things. In the current web, companies employ many
ponderous technological systems to perform those functions. In these systems, we
are like ghosts in the machines. We have "accounts" in companies' systems, but
no good way to recognize, remember, and react to them or anyone else. We are not
digital embodied.

One of the great benefits of embodiment is the ability to form and
operationalize rich digital relationships. I've written a lot about the nature
of digital relationships.

Relationships and Identity Authentic Digital Relationships Ephemeral
Relationships Operationalizing Digital Relationships Relationships in the
Self-Sovereign Internet of Things The Architecture of Identity Systems Are
Transactional Relationships Enough? Fluid Multi-Pseudonymity

One of the discussions at VRM Day caused me to think about a feature of digital
relationships I hadn't considered before. Someone said that if you think about a
graph with people (or things, organizations, and so on) as the nodes, the
relationships are the edges, like so1:

A single, bi-directional relationship (click to enlarge)

In this figure Alice and Bob have a bi-directional relationship. This is how
I've normally thought about it and how I'd have drawn it. But in today's
discussion, someone said that the relationship is shared and that Alice and Bob
both control it. But I realized that viewpoint is too simple. Specifically,
Alice and Bob each have a different perspective of that relationship and will
use it separately.

For example, imagine that Alice is the cashier at a grocery store and Bob is a
customer. Alice gives great service, so Bob seeks her out when he shops. Alice
on the other hand has no particular recollection of Bob from encounter to
encounter. For Alice, the relationship is ephemeral, but for Bob, it's longer
term. The nature of each relationship is different. So, we might look at it like
this:

Two uni-directional relationships (click to enlarge)

But after discussing it some more, I realized that these relationships aren't
independent. They're entangled like this:

Entangled relationships (click to enlarge)

In the example I gave above, as Bob seeks out Alice more and more, Alice might
come to recognize him and call him by name, changing the nature of her
relationship with Bob. And that may influence the nature of Bob's relationship
with Alice. Over time, these interactions influence both relationships. So,
while Alice and Bob both have control over their relationship with the other,
actions by one influence the other.

I frequently say that we don't build identity systems to manage identities, but
rather to manage relationships. The problem with contemporary identity systems
is that they are all one sided, controlled by one party—almost always a company.
As I've said before, people are not digitally embodied and thus have no good way
to manage their online relationships. As we strive to build better digital
identity systems, I think it's paramount that we build systems that provide
people with tools that embody them and provide them with the ability to
operationalize their online relationships. These are more than decentralized;
they are self-sovereign.

Notes Peer decentralized identifiers (DIDs) are a great technology for creating
bi-directional relationships.

Tags: identity relationships


MORE
April 250
Today, 17:55ongoing by Tim Bray (Feed)
Mobile Typing Pain

I ran a Fediverse poll asking how people go about entering text on mobile
devices. The results shocked me: Half the population just taps away. Do you?
Read on for details and speculation.

This ongoing fragment embeds links to previous fragments, because I’ve been
worrying and writing about this problem for a long time. Which in itself is
interesting, more evidence that the problem is hard.

The poll post and (long) chain

of responses are here on Mastodon.

People care

First, 3.5K poll responses is more than I usually see on the Fediverse; evidence
that plenty of people have feelings about this. To reinforce that impression,
scroll down through the responses (there are dozens). Many say, essentially,
“Entering text on a mobile device is too hard, so I don’t.”

I’m one of those; I regularly start entering a message into a phone, stop, get
up, and walk across the room to a real keyboard.

Tap tap wow

I widened my eyes when I saw that half the respondents testify to tapping along
letter by letter. I could never. But then I have big fat farmer’s fingers with
soft ends, and am not terribly dextrous.

But, probably, I shouldn’t have been surprised; 21 years ago in this blog I
remarked that “it's pretty impressive to watch a Japanese person pounding text
into their PDA at high speed using just their thumbs.” And today I watch teenage
digits dance on devices like maddened maenads; they seem not to find it tedious.

Swiping etc

A quarter of poll respondents reported swiping words into their phones.

I mentioned above that people have been working on this for a long time. Check
out this progress report from ongoing in 2011. It’s worth noting that Android’s
input method being replaceable was important in driving this innovation.

My own proposal, the Digitator, has failed to break through.

That piece concludes “Anyhow, I’m pretty sure that something will come along.”
But on the evidence it hasn’t, really.

The AI angle: Auto-predict and voice

The least popular poll options were tap-plus-autopredict and voice. I guess I’m
weird, because those are what I mostly use. I suspect many others should too but
don’t, probably because they tried those things a while ago and haven’t
revisited them recently.

In my experience (which, granted, is almost all on Google Pixel devices) the
autopredict and voice options have gotten stronger with almost every release.
Not just a little bit stronger, either. Perhaps it’s just because I’m the white
male Anglophone “canonical human” that designers build for, but I get
dramatically better results than I used to.

Now obviously, most reasonable people will only talk to their phone when they’re
in a private place, which limits the use of that option. But if you can find
privacy, the voice option is getting remarkably good.

Which is to say, I can enter message or email text in at a pace that is
sometimes adequate. Do I enjoy doing this? No, I hate it, as I noted above, and
will make a real effort to switch to a keyboard.

In particular if what I want to enter is important, might matter.

Because anything that matters deserves editing, and it’s rare indeed that I hit
“Send” on a first draft. And while brute-force text entry is edging into
adequacy, editing remains a pool of pain.

Subtext

Two and a half decades into this millennium, the most popular communication
products are optimized for consumption and barely adequate for creation. If I
were paranoid and cynical, I might suspect that this is no accident. Oh wait, I
am. But in fact I think it’s just a hard problem.


MORE
April 250
22.6.2024, 19:00Platformonomics (Feed)
BREAKING: European Union Designated an Advanced Persistent Threat
European Union Designated an Advanced Persistent Threat Threat Actor Seeks to
Degrade
Global Internet

APRIL 22, 2024 – The European Union has been designated an Advanced Persistent
Threat (APT) group for its systematic efforts to undermine cybersecurity and
fundamental capabilities of the Internet around the globe.

Advanced Persistent Threat groups are sophisticated adversaries that use
multiple attack vectors to pursue malicious objectives over an extended period.

The European Union threat actor has repeatedly attempted to undermine Internet
privacy, security, and utility. Attacks include efforts to backdoor encryption
systems, shatter the iOS security sandbox, prohibit end-to-end secure messaging,
arbitrarily censor “unwanted” digital speech, and preemptively restrict the use
of generative AI. The European Union has aggressively tried to expand its
attacks beyond Europe, and often masks its attacks as purported privacy and
security protections.

“The European Union isn’t actually very advanced – they really have no clue
about technology or how it works – but they are extraordinarily persistent in
their efforts to undermine digital privacy and security not just in Europe, but
globally.”

The European Union got its start as a threat actor with denial-of-service
attacks through ineffectual and unnecessary cookie consent dialogs. Global
acquiescence to that GDPR attack emboldened the threat actor to pursue a wide
variety of additional attack vectors (aka “the Brussels Effect”).

In accordance with APT group naming conventions, the European Union threat actor
has been dubbed “les fromages qui puent” (“Stinky Cheese”).


MORE
April 220
24.6.2024, 6:00SEI Blog (Feed)
The Great Fuzzy Hashing Debate
This post details a debate among two researchers over whether there is utility
in applying fuzzy hashes to instruction bytes.

MORE
April 220
Today, 9:19mark nottingham (Feed)
Considerations for AI Opt-Out

Creating a Large Language Model (LLM) requires a lot of content – as implied by
the name, LLMs need voluminous input data to be able to function well. Much of
that content comes from the Internet, and early models have been seeded by
crawling the whole Web.

This now widespread practice of ingestion without consent is contentious, to put
it mildly. Content creators feel that they should be compensated or at least
have a choice about how their content is used; AI advocates caution that without
easy access to input data, their ability to innovate will be severely limited,
thereby curtailing the promised benefits of AI.

The Policy Context

In the US, the Copyright Office has launched an initiative to examine this and
other issues surrounding copyright and AI. So far, they have avoided addressing
the ingestion issue, but nevertheless it has come up repeatedly in their public
proceedings:

“The interests of those using copyrighted materials for AI ingestion purposes
must not be prioritized over the rights and interests of creators and copyright
owners.” – Keith Kupferschmid, Copyright Alliance

“Training of AI language models begins with copying, which we believe has
infringed our copyrights and has already deprived us of hundreds of millions of
dollars in rightful revenues.  The additional violation of our moral right of
attribution makes it impossible to tell which of our works have been copied to
train AI and thus frustrates redress for either the economic infringement or the
violation of our moral right to object to use of our work to train AI to
generate prejudicial content. […] OpenAI, for example, has received a billion
dollars in venture capital, none of which has been passed on to the authors of
the training corpus even though, without that training corpus, chatGPT would be
worthless.” – Edward Hasbrouck, National Writers Union

It’s uncertain when (or if) the Copyright Office will provide more clarity on
this issue. Also relevant in the US are the outcomes of cases like Getty Images
(US), Inc. v. Stability AI, Inc.

However, Europe has been more definitive about the ingestion issue. Directive
2019/790 says:

The [exception for copyright] shall apply on condition that the use of works and
other subject matter referred to in that paragraph has not been expressly
reserved by their rightholders in an appropriate manner, such as
machine-readable means in the case of content made publicly available online.1

This is reinforced by the recently adopted AI Act:

Any use of copyright protected content requires the authorisation of the
rightsholder concerned unless relevant copyright exceptions and limitations
apply. Directive (EU) 2019/790 introduced exceptions and limitations allowing
reproductions and extractions of works or other subject matter, for the purpose
of text and data mining, under certain conditions. Under these rules,
rightsholders may choose to reserve their rights over their works or other
subject matter to prevent text and data mining, unless this is done for the
purposes of scientific research. Where the rights to opt out has been expressly
reserved in an appropriate manner, providers of general-purpose AI models need
to obtain an authorisation from rightsholders if they want to carry out text and
data mining over such works.

In other words, European law is about to require commercial AI crawlers to
support an opt-out. However, it does not specify a particular mechanism: it only
says that consent must be ‘expressly reserved in an appropriate manner.’

So, what might that opt-out signal look like?

Robots.txt as an Opt-Out

Since most of the publicly available content on the Internet is accessed over
the Web, it makes sense to consider how an opt-out might be expressed there as a
primary mechanism. The Web already has a way for sites to opt-out of automated
crawling: the robots.txt file, now specified by an IETF Standards-Track RFC.

At first glance, robots.txt intuitively maps to what’s required: a way to
instruct automated crawlers on how to treat a site with some amount of
granularity, including opting out of crawling altogether. Some LLMs have latched
onto this it already; for example, OpenAI allows their crawler to be controlled
by it.

There are a lot of similarities between gathering Web content for search and
gathering it for an LLM: the actual crawler software is very similar (if not
identical), crawling the whole Web requires significant resources, and both uses
create enormous potential value not only for the operators of the crawlers, but
also for society.

However, it is questionable whether merely reusing to robots.txt as the opt-out
mechanism is sufficient to allow rightsholders to fully express their
reservation. Despite the similarities listed above, it is hard to ignore the
ways that LLM ingest is different.

That’s because Web search can be seen as a service to sites; it makes them more
discoverable on the Web, and is thus symbiotic – both parties benefit. LLM
crawling, on the other hand, doesn’t have any benefits to the content owner, and
may be perceived as harming them.

Through the lenses of those different purposes and their associated power
dynamics, a few issues become apparent.

1. Usability and Ecosystem Impact

Robots.txt allows sites to target directives to bots in two different ways: by
path on the site (e.g., /images vs. /users) and by User-Agent. The User-Agent
identifies the bot, allowing sites to specify things like “I allow Google to
crawl my site, but not Bing.” Or, “I don’t allow any bots.”

That might be adequate for controlling how your site appears in search engines,
but problematic when applied to AI. Let’s look at an example.

To stop OpenAI from crawling your site, you can add:

User-Agent: GPTBot Disallow: /

However, that directive doesn’t apply to Google, Mistral, or any other
LLM-in-waiting out there; you’d have to target each individual one (and some
folks are already advising on how to do that).

If you miss one, that’s your fault, and it’ll be in that model forever, so
careful (or just frustrated) people might decide to just ban everything:

User-Agent: * Disallow /

But that has the downside of disallowing AI and search crawlers – even though
presence in search engines is often critical to sites. To avoid that, you would
have to enumerate all of the search engines and other bots that you want to
allow, creating more work.

Significantly, doing so could also have a negative effect on the Web ecosystem:
if sites have a stronger incentive to disallow unknown bots thanks to AI, it
would be much harder to responsibly introduce new crawler-based services to the
Web. That would tilt the table even further in the favour of already established
‘big tech’ actors.

There are two easy ways to fix these issues. One would be to define a special
User-Agent that applies to all AI crawlers. For example:

User-Agent: AI-Ingest Disallow: /

The other approach would be to create a new well-known location just for AI –
for example /.well-known/ai.txt. That file might have the same syntax as
robots.txt, or its notoriously quirky syntax could be ditched for something more
modern.

Either solution above would make it easy for a site to opt-out of AI crawling of
any sort without enumerating all of the potential AI crawlers in the world, and
without impacting their search engine coverage or creating ecosystem risk.

I suspect that many have been assuming that one of these things will happen;
they’re fairly obvious evolutions of existing practice. However, at least two
more issues are still unaddressed.

2. Previously Crawled Content

Web search and LLMs also differ in how they relate to time.

A search engine crawler has a strong interest in assuring that its index
reflects the current Web. LLM crawlers, on the other hand, are ravenous without
regard to its age or current availability on the Web. Once ingested content
forms part of a model, they add value to that model for the lifetime of its use
– and the model often persists for months or even years after the ingested
content was obtained. Furthermore, that content might be reused to create future
models, indefinitely.

That means that a content owner who isn’t aware of the LLM crawler at crawl time
doesn’t have any recourse. From the Copyright Office sessions:

We believe that writers should be compensated also for past training since it
appears that the massive training that has already occurred for GPT and Bard to
teach the engines to think and to write has already occurred[.] – Mary
Rasenberger, The Authors Guild

This shortcoming could be addressed by a relatively simple measure: stating that
the policy for a given URL applies to any use of content obtained from that URL
at model creation time, regardless of when it was obtained.

A significant amount of detail would need to be specified to make this work, of
course. It would also likely necessitate some sort of grandfathering or
transition period for existing models.

Needless to say, the impact of this kind of change could be massive: if 90% of
the sites in the world opt out in this fashion (a la App Tracking Transparency),
it would be difficult to legally construct a new model (or at least market or
use such a model in Europe, under the forthcoming rules).

On the other hand, if that many people don’t want to allow LLMs to use their
content when offered a genuine chance to control it, shouldn’t their rights be
honoured? Ultimately, if that’s the outcome, society will need to go back to the
drawing board and figure out what it values more: copyright interests or the
development of LLMs.

3. Control of Metadata

Another issue with reusing robots.txt is how that file itself is controlled. As
a site-wide metadata mechanism, there is only one controller for robots.txt: the
site administrator.

That means that on Facebook, Meta will decide whether your photos can be used to
feed AI (theirs or others’), not you. On GitHub, Microsoft will decide how your
repositories will be treated. And so on.

While robots.txt is great for single-owner sites (like this one), it doesn’t
meet the needs of a concentrated world – it leverages the power that accrues to
a small number of platform owners to decide policy for all of their users.

Avoiding that outcome means that users need to be able express their preference
in the content itself, so that it persists no matter where it ends up. That
means it’s necessary to be able to embed policy in things like images, videos,
audio files, document formats like PDF, Office, and ePub, containers like ZIP
files, file system paths for things like git repos, and so on. Assuming that a
robots.txt-like approach is also defined, their relative precedence will also
need to be specified.

Luckily, this is not a new requirement – our industry has considerable
experience in embedding such metadata into file formats, for use cases like
content provenance. It just needs to be specified for AI control.

What’s Next?

Policy decisions like that just made by Europe might be the drivers of change in
LLM ingest practices, but I hope I’ve shown that the technical details of that
‘appropriate manner’ of opting out can significantly steer power between AI
companies and content owners.

Notably, while the worldwide copyright regime is explicitly opt-in (i.e., you
have to explicitly offer a license for someone to legally use your material,
unless fair use applies), the European legislation changes this to opt-out for
AI.2 Given that, offering content owners a genuine opportunity to do so is
important, in my opinion.

I’ve touched on a few aspects that influence that opportunity above; I’m sure
there are more.3 As I implied at the start, getting the balance right is going
to take careful consideration and perhaps most importantly, sunlight.

However, It’s not yet clear where or how this work will happen. Notably, the
standardisation request to the European Standardisation Organisations in support
of safe and trustworthy artificial intelligence does not mention copyright at
all. Personally, I think that’s a good thing – worldwide standards need to be in
open international standards bodies like the IETF, not regionally fragmented.

In that spirit, the IETF has recently created a mailing list to discuss AI
control. That’s likely the best place to follow up if you’re interested in
discussing these topics.

See also Recital 18. ↩

And I suspect other jurisdictions might follow the same approach; time will
tell. ↩

For example, some of the input to the Copyright Office mentioned group licensing
regimes. An opt-out mechanism could be adapted to support that. ↩


MORE
April 220
22.6.2024, 19:00Platformonomics (Feed)
Platformonomics TGIF #48: April 19, 2024

Platformonomics TGIF is a weekly roll-up of links, comments on those links, and
perhaps a little too much tugging on my favorite threads.

Get Platformonomics Updates By Email

Subscribe

The recovery from Clown Week continues.

News Blockbuster News of the Week

Google announces “We are a place of business”. Who knew?

Some employees (and a growing number of ex-employees) are still struggling with
this concept:

Previous:

Google Goes Off the Guardrails

Related:

Nine Google employees arrested after eight-hour sit-in protest, Google fires 28
employees after sit-in protest over Israel cloud contract,

Is Quantum Computing a Scam?

The strongest evidence for quantum computing being a scam is IBM’s avid
involvement (or at least incessant PR). IBM has an unrivaled track record in the
21st century at selling futures that never materialize (e.g. Watson, blockchain,
IBM Cloud, smart cities, enterprise-class Second Life). Somehow they never get
past the elaborate TV ad campaigns.

Previous:

Follow the CAPEX: The Clown Car Race Checkered Flag, Introducing Platformonomics
ClownWatch, IBM and the Art of Misleading Investors, This Week in Securities
Fraud, Clown CAPEX, IBM’s “Cloud” Business (or Lack Thereof), IBM: How Much
Longer Do the Good Times Last?, IBM’s Lost Decade

Power Crunch

With even Oracle discovering that electricity is the long pole for large scale
AI deployments, I think we can now safely say this is common knowledge. Told ya!

Previous:

EU Insanity: AI Energy Suicide, EU Insanity: AI Energy Suicide (Part Deux),
Cloud Power Up

Cantwell Recants?

Last week:

This week:

Great news if true. Hopefully she will take great pride at bringing such
principled and material improvements to this legislation. I still have not heard
from her office, despite checking the box that “I Would Like A Response From The
Senator”.

Previous:

Tik Tok Traitors: Maria Cantwell (Continued), Tik Tok Traitors: Maria Cantwell,
Tik Tok’s Call to Arms: Useful Idiots, Fellow Travelers, Registered and
Unregistered Foreign Agents Unite!, Contact Your Senators

Related:

This senator wants an online privacy law. She’s slowed efforts for years.,
Chinese diplomats are quietly meeting with Hill staffers about TikTok

Private Equity in Action: The VMware Saga May Never End

Previous:

Software Migration Alerts, Private Equity in Action: VMware “Unease” Continues,
Perfidious Private Equity, Private Equity in Action: VMware Yet Again, Private
Equity in Action: VMware (Continued), Private Equity in Action: VMware, Private
Equity Will Buy and Ruin More Software Companies, Software Migration Alert:
VMware (Continued), Software Migration Alert: VMware, Private Equity in Action:
VMware Customer Complaints “Unwarranted”

Related:

Broadcom questioned by EU over VMware licensing changes

Private Equity Raises More Money to Ruin Software Companies

When Vista Equity (or any other PE firm) comes amalgamating, it is time to start
migrating.

Previous:

Software Migration Alerts, Perfidious Private Equity

EU Insanity: Regulatory Contortions

The Information responds on my behalf:

Driving the opinion was European privacy group NOYB, which last November
complained that Meta’s offering—of a free service supported by ads or one that
was ad-free but cost money—wasn’t structured fairly. The privacy group argued
that the subscription was a “privacy fee.” If extended to other services, such a
fee would require people to spend thousands of euros a year to, for instance,
keep their phone “private,” the group said.

How about this idea as an alternative—Europeans who don’t want their data used
could simply stop using Meta’s services? Where is it written that having access
to a social media site is a fundamental human right? What happened to the idea
that a private company has a right to make money? Seriously, if NOYB dislikes
Meta so much, why on earth are its members using Facebook or Instagram to start
with?

Previous:

EU Insanity: Not Learning From Regulatory Failure, EU Announces DMA Struggle
Sessions, EU Insanity: AI Regulatory Suicide, EU Insanity: AI Energy Suicide, EU
Insanity: The Kafka Singularity, EU Insanity: AI Energy Suicide (Part Deux), EU
Insanity: Mistral Edition, Move Fast and Regulate Things (You Don’t Understand),
The EU Will Continue to Fiddle While Rome Burns, When “Move Fast and Regulate
Things” Breaks Down, AI Regulation: Move Fast and Regulate Things

UK Announces Big Tech Struggle Sessions “For The Children”

The struggle sessions will continue until Europe has a tech industry (i.e.
indefinitely).

Previous:

EU Announces DMA Struggle Sessions

Mercantilist Showdown

“Insiders say the immediate impulse of EU policymakers confronting the West
Taiwanese threat to the European auto industry was to levy new fines on American
Big Tech companies.”

Maybe, just maybe, Europe is starting to wake up to its real challenges (even if
the previous two items suggest otherwise).

Previous:

Trade Surplus Region Complains Trade Deficit Region is “Protectionist”,
Collapsing Economic Models: German Edition, EU Rethinking Its Stance on
Mercantilism, German Auto Industry Reaches Hail Mary Stage, How Do You Say
“Burning Platform” in German?, Volkswagen’s Death Throes Continue, The Fate of
the European Economy: Automotive Edition, Europe: Investing in Slave Labor

Related:

Trade Wars Are Class Wars

Viewer Mail – Episode 2

Q: What about Apple’s CAPEX?

A: I haven’t tracked Apple because they don’t have a public cloud. Apple’s CAPEX
spend is both flat and underwhelming in absolute terms. They spend on
manufacturing tooling so the proportion going to data centers is even smaller.
In the past they have relied on the hyperclouds for storage. I don’t know if
they have migrated entirely to their own data centers or not. We’ll see if Apple
gets more excited about CAPEX in the generative AI era.

Got more Viewer Mail questions?

Previous:

Viewer Mail Episode 1


MORE
June 60
Today, 17:55ongoing by Tim Bray (Feed)
Meta.ai Oh My!

“Meet Your New Assistant” says the announcement, going on with “Built With Llama
3”. And oh my goodness has it ever got a lot of coverage. So I thought I might
as well try it.

My first cut was a little unfair; I asked it about a subject on which I am
unchallenged as the world’s leading expert: Tim Bray. (That’s probably
overstating it: My wife is clearly in the running.)

So I asked meta.ai “What does Tim Bray think of Google?” Twice; once on my phone
while first exploring the idea, and again later on my computer. Before I go on,
I should remark that both user interfaces are first-rate: Friction-free and
ahead of the play-with-AI crowd. Anyhow, here are both answers; it may be
relevant that I was logged into my long-lived Facebook account:

The problem isn’t that these answers are really, really wrong (which they are).
The problem is that they are terrifyingly plausible, and presented in a tone of
serene confidence. For clarity:

I am not a Computer Scientist. Words mean things.

I worked for Google between March of 2010 and March of 2014.

I was never a VP there nor did I ever have “Engineer” in my title.

I did not write a blog post entitled “Goodbye, Google”. My exit post, Leaving
Google, did not discuss advertising nor Google’s activities in China, nor in
fact was it critical of anything about Google except for its choice of
headquarters location. In fact, my disillusionment with Google (to be honest,
with Big Tech generally) was slow to set in and really didn’t reach critical
mass until these troubling Twenties.

The phrase “advertising-based business model”, presented in quotes, does not
appear in this blog. Quotation marks have meaning.

My views are not, nor have they been, “complex and multifaceted”. I am
embarrassingly mainstream. I shared the mainstream enchantment with the glamor
of Big Tech until, sometime around 2020, I started sharing the mainstream
disgruntlement.

I can neither recall nor find instances of me criticizing Google’s
decision-making process, nor praising its Open-Source activities.

What troubles me is that all of the actions and opinions attributed to meta.ai’s
version of Tim Bray are things that I might well have done or said. But I
didn’t.

This is not a criticism of Meta; their claims about the size and sophistication
of their Llama3 model seem believable and, as I said, the interface is nifty.

Is it fair for me to criticize this particular product offering based on a
single example? Well, first impressions are important. But for what it’s worth,
I peppered it with a bunch of other general questions and the pattern repeats:
Plausible narratives containing egregious factual errors.

I guess there’s no new news here; we already knew that LLMs are good at
generating plausible-sounding narratives which are wrong. It comes back to what
I discussed under the heading of “Meaning”. Still waiting for progress.

The nice thing about science is that it routinely features “error bars” on its
graphs, showing both the finding and the degree of confidence in its accuracy.

AI/ML products in general don’t have them.

I don’t see how it’s sane or safe to rely on a technology that doesn’t have
error bars.


MORE
April 190
Today, 14:28Phil Windley's Technometria (Feed)
Web 2.0 is Collapsing Under its Own Weight

Summary: The overhead of performing even simple tasks online is getting larger
and larger. I question the security of almost all these supposedly "secure"
messaging systems. And I'm tired of the 'Utopia of Rules' mindset pervasive in
every organization. It's exhausting how they expect customers to constantly
adapt to their needs.

I don't know if you recall the game Kerplunk. It's a classic children's game
that has been around for decades. I remember playing it with my sister. The
basic setup involves a transparent plastic tube, a number of sticks, and
marbles. The sticks are threaded through the tube to form a web or nest at the
bottom on which the marbles rest. We'd take turns removing a stick at a time,
trying not to let any marbles fall through the web and out of the tube. At some
point, the remaining sticks can't hold the marbles and everything falls down.

The modern web reminds me more and more of a big Kerplunk game and I think the
marbles are about to fall. What started out as an easier way to do things like
shop, bank, and get health care information has become increasingly complex over
time. More and more of the email I receive seems to be simply directing me to
log into some bespoke system to retrieve a message or engage in some workflow.
And even with a password manager, the act of logging in is often a chore with
different user interfaces, custom MFA requirements, and weird rules for
passwords. Once you're on the system, session time-outs induce their own form of
anxiety since stepping away for a few minutes to attend to something else might
require going through the whole Kafkaesque process all over again. The modern
web has turned into a dystopian theater of the absurd where even reading a
simple appointment reminder from your doctor requires several minutes of
stress-inducing interaction with baroque systems and processes.

And it's not just doctors, of course, banks, government agencies, hospitals,
ecommerce sites, and customer service systems all adopt these special purpose
messaging systems. If you ask these organizations why they use bespoke messaging
systems, they'll list things like "timely and improved communication,"
"convenience," and "privacy and security." But the real reason is that it's more
convenient for them because these systems are integrated with their backends and
make their processes more manageable. There's certainly nothing about them
that's more convenient, timely, or better than email for their customers1.

I also question the privacy and security premise. Email can be insecure. And
your email provider can see the contents of your emails. But the messaging
system run by your doctor or bank is likely less secure than the email systems
run by Apple, Google, and the others. And achieving privacy by making everything
incompatible so that you have to use a different system for each correspondent
is like chopping off your finger to prevent hangnails.

How did we get here? Bureaucracy. Not just government bureaucracy, but
bureaucracy of all kinds. In Utopia of Rules2, David Graeber talks about how
power imbalances force the less powerful group to perform what he calls
interpretive labor, the work of understanding and implementing what's better or
more convenient for the more powerful partner. People are not equal participants
in online interactions. We don't have the tools to be fully embodied online3.
Because of this we are forced to play by the rules organizations online who are
digitally embodied with servers, identity systems, customer management systems,
and so on. And part of that is being forced to use their inconvenient and anemic
messaging systems.

What's the answer? People need tools. I think digital wallets (a bad name for an
important tool), autonomic (peer) identifiers with strong cryptography, and
verifiable credentials are a huge step forward. These tools provide the means
for people to be peers online rather that mere ghosts in someone else's machine.
That's why I insist on using the term self-sovereign rather than decentralized
to describe these systems. Cogito Ergo Sum.

Notes For a deeper dive into why one-off messaging systems are never as good as
email, see Rich Sharing and Personal Channels. Email and other useful messaging
systems exhibit a property called rich sharing that makes them much more robust
that the simple idea of "sharing a message" would bring to mind. If you're
interested in power imbalances and how they come about, I can't recommend
Graeber's book highly enough. He had such a keen understanding of this problem
and wrote about it in a way that's both informative and entertaining. I talk
about this in more detail in Chapter 17 of Learning Digital Identity when I
discuss authentic digital relationships.

Photo Credit: Playing Kerplunk from DALL-E (public domain) Prompt: Draw a
picture of a boy and girl playing kerplunk that's 1200x500 pixels

Tags: web identity authentication authorization verifiable+credentials


MORE
April 150
24.6.2024, 6:00SEI Blog (Feed)
Comparing the Performance of Hashing Techniques for Similar Function Detection
This blog post explores the challenges of code comparison and presents a
solution to the problem.

MORE
April 150
Today, 17:55ongoing by Tim Bray (Feed)
Topfew Release 1.0

Back in 2021-22, I wrote a series of blog posts about a program called “topfew”
(tf from your shell command-line). It finds the field values (or combinations of
values) which appear most often in a stream of records. I built it to explore
large-scale data crunching in Go, and to investigate how performance compared to
Rust. There was plentiful input, both ideas and code, from Dirkjan Ochtman and
Simon Fell. Anyhow, I thought I was finished with it but then I noticed I was
using the tf command more days than not, and I have pretty mainstream
command-line needs. Plus I got a couple of random pings about whether it was
still live. So I turned my attention back to it on April 12th and on May 2nd
pushed v1.0.0.

I added one feature: You can provide a regex field separator to override the
default space-separation that defines the fields in your records. Which will
cost you a little performance, but you’re unlikely to notice.

Its test coverage is much improved and, expectedly, there are fewer bugs. Also,
better docs.

Plan

I think it’s pretty much done, honestly can’t think any useful new features. At
some point, I’ll look into Homebrew recipes and suchlike, if I get the feeling
they might be used.

Obviously, please send issues or PRs if you see the opportunity.

Who needs this?

It’s mostly for log files I think. Whenever I’m poking around in one of those I
find myself asking questions like “which API call was hit most often?” or “Which
endpoint?” or “Which user agent?” or “Which subnet?”

The conventional hammer to drive this nail has always been something along the
lines of:

awk '{print $7}' | sort | uniq -c | sort -nr | head

Which has the advantage of Just Working on any Unix-descended computer. But can
be slow when the input is big, and worse than linear too. Anyhow, tf is like
that, only faster. In some cases, orders of magnitude faster. Plus, it has
useful options that take care of the grep and sed idioms that often appear
upstream in the pipe.

Topfew’s got a decent README so I’m not going invest any more words here in
explaining it.

But it’s worth pointing out that it’s a single self-contained binary compiled
from standalone Go source code with zero dependencies.

Performance

This subject is a bit vexed. After I wrote the first version, Dirkjan
implemented it in Rust and it was way faster, which annoyed me because it ought
to be I/O-bound. So I stole his best ideas and then Simon chipped in other good
ones and we optimized more, and eventually it was at least as fast as the Rust
version. Which is to say, plenty fast, and probably faster than what you’re
using now.

But you only get the big payoff from all this work when you’re processing a
file, as opposed to a stream; then tf feels shockingly fast, because it divides
the file up into segments and scans them in parallel. Works remarkably well.

Unfortunately that doesn’t happen too often. Normally, you’re grepping for
something or teeing off another stream or whatever. In which case, performance
is totally limited by reading the stream; I’ve profiled the hell out of this and
the actual tf code doesn’t show up in any of the graphs, just the I/O-related
buffer wrangling and garbage collection. Maybe I’m missing something. But I’m
pretty sure tf will keep up with any stream you can throw at it.

Tooling

Over the years I’ve become an adequate user of GitHub CI. It’s good to watch
that ecosystem become richer and slicker; the things you need seem to be there
and for an OSS hobbyist like me, are generally free. Still, it bothers me that
Everything Is On GitHub. I need to become aware of the alternatives.

I still live in JetBrains-land, in this case specifically Goland, albeit
unfashionably in Light mode. It scratches my itches.

Anyhow, everything is easier if you have no dependencies. And our whole
profession needs to be more thoughtful about its dependencies.

Dirty secret

I’ve always wanted to ship a two-letter shell command that someone might use.
Now I have. And I do think tf will earn a home in a few folks’ toolboxes.


MORE
May 40
22.6.2024, 19:00Platformonomics (Feed)
Platformonomics TGIF #47: April 12, 2024

Platformonomics TGIF is a weekly roll-up of links, comments on those links, and
perhaps a little too much tugging on my favorite threads.

Get Platformonomics Updates By Email

Subscribe

Clown Week is here!

News Follow the CAPEX: The Clown Car Race Checkered Flag

As the Clown Week cornerstone, we call a winner in the CAPEX Clown Car Race:

(Oracle left, IBM right)

This visit to Clown City includes picking a victor between IBM and Oracle to be
the cloudiest clown, the biennial telling of the joke about Oracle’s CAPEX being
indistinguishable from the x-axis, and wondering why Oracle is struggling so
much with its modest cloud infrastructure build-out. I also express
disappointment that not one of Bank of America, GAIA-X, and Trump Media ever
achieved epic superclown status, despite so much potential.

Reminder: I don’t set companies’ CAPEX budgets, I just work with the decisions
they’ve made.

Previous:

Follow the CAPEX: Keeping Up With The Clowns, Follow the CAPEX: Separating the
Clowns from the Clouds, Follow the CAPEX: Clown Watch, This Week in Securities
Fraud, Clown Week Preview: Oracle CAPEX

Platformonomics ClownWatch

This Clown Week we say farewell to cloud-washing clowns and shift our focus to
AI-washing clowns. The Laws of Clownodynamics tell us clownishness is never
created nor destroyed.

Our new Platformonomics ClownWatch service monitors companies for large
disparities between flowery AI rhetoric and their actual CAPEX spend.

Previous:

Introducing Platformonomics ClownWatch

Software Migration Alert: Model N

Private equity has come for Model N. If you are a Model N customer, be warned
the PE clowns are coming to ruin your software. You need to immediately start
your migration plan.

We have a new summary page of Software Migration Alerts.

Previous:

Perfidious Private Equity, Private Equity Buys Software Company from Private
Equity: Avetta, Private Equity Buys IT Company from Private Equity: Presidio,
Private Equity in Action: VMware Customer Complaints “Unwarranted”, Software
Migration Alert: Jama Software, Software Migration Alert: Everbridge, Private
Equity in Action: VMware (Continued), Private Equity in Action: VMware, Private
Equity Will Buy and Ruin More Software Companies, Software Migration Alert:
VMware (Continued), Software Migration Alert: VMware, Software Migration Alert:
Rover, Software Migration Alert: Avid Technology, Software Migration Alert:
EngageSmart, This Week’s Software Migration Alerts

Cloud Repatriation is Going Just Great: Radio Silence from Equinix

Not just clowns but fraudulent clowns?

Previous:

Cloud Repatriation is Going Just Great: Equinix, New Cloud Repatriation
Hypothesis, Cloud Repatriation Where Art Thou?, Cloud Repatriation: The Search
Continues, Cloud Repatriation: Still Not a Thing, Platformonomics Repatriation
Index – Q1 2023: Surf’s Up?

EU Insanity: AI Energy Suicide (Part Deux)

Needless to say, there is no Clown Week without the European Union!

Previous:

EU Insanity: Not Learning From Regulatory Failure, EU Announces DMA Struggle
Sessions, EU Insanity: AI Regulatory Suicide, EU Insanity: AI Energy Suicide, EU
Insanity: The Kafka Singularity, EU Insanity: Mistral Edition, Move Fast and
Regulate Things (You Don’t Understand), The EU Will Continue to Fiddle While
Rome Burns, When “Move Fast and Regulate Things” Breaks Down, AI Regulation:
Move Fast and Regulate Things, Cloud Power Up

Related:

AI Datacenter Energy Dilemma – Race for AI Datacenter Space, Irish power crunch
could be prompting AWS to ration compute resources

Antitrust Incoherence: Roomba Aftermath Edition

In the absence of a coherent doctrine that tells everyone where the lines are on
the field, the regulators are just clowns. Where is the regulatory response to
Jassy’s points?

Previous:

Antitrust Incoherence: Apple Edition, Antitrust Incoherence: Spotify Edition,
Antitrust Incoherence: Roomba Edition, The Incoherence Doctrine: The FTC Strikes
Again, A New Antitrust Doctrine, The DOJ Play at Home Game

McKinsey a Source of Bad Banking Advice (Shocking!)

Evil clowns are a thing.

Add attempting to destabilize the banking system to McKinsey’s long list of
crimes (opioids crisis, insider trading, also Enron, other autocracies, hybrid
cloud, et al.).

I reiterate my call for McKinsey to get the corporate death penalty. Would be a
huge net win for civilization.

Someone described McKinsey this week as one of the first LLMs: they’re eloquent
even when hallucinating wildly.

Previous:

A Call for the Corporate Death Penalty, McKinsey Karma?

Related:

McKinsey: Last Week Tonight with John Oliver

Putting the A and I into Authoritarianism

I had assumed the EA/AI doomer clowns had packed it in after the conviction of
poster boy SBF and their inability to articulate any causal extinction
mechanisms. Or at least assumed that the problem was relegated to the EU. But it
seems they are still on the march in the US.

We should probably start treating Asana like a private equity company (i.e.
don’t use them).

Previous:

EU Insanity: Not Learning From Regulatory Failure, EU Insanity: AI Regulatory
Suicide, EU Insanity: The Kafka Singularity, EU Insanity: Mistral Edition, Move
Fast and Regulate Things (You Don’t Understand), The EU Will Continue to Fiddle
While Rome Burns, When “Move Fast and Regulate Things” Breaks Down, AI
Regulation: Move Fast and Regulate Things

Related:

How a billionaire-backed network of AI advisers took over Washington


MORE
June 60
Follow the CAPEX: The Clown Car Race Checkered Flag
Oracle, left. IBM, right.

We’ve used capital expenditures (CAPEX) to separate the clouds from the clowns
for over a decade. Clouds consume copious CAPEX. Clowns confabulate.

It is time for another inspection tour of Clown City (previous visits: 2018,
2020, 2022). Once a bustling stop on the cloud-washing promotional circuit,
Clown City is now but a shadow of its former self. Remember (colorfully
bewigged) visitors like Cisco Cloud, HPE Helion, Rackspace, Verizon Cloud, and
VMware Cloud? They all eventually learned – the hard way – that clouds are built
with CAPEX, not chatter.

Sadly, a once-promising new generation of potential clowns never rose to the
occasion of repopulating Clown City and entertaining us with the gap between
their cloud rhetoric and their CAPEX spending.

The only activity left on the sleepy streets of Clown City is the CAPEX Clown
Car Race between the last two of our self-proclaimed cloud pretenders: IBM and
Oracle. It is time to declare a winner in this (admittedly bush league) race!

The CAPEX Clown Car Race checkered flag marks the end of the cloud-washing era.
But clown fans should not despair. Clowns are eternal. The First Law of
Clownodynamics says clownishness is neither created nor destroyed.

There are always companies cloaking themselves in the new hotness without
possessing the proper prerequisites for participation. When the hotness moves,
so do the clowns.

But some domains require real investment, which means the all-seeing-eye of
CAPEX still exposes poseurs.

Hello AI clowns! Welcome to Platformonomics ClownWatch

CAPEX might be an even better tell in AI than it was with cloud, because GPUs
ain’t cheap and software now has COGS (a development which may prove
regrettable).

But first, we have a drag race approaching the finish line.

The CAPEX Clown Car Race

Oracle got off to a very poor start in our race, with Larry Ellison harrumphing
about cloud computing:

Maybe I’m an idiot, but I have no idea what anyone is talking about. What is it?
It’s complete gibberish. It’s insane. When is this idiocy going to stop?”

Eventually Oracle got cloud religion after some embarrassing but certainly not
unexpected attempts to substitute hot air for CAPEX. Nothing motivates like
losing half your database market share.

Through 2022 it even looked like Oracle might be serious in the only way that
matters: up-and-to-the-right CAPEX spending.

But 2023 saw a catastrophic 27% drop in Oracle’s CAPEX spending. While their
absolute and directional spending is both disappointing and probably disastrous,
it doesn’t matter because their opponent in the CAPEX Clown Car Race is IBM
(company motto: “Down and to the right”).

There is a deeply philosophical question in how you win a clown car race. Is the
winner the most clowny or the least clowny? With next to no consideration, we’re
going with the latter and calling a winner.

Congratulations Oracle – based on your CAPEX, you’re the cloudiest of the
clowns! But Oracle is still a clown, and certainly not a hyperscale cloud (there
is no “big four” despite the best efforts of Oracle PR to slip away from the
children’s table).

And IBM gets a clowniest clown participation trophy. IBM — I’ve always believed
you could do this!

Below we’ll look at IBM and Oracle’s CAPEX spending, amuse ourselves with the
current fairy tales they’re telling, ask what is going on at Oracle, and then
make the brutal comparison with the CAPEX boomtown of Cloud City. And we’ll end
with a quick “where are they now” look at that once-promising new generation of
aspiring clowns.

Get Platformonomics Updates By Email

Subscribe

IBM

Company motto: “Down and to the Right”

In what is not even remotely a surprise, IBM hit new record lows in 2023 for
both absolute CAPEX spending and CAPEX spending as a percentage of revenue. The
hypers of “hybrid cloud” cut CAPEX by 20%, spending less than $1.5 billion
(across the entire company, not just cloud infrastructure), which is less than
half of what they spent as recently as 2020.

These numbers confirm there is no cloud in “hybrid cloud”. Like soylent green,
“hybrid cloud” is people – consultants (and mostly offshore consultants). There
is no cloud business at IBM (as we have been saying for over a decade).

In perhaps related news, the leader of IBM Cloud quietly departed for another IT
consulting company (one that doesn’t pretend to be a platform company).

But don’t shed too many tears for IBM. The company is in the midst of a bold
strategic transformation. A transformation from a cloud to an AI . Because the
absence of CAPEX in either domain ensures you are a .

Meet the new shell game. Same as the old shell game.

IBM is now running its cloud playbook for AI. The company opened their Q4 2023
earnings release wanting you to think they are a generative AI company:

“In the fourth quarter, we grew revenue in all of our segments, driven by
continued adoption of our hybrid cloud and AI offerings. Client demand for AI is
accelerating and our book of business for watsonx and generative AI roughly
doubled from the third to the fourth quarter,” said Arvind Krishna, IBM chairman
and chief executive officer.

You’d almost think IBM was seeing material revenue from AI. But when asked to
“elaborate specifically on exactly what the book of business means” by a veteran
Wall Street analyst familiar with IBM’s antics, much bobbing and weaving ensues:

So Toni, on the AI book of business, this is not all revenue in the quarter. I
would just begin with that statement to set it straight. At this stage, we
wanted to start looking at what is our momentum, what is the sentiment from our
clients. So we went to a measure that is more reflective of, I’ll use the word
signings. What is the commitment the clients are making to us? Consulting is
straightforward. It is the signings. Consulting signings are anywhere from 12 to
24 months on average is how much time they play out over there. And on software,
it’s what they’re committing to. And we are using SaaS ACV. So it’s a 12-month
commitment, which is typical for as a service as well as, since we do offer our
portfolio both ways as license or as a service, it includes the license piece as
well. Now over a long-term, let’s call it a couple of years or more, yes, the
book of business should turn into an amount of revenue in a quarter, but that’s
going to take a bit of time to catch up. But we felt that this gives the better
indicator right now of what is our traction and what is our acceleration in that
part of the business.

This is the same game IBM played for cloud. Chant the buzzwords and hope
investors won’t notice the company isn’t actually investing and doesn’t have
competitive products.

As they did with cloud, IBM’s AI numbers are not something they actually report
to the SEC, yet lead with in investor communications. I don’t know if that is
securities fraud, but I know IBM has no GPUs.

Oracle

Company motto: “That’s Not Just a Rack, It’s an Entire Region”

Data note: we do Oracle a solid by comparing their March 2023 through February
2024 (FYQ4 to FYQ3) results to the 2023 calendar year results of other
companies. This lets Oracle bring a full two months from their glorious future
to the comps. Hopefully that doesn’t give them too unfair an advantage.

When we last checked in on the database vampire, we noted they were (finally)
ramping their CAPEX:

Oracle, on the other hand, is at long last starting to put its money where its
mouth has been (though the mouth’s lead looks insurmountable). The database
vampire doubled their annual CAPEX spend in the last year (using their March
2021 through February 2022 results). The company promised to spend $4 billion in
CAPEX in FY22 and it looks like will hit that mark by May. And they surpassed
IBM in annual CAPEX spending this year.

But after spending more than $8 billion in our ersatz calendar 2022, Oracle has
fallen off the CAPEX wagon. Oracle CAPEX declined 27% to under $6 billion from
2022 to 2023.

Oracle’s commentary in the last several quarters has (commendably!) focused on
CAPEX, with the CAPEX discussion taking up more time in each successive
conference call. Oracle repeatedly tells us they have “enormous” cloud demand,
are capacity constrained, and are building out data centers incredibly fast. But
something has gone off the rails in Oracle’s cloud infrastructure build-out
because their CAPEX spending has declined by double digits in each of the last
three quarters.

In FY23 Q1 (Sept 2023), Oracle says:

Because we have far more demand than we can supply, our biggest challenge is
building data centers as quickly as possible.

Capital expenditures were $8.3 billion over the last four quarters and we are
clearly beginning to see the cash flow benefits stemming from our cloud
transformation. CapEx was $1.3 billion in Q1 as we continue to build capacity
for bookings and our customers’ growing needs.

Given the demand we have and see in the pipeline, I expect that fiscal year 2024
CapEx will be similar to this past year’s CapEx.

Yet Oracle CAPEX for FYQ1 was down 24%. And they reiterate the guidance to spend
$8+ billion in FY24 CAPEX (ending May 2024).

In FY2024 Q2 (Dec 2023), Oracle told us:

The demand for Oracle’s Cloud Infrastructure and Generative AI is consistently
increasing quarter after quarter.

We have to build 100 additional cloud data centers because there are billions of
dollars more in contracted demand than we currently can supply. Cloud
Infrastructure demand is huge and growing at an unprecedented rate.         

CapEx was $1.1 billion in Q2 as we continue to build capacity for bookings and
our customers’ growing needs. Given the enormity of our pipeline and backlog, I
expect CapEx will be somewhere around $8 billion this fiscal year, meaning our
second half CapEx will be considerably higher as we bring online more capacity.

Yet Q2 CAPEX was down 56% and the lowest spend in eight quarters. And they
slightly reduce guidance for the fiscal year ending in May 2024 to $8 billion.

By FY2024 Q3 CAPEX discussion dominated their conference call remarks (as it
should in all calls):

Where if not for some continuing supply constraints, consumption growth would
have been even higher.

Now while we spent $2.1 billion on CapEx this quarter, the $1.7 billion in the
cash-flow statements is slightly lower just due to the timing of payments. So
the $2.1 billion is actually what we spent and will pay for.

We are working as quickly as we can to get the cloud capacity built out given
the enormity of our backlog and pipeline. I expect the CapEx will be somewhere
around $7 billion to $7.5 billion this fiscal year, meaning our Q4 CapEx should
be considerably higher.

We have enormous amounts of demand. I tried to make that clear last quarter, and
we have more capacity coming online. But we have tried to – we’re trying to
focus on much larger chunks of data center capacities and electricity and all of
that and that’s just – that all to come.

Let me add that Oracle has been building data centers at a record level and a
lot of people I think are aware that we can build fairly small data centers to
get started when we want to.

We can go very small. We can get a full cloud data center with Ultra services in
10 racks.

But this is what I want to point out. We’re also building the largest data
centers in the world that we know of. We’re building an AI data center in the
United States where you could park eight Boeing 747s nose-to-tail in that one
data center. So, we are building large numbers of data centers, and we were –
and some of those data centers are smallish, but some of those data centers are
the largest AI data centers in the world.

So, we’re bringing on enormous amounts of capacity over the next 24 months
because the demand is so high, we need to do that to satisfy our existing set of
customers.

So, for fiscal year ’25, looking at about $10 billion in CapEx because it’s also
involves not only some big centers, but it also involves expansions of existing
centers. So we’ve already got some areas that we will be filling out. So at
least preliminarily, we’re looking at $10 billion for next year. And then it’s
not too complicated to figure out the math here when I’m looking at somewhere
between $7 billion and $7.5 billion for the full year and you’ve got all the
numbers for one, two, and three at this point.

And I would include for Q3 the one we just are announcing. I would add in the
amount we haven’t paid yet as the CapEx number for this quarter. Okay? And then
I guess that would be and then Larry gets the second question. But anyway, so
$2.1 billion for this quarter and you’ve got Q1 and Q2 and I’m going to be
somewhere between $7 billion and $7.5 billion for the full year, which is
actually a little bit lower than I thought. But we were able to do pretty well.
You know-how we spend very carefully.

So, there is a tremendous amount of demand, the data centers take longer to
build, and we would like that said, we are getting very good at building them
quickly and getting the building the power and the communication links in, we’re
doing faster than we have ever happened in the past.

It’s very different than it used to be. So, we’re able to bring additional
capacity online very quickly if we have that the electric power and the
communication lines. So, is the long pole in the tent is actually building the
structure, connecting the electricity, connecting the communication lines.

Despite claiming they’re building data centers “at a record level”, Q3 CAPEX was
down 36%. They strangely want credit for more CAPEX spending than what they
reported to SEC (again, securities fraud questions are not addressed here). And
Larry tried to distract with a soliloquy on how many 747s would fit in a future
Oracle data center (when you only fly private, you may not realize airlines
don’t fly 747s any more).

Guidance for the full fiscal year CAPEX spend has come down by over a billion
dollars. The lower target still requires a huge ramp in Q4, and roughly doubling
spending from Q3. But most of all, we see the first acknowledge of constraints
on data centers build-outs, in particular electron availability.

And this double-digit CAPEX decline occurs amidst the AI frenzy, a trend Oracle,
of course, claims to be leading. 2023 was an odd year to slash infrastructure
investment if you’re big in AI (or say you are).

The fundamental question about Oracle’s cloud infrastructure ambitions is does
the company just not know how to build data centers with more than a couple
racks or have they concluded they don’t want to (or can’t) spend the kind of
money this game requires?

Oracle has been on a very public cloud scale learning curve. First they
pretended they didn’t need much CAPEX because, of course, Oracle’s stuff was
just so much better than anyone else’s.

Then they won a fraction of Zoom’s cloud business, hyped it in time-honored
Oracle fashion, and were humbled when they couldn’t fulfill the “extreme”
capacity demanded.

While it is cute that Oracle counts a single rack as a data center and ten racks
is a “Ultra” data center, the hyperclouds operate at many orders of magnitude
greater scale. Oracle is now trying to transition beyond bonsai data centers and
seems to be struggling with the concrete and steel world of data centers (with
their need for building permits, connectivity and power).

No doubt Oracle will someday learn why hyperclouds are building smaller data
centers and not 747 hangars.

The hyperclouds have decades of experience building cloud-scale data centers at
this point. They manage multi-year lead times and are deeply engaged in the “mad
scramble” for power, which is the biggest constraint for data centers today.

Oracle is now guiding to a FY25 CAPEX spend of – gasp — $10 billion, but given
their recent inability to actually spend to their guidance, that may still be
optimistic.

The other possibility is Oracle is having second thoughts about the financial
hit associated with transforming an asset-light software company into a serious
cloud infrastructure player. In 2022, Oracle’s CAPEX as a percentage of revenue
hit 18%, comparable to what hyperclouds can spend. But that spend also came with
a roughly $2 billion profit hit. They quickly reined in that percentage by over
a third in 2023.

But the financial issue isn’t just the impact on Oracle profits. Tragically,
they aren’t actually competing with fellow clown car IBM. They’re competing with
Amazon, Google and Microsoft and must soberly contemplate their financial
wherewithal to play in that league.

For all of Oracle’s ramp in CAPEX in recent years, they are over $150 billion
behind cumulative hypercloud spending. And they are falling further and further
behind with every passing quarter. In 2023, AWS spent $24.8 billion on CAPEX
(over four times Oracle), Google spent $32.3 billion (over five times Oracle),
and Microsoft $41.2 billion (almost seven times Oracle). Microsoft spent more on
CAPEX in just 2023 than Oracle has in its entire nearly 50 year history.

Oracle’s $10 billion in CAPEX guidance for FY25 adds a nice incremental zero to
their spend (and hits a milestone the hyperclouds all passed in 2016), but it
still doesn’t come close to what the hyperclouds are spending. If you add up
Oracle’s actual CAPEX spending and guidance for their fiscal years 2022 through
2025, it doesn’t match what Google and Microsoft spent in 2023.

To give a sense of how dramatically Oracle is being left behind, here is the
combined CAPEX spending of AWS, Google and Microsoft (corporate level spend for
Google and Microsoft, but ~80% of that is cloud infrastructure), minus Oracle’s
spending (which is also corporate level). The gap has grown from $20 billion in
2015 to over $90 billion in 2023.

For completeness, to make Oracle’s clown car status abundantly clear, and to
make my biennial favorite joke about Oracle being indistinguishable from the
x-axis, here is annual and cumulative CAPEX for both the hyperclouds and Oracle.

The Oracle plan to catch up to the hyperclouds remains unclear to me.

What Ever Happened to…

While our once-promising cohort of new clowns disappointed in both their
hyperbole and investment, they remain entertaining in other ways:

Bank of America – still dominant as the worldwide leader in financial crime, but
is also now insolvent in the wake of rising interest rates. And still have spent
ZERO on CAPEX since 2011 (bank accounting FTW). Their alum and pal at IBM who
was leading the charge for a compliance-oriented private cloud doohickey bailed,
so expect B of A to eventually be the last bank to embrace the public cloud.
GAIA-X – the EU’s hand waving assertion of their strategic cloud autonomy “based
on European values” is, in the words of one booster, “a paper monster that will
exist but will not have any impact in the market”. Others say it “failed to
reach critical mass or really take off”, “operates too much at the meta level”,
is “still a long way from clear, tangible use cases” and “never got out of this
theoretical sinkhole”. Even IBM spends more on CAPEX in a year than GAIA-X’s
multi-year budget. Trump Media & Technology Group – “Net cash used in investing
activities for the twelve months ended December 31, 2023 was approximately $2.2
thousand” ($2,200.00). Emphatically serverless!

We now return to our regular programming.










































MORE
June 60
24.6.2024, 6:00SEI Blog (Feed)
The Latest Work from the SEI: an OpenAI Collaboration, Generative AI, and Zero
Trust
This post highlights the latest work from the SEI in the areas of generative AI,
zero trust, large language models, and quantum computing.

MORE
April 100
22.6.2024, 19:00Platformonomics (Feed)
Platformonomics TGIF #46: April 5, 2024

Platformonomics TGIF is a weekly roll-up of links, comments on those links, and
perhaps a little too much tugging on my favorite threads.

Get Platformonomics Updates By Email

Subscribe

The clown culture of over-promising and under-delivering has rubbed off, but
(CAPEX) Clown Week is still coming! The artwork is fabulous.

I did a return appearance on the TelcoDR podcast to talk about, what else, cloud
CAPEX. Even restrained myself to just one small shot at McKinsey.

News NVIDIA’s Mainframe Envy

I have no quibble with the data center as the new unit of compute, but am
skeptical NVIDIA is going to pull off its own vertically integrated stack
spanning silicon, systems and software. Beyond NVIDIA, data center
infrastructure is exciting again! Also we may have to integrate nuclear power
plants into the new stack.

Previous:

Follow the CAPEX: Triangulating NVIDIA, Cloud Power Up

Amazon Desperately Seeks Compression Algorithm for Experience

Previous:

Most Greatest AI Company in Entire Universe Makes Major Strategy Pivot, SEC
Embraces Platformonomics ClownWatch, Warns Amazon and IBM, The AWS Generative AI
Soap Opera, The Amazon AI Inferiority Complex

Aspiring Technology Company Neither Interested in Technology Nor Understands
Technology

The New York Times’ CEO says “we are trying to more firmly establish the Times
as a world-class digital product and tech company.” Yet the database is
seemingly too advanced for them.

Previous:

Living in a Glass House: The New York Times On Tech, Living in a Glass House:
What is the New York Times?, Does the New York Times Editorial Page Ever Talk to
the Business Side of the House?

Private Equity Buys Software Company from Private Equity: Avetta

Usually it takes just one private equity firm to ruin a software company.
Imagine what happens to a software company that gets flipped among private
equity firms (because markups have to happen, even if the market won’t pay
them).

Previous:

Perfidious Private Equity, Software Migration Alert: Jama Software, Software
Migration Alert: Everbridge, Private Equity in Action: VMware (Continued),
Private Equity in Action: VMware, Private Equity Will Buy and Ruin More Software
Companies, Software Migration Alert: VMware (Continued), Software Migration
Alert: VMware, Software Migration Alert: Rover, Software Migration Alert: Avid
Technology, Software Migration Alert: EngageSmart, This Week’s Software
Migration Alerts

Private Equity Buys IT Company from Private Equity: Presidio

It is private equity all the way down:

The company has had several private equity owners since it was founded in 2003.

American Securities acquired Presidio in 2011 for an undisclosed amount and had
sold it to New York-based Apollo Global Management Inc three years later for
about $1.3 billion, Reuters reported at the time.

BC Partners had acquired the company in 2019 and delisted it from the Nasdaq in
a $2.1 billion take-private deal, two years after Apollo Global took it public.

Presidio had acquired several smaller rivals, including Irish IT consulting firm
Arkphire and Charlotte, North Carolina-based ROVE, under BC Partners’ ownership.

Private Equity in Action: VMware Customer Complaints “Unwarranted”

Private equity and the EU are a great pairing and I hope they get to spend a lot
of time together:

Despite Broadcom execs’ efforts to convince people that its changes are
reasonable and will eventually end up financially benefitting stakeholders,
there’s still effort from industry groups to get federal regulators involved
with how Broadcom is running VMware.

As reported by Dutch IT magazine Computable on Friday, representatives from
Beltug, a Belgian CIO trade group; Le Cigref, a French network of companies
interested in digital technology; the CIO Platform Nederland association for
CIOs and CDOs; and VOICE e.V., a German association for IT decisionmakers, sent
a letter [PDF] to European Commission President Ursula von der Leyen and
European Commissioner Thierry Breton on Thursday to “strongly condemn”
Broadcom’s businesses practices and ask the commission to take action.

Previous:

Private Equity in Action: VMware “Unease” Continues, Perfidious Private Equity,
Private Equity in Action: VMware Yet Again, Private Equity in Action: VMware
(Continued), Private Equity in Action: VMware, Private Equity Will Buy and Ruin
More Software Companies, Software Migration Alert: VMware (Continued), Software
Migration Alert: VMware

EU Insanity: Not Learning From Regulatory Failure

So do you admit the errors of your micromanaging regulatory ways or just pack up
your tool bag and take it to a new market?

Too bad about the European telco market. But at least there isn’t much of a
European tech market to ruin.

Previous:

EU Announces DMA Struggle Sessions, EU Insanity: AI Regulatory Suicide, EU
Insanity: AI Energy Suicide, EU Insanity: The Kafka Singularity, EU Insanity:
Mistral Edition, Move Fast and Regulate Things (You Don’t Understand), The EU
Will Continue to Fiddle While Rome Burns, When “Move Fast and Regulate Things”
Breaks Down, AI Regulation: Move Fast and Regulate Things

Tik Tok Traitors: Maria Cantwell (Continued)

Maria Cantwell (D-Washington D-West Taiwan?) looking to poison pill the Tik Tok
bill in the Senate? Feel free to ask her office what is going on.

Previous:

Tik Tok Traitors: Maria Cantwell, Tik Tok’s Call to Arms: Useful Idiots, Fellow
Travelers, Registered and Unregistered Foreign Agents Unite!, Contact Your
Senators

Useful Idiots, Fellow Travelers and Unregistered Foreign Agents: NYC Mayor Eric
Adams

Previous:

Useful Idiots, Fellow Travelers and Unregistered Foreign Agents: Elon Musk Part
II, The Donald, Jeff Yass, Microsoft Bing, DC Lobbyists Part III, Hollywood, DC
Lobbyists Part II, NBA, Warner Bros. Discovery, Germany, DC Lobbyists,
Elon/Twitter, The Hugo Awards, Harvard University, Wall Street, Apple


MORE
June 60
Today, 17:55ongoing by Tim Bray (Feed)
OSQI

I propose the formation of one or more “Open Source Quality Institutes”. An OSQI
is a public-sector organization that employs software engineers. Its mission
would be to improve the quality, and especially safety, of popular Open-Source
software.

Why?

The XZ-Utils backdoor (let’s just say #XZ) launched the train of thought that
led me to this idea. If you read the story, it becomes obvious that the key
vulnerability wasn’t technical, it was the fact that a whole lot of Open-Source
software is on the undermaintained-to-neglected axis, because there’s no
business case for paying people to take care of it. Which is a problem, because
there is a strong business case for paying people to attack it.

There are other essential human activities that lack a business case, for
example tertiary education, potable water quality, and financial regulation. For
these, we create non-capitalist constructs such as Universities and Institutes
and Agencies, because society needs these things done even if nobody can make
money doing them.

I think we need to be paying more attention to the quality generally, and safety
especially, of the Open-Source software that has become the underlying platform
for, more or less, our civilization. Thus OSQI.

They’re out to get us

For me, the two big lessons from #XZ were first, the lack of resources
supporting crucial Open-Source infrastructure, but then and especially, the
demonstration that the attackers are numerous, skilled and patient. We already
knew about numerous and skilled but this episode, where the attacker was already
well-embedded in the project by May 2022, opened a few eyes, including mine.

The advantage, to various flavors of malefactor, of subverting core pieces of
Open-Source infrastructure, is incalculable. #XZ was the one we caught; how many
have we missed?

What’s OSQI?

It’s an organization created by a national government. Obviously, more nations
than one could have an OSQI.

The vast majority of the staff would be relatively-senior software engineers,
with a small percentage of paranoid nontechnical security people (see below).
You could do a lot with as few as 250 people, and the burdened cost would be
trivial for a substantial government.

Since it is a matter of obvious fact that every company in the world with
revenue of a billion or more is existentially dependent on Open Source, it would
be reasonable to impose a levy of, say, 0.1% of revenue on all such companies,
to help support this work. The money needn’t be a problem.

Structure

The selection of software packages that would get OSQI attention would be left
to the organization, although there would be avenues for anyone to request
coverage. The engineering organization could be relatively flat, most people
giving individual attention to individual projects, then also ad-hoc teams
forming for tool-building or crisis-handling when something like #XZ blows up.

Why would anyone work there?

The pay would be OK; less than you’d make at Google or Facebook, but a decent
civil-service salary. There would be no suspicion that your employer is trying
to enshittify anything; in fact, you’d start work in the morning confident that
you’re trying to improve the world. The default work mode would be remote, so
you could live somewhere a not-quite-Google salary would support a very
comfortable way of life. There would be decent vacations and benefits and
(*gasp*) a pension.

And there is a certain class of person who would find everyday joy in peeking
and poking and polishing Open-Source packages that are depended on by millions
of programmers and (indirectly) billions of humans. A couple of decades ago I
would have been one.

I don’t think recruiting would be a problem.

So, what are OSQI’s goals and non-goals?

Goal: Safety

This has to come first. If all OSQI accomplishes is the foiling of a few
#XZ-flavor attacks, and life becoming harder for people making them, that’s just
fine.

Goal: Tool-building

I think it’s now conventional wisdom that Open Source’s biggest attack surfaces
are dependency networks and build tools. These are big and complex problems, but
let’s be bold and set a high bar:

Open-Source software should be built deterministically, verifiably, and
reproducibly, from signed source-code snapshots. These snapshots should be free
of generated artifacts; every item in the snapshot should be human-written and
human-readable.

For example: As Kornel said, Seriously, in retrospect, #autotools itself is a
massive supply-chain security risk. No kidding! But then everyone says “What are
you gonna do, it’s wired into everything.”

There are alternatives; I know of CMake and Meson. Are they good enough? I don’t
know. Obviously, GNU AutoHell can’t be swept out of all of the fœtid crannies
where it lurks and festers, but every project from which it is scrubbed will
present less danger to the world. I believe OSQI would have the scope to make
real progress on this front.

Non-goal: Features

OSQI should never invest engineering resources in adding cool features to
Open-Source packages (with the possible exception of build-and-test tools). The
Open-Source community is bursting with new-features energy, most coming from
people who either want to scratch their own itch or are facing a real blockage
at work. They are way better positioned to make those improvements than anyone
at OSQI.

Goal: Maintenance

Way too many deep-infra packages grow increasingly unmaintained as people age
and become busy and tired and sick and dead. As I was writing this, a plea for
help came across my radar from Sebastian Pipping, the excellent but unsupported
and unfunded maintainer of Expat, the world’s most popular XML parser.

And yeah, he’s part of a trend, one that notably included the now-infamous
XZ-Utils package.

And so I think one useful task for OSQI would be taking over (ideally partial)
maintenance duties for a lot of Open-Source projects that have a high ratio of
adoption to support. In some cases it would have to take a lower-intensity form,
let’s call it “life support”, where OSQI deals with vulnerability reports but
flatly refuses to address any requests for features no matter how trivial, and
rejects all PRs unless they come from someone who’s willing to take on part of
the maintenance load.

One benefit of having paid professionals doing this is that they will blow off
the kind of social-engineering harassment that the #XZ attacker inflicted on the
XZ-Utils maintainer (see Russ Cox’s excellent timeline) and which is
unfortunately too common in the Open-Source world generally.

Goal: Benchmarking

Efficiency is an aspect of quality, and I think it would be perfectly reasonable
for OSQI to engage in benchmarking and optimization. There’s a non-obvious
reason for this: #XZ was unmasked when a Postgres specialist noticed performance
problems.

I think that in general, if you’re a bad person trying to backdoor an
Open-Source package, it’s going to be hard to do without introducing performance
glitches. I’ve long advocated that unit and/or integration tests should include
a benchmark or two, just to avert well-intentioned performance regressions; if
they handicap bad guys too, that’s a bonus.

Goal: Education and evangelism

OSQI staff will develop a deep shared pool of expertise in making Open-Source
software safer and better, and specifically in detecting and repelling multiple
attack flavors. They should share it! Blogs, conferences, whatever. It even
occurred to me that it might make sense to structure OSQI as an educational
institution; standalone or as a grad college of something existing.

But what I’m talking about isn’t refereed JACM papers, but what my Dad, a
Professor of Agriculture, called “Extension”: Bringing the results of research
directly to practitioners.

Non-goal: Making standards

The world has enough standards organizations. I could see individual OSQI
employees pitching in, though, at the IETF or IEEE or W3C or wherever, with work
on Infosec standards.

Which brings me to…

Non-goal: Litigation

Or really any other enforcement-related activity. OSQI exists to fix problems,
build tools, and share lessons. This is going to be easier if nobody (except
attackers) sees them as a threat, and if staff don’t have to think about how
their work and findings will play out in court.

And a related non-goal…

Non-goal: Licensing

The intersection between the class of people who’d make good OSQI engineers and
those who care about Open-Source licenses is, thankfully, very small. I think
OSQI should accept the license landscape that exists and work hard to avoid
thinking about its theology.

Non-goal: Certification

Once OSQI exists, the notion of “OSQI-approved” might arise. But it’d be a
mistake; OSQI should be an engineering organization; the cost (measured by
required bureaucracy) to perform certification would be brutal.

Goal: Transparency

OSQI can’t afford to have any secrets, with the sole exception of
freshly-discovered but still-undisclosed vulnerabilities. And when those
vulnerabilities are disclosed, the story of their discovery and characterization
needs to be shared entirely and completely. This feels like a bare-minimum basis
for building the level of trust that will be required.

Necessary paranoia

I discussed above why OSQI might be a nice place to work. There will be a
downside, though; you’ll lose a certain amount of privacy. Because if OSQI
succeeds, it will become a super-high-value target for our adversaries. In the
natural course of affairs, many employees would become committers on popular
packages, increasing their attractiveness as targets for bribes or blackmail.

I recall once, a very senior security leader at an Internet giant saying to me
“We have thousands of engineers, and my job requires me to believe that at least
one of them also has another employer.”

So I think OSQI needs to employ a small number of paranoid traditional-security
(not Infosec) experts to keep an eye on their colleagues, audit their finances,
and just be generally suspicious. These people would also worry about OSQI’s
physical and network security. Because attackers gonna attack.

Pronunciation

Rhymes with “bosky”, of course. Also, people who work there are OSQIans. I’ve
grabbed “osqi.org” and will cheerfully donate it in the long-shot case that this
idea gets traction.

Are you serious?

Yeah. Except for, I no longer speak with the voice of a powerful employer.

Look: For better or for worse, Open Source won. [Narrator: Obviously, for
better.] That means it has become crucial civilizational infrastucture, which
governments should actively support and maintain, just like roads and dams and
power grids.

It’s not so much that OSQI, or something like it, is a good idea; it’s that not
trying to achieve these goals, in 2024, is dangerous and insane.


MORE
April 30
24.6.2024, 6:00SEI Blog (Feed)
Applying Large Language Models to DoD Software Acquisition: An Initial
Experiment
This SEI Blog post illustrates examples of using LLMs for software acquisition
in the context of a document summarization experiment and codifies the lessons
learned from this experiment and related work on applying generative AI to
software engineering.

MORE
April 10
23.6.2024, 18:04The Architect Elevator (Feed)
The Serverless Illusion
Abstractions can become illusions. Is Serverless one of them?

MORE
April 170
Today, 17:55ongoing by Tim Bray (Feed)
A057X

Yes, “A057X” is cryptic, but my new lens’s official monicker is “150-500mm
F/5-6.7 Di III VC VXD” so let’s stick with that part number. It’s from Tamron
and this is the Fujifilm X-Mount variation. Lens-geeking is my favorite part of
photo-geeking and it’s great that more manufacturers are opening up to
third-party lens builders.

Last May I wrote that I wanted a big-ass super-telephoto and now I have one.
Let’s start with a little comparo. Here is (roughly) the same 2km-away landscape
shot on the decades-old Tokina 400mm I’d been using since 2009, and on the new
Tamron.

If you care about this sort of thing you might want to enlarge these. A person
is visible in the bottom picture, and another if you’re using Lightroom on a 4K
screen.

Now let’s be honest; the color and flavor of the earlier picture is nicer,
because the sun was just right; that’s why I strapped on the old glass. But the
new-lens picture shows that yes, we do still make progress in analog
technologies, and given the same light, there’d be more you could do with with
today’s lens.

Anyhow, here’s what it looks like.

That’s on a Fujifilm X-T2, one of the bulkier of Fuji’s X-cameras. What’s not
instantly obvious is that the camera and lens are sitting on the lens’s tripod
shoe. That camera is now eight years old and needs to be replaced, but I’m not
fully won over by the latest X-cams and the lens was an easier trigger to pull.

The reviews all said “Considering what it does, it’s amazingly small and light!”
Maybe, but in fact it’s a big freakin’ heavy hunk of metal and glass. A tripod
really helps.

For the birds

Tripod? But everyone seems to think that this kind of lens is for shooting birds
in flight. So I took it to our cabin this weekend to test that hypothesis. Thus
I learned that you really can’t shoot birds unless you’re hand-holding the
camera. And even then, you can’t unless you’ve been practicing. I managed to get
one picture of a bird in flight, but it was just a seagull and not a terribly
handsome one either.

Then a couple of visitors settled at the top of a nearby Douglas Fir. Here’s
one. Yes, the sky was that blue.

Isn’t it handsome? If you look close, though, its tail is jammed against a
branch. But then it bent over to peer out at something.

Aren’t those feathers beautiful? This was a big-ass tree and I wasn’t right next
to it, either. Yay Tamron.

Little, big

Turns out this thing can focus relatively close-in for an item of its ilk, so
you can do, um what would one call it, macro-at-a-distance?

That’s a teeny little blossom. But when I’m looking out over the water, I always
end up taking pictures of the mountains on the other side.

That one is a damn long way away. The picture suffers from being reduced to fit
into your browser. I wish I could give everyone in the world Lightroom and a
good 4K monitor.

Note that…

None of the pictures via this lens could have been captured on any mobile-phone
camera in the world. You have to go pretty far these days to get into that
territory.


MORE
April 10
22.6.2024, 19:00Platformonomics (Feed)
Platformonomics TGIF #45: March 29, 2024

Platformonomics TGIF is a weekly roll-up of links, comments on those links, and
perhaps a little too much tugging on my favorite threads.

Get Platformonomics Updates By Email

Subscribe

The last two weeks have been eventful, so we have a lot to cover after my
absence. Buckle up!

News Amazon CAPEX Plans

Bloomberg looks at cloud CAPEX and the accompanying “mad scramble” for
electricity to power those data centers.

A $150 billion number over the next 15 years is unimpressive and inadequate ($10
billion a year). AWS spent $133 billion on CAPEX over the last nine years and
averaged $22 billion a year over the last four years. That’s way too low to be
competitive (GPUs are expensive!). But after their first ever decline in CAPEX
spending in 2023, who knows what is going on at AWS.

Previous:

Follow the CAPEX: Cloud Table Stakes 2023 Retrospective, Follow the CAPEX:
Triangulating NVIDIA, Follow the GPUs: Cloud CAPEX, more CAPEX, Cloud Power Up,
Behold the Nuclear Cloud, Nuclear Powered Cloud Data Centers Are Happening, A
Nuclear-Powered Cloud, When Nuclear-Powered Data Centers?

Microsoft/OpenAI CAPEX Plans

It isn’t $7 trillion but still a good start. Presumably nuclear-powered. And
perhaps off-world? Orbital computing is on my list of topics to chronicle.

Previous:

Follow the CAPEX: Cloud Table Stakes 2023 Retrospective, Follow the CAPEX:
Triangulating NVIDIA

Antitrust Incoherence: Apple Edition

Hipster antitrust remains intellectually incoherent and provides no clear lines
for anyone on the playing field. The Apple case deserves more than a few
one-liners, but I have learned the audience for musings on antitrust is quite
limited. But I may have to write a follow-up to my last treatise on the current
state of antitrust. Or at least dig into the many ridiculous connections being
made around this case between Apple and Microsoft.

Previous:

Antitrust Incoherence: Spotify Edition, Antitrust Incoherence: Roomba Edition,
The Incoherence Doctrine: The FTC Strikes Again, A New Antitrust Doctrine, The
DOJ Play at Home Game

EU Announces DMA Struggle Sessions

The struggle sessions will continue until Europe has a tech industry (i.e.
indefinitely).

Previous:

EU Insanity: AI Regulatory Suicide, EU Insanity: AI Energy Suicide, EU Insanity:
The Kafka Singularity, EU Insanity: Mistral Edition, Move Fast and Regulate
Things (You Don’t Understand), The EU Will Continue to Fiddle While Rome Burns,
When “Move Fast and Regulate Things” Breaks Down, AI Regulation: Move Fast and
Regulate Things

The Most Important Issue in AI

The topic everyone is talking about is how the AI market landscape is unfair to
venture capitalists. High capital requirements? Unfair! The importance of
distribution and customer feedback? Unfair! Incumbents paying attention? Unfair!
We all need to focus on how to make life easier for our most important class:
VCs.

Previous:

The VC Antitrust Doctrine, VC-induced Indigestion, The Looming Layoffs for
Venture Capitalists, Ghost VCs, Useful Idiots, Fellow Travelers and Unregistered
Foreign Agents: Jeff Yass

Thought Leadership (in Mascots) Shot: Chaser:

Know your strengths! Don’t pretend to be a software company if you’re not one!

Previous:

Salesforce Rallies Its Deep Bench of AI Experts (and Cartoon Characters),
Bluster, Bombast and Bullshit: Just Another Day at Salesforce, Salesforce’s
Shamelessness is Staggering, Salesforce Gonna Salesforce, The Spectacle that is
Dreamforce

Most Greatest AI Company in Entire Universe Makes Major Strategy Pivot

“Generative AI is poised to be the most transformational technology of our time,
and we believe our strategic collaboration with Anthropic will further improve
our customers’ experiences, and look forward to what’s next,” said Swami
Sivasubramanian, vice president of data and AI at AWS cloud provider.

Amazon spends $2.75 billion on AI startup Anthropic in its largest venture
investment yet

Amazon finally starts talking about their impact on the generative AI market in
the future, not present tense. Progress! And I’ve completely lost track of how
many steps we are into the AI marathon!

Previous:

SEC Embraces Platformonomics ClownWatch, Warns Amazon and IBM, The AWS
Generative AI Soap Opera, The Amazon AI Inferiority Complex

Cloud Repatriation is Going Just Great: Equinix

Prime cloud repatriation beneficiary Equinix has had to resort to the WorldCom
playbook in order to put up their mediocre growth numbers? I still need to make
fun of the AI data center bubble.

Previous:

New Cloud Repatriation Hypothesis, Cloud Repatriation Where Art Thou?, Cloud
Repatriation: The Search Continues, Cloud Repatriation: Still Not a Thing,
Platformonomics Repatriation Index – Q1 2023: Surf’s Up?

Related:

Web3 is Going Just Great (a recurring look at cloud repatriation’s sibling VC
thought leadership scheme)

Product Market Fit for Google Gemini?

As clichéd as “Orwellian” has become, Orwell remains underappreciated.

Previous:

Google: Still Not Code Red?, Google Goes Off the Guardrails, AI Marketing
Hallucinations: Google Edition, The Google Pivot Away from “AI-Safety First”,
“Google has more than 160 people working full time“, Google: Code Red Now?

Victory Lap: Instability AI

Called it.

Previous:

Instability AI, Stability AI: Not So Stable, Seeing Post Facto Red Flags Ex Ante

Software Migration Alert: Jama Software

As we all now find ourselves chanting, when private equity comes amalgamating,
it is time to start migrating.

Previous:

Perfidious Private Equity, Software Migration Alert: Everbridge, Private Equity
in Action: VMware (Continued), Private Equity in Action: VMware, Private Equity
Will Buy and Ruin More Software Companies, Software Migration Alert: VMware
(Continued), Software Migration Alert: VMware, Software Migration Alert: Rover,
Software Migration Alert: Avid Technology, Software Migration Alert:
EngageSmart, This Week’s Software Migration Alerts

MariaDB Saved from Clutches of Private Equity

Software Migration Alert averted!

Previous:

Perfidious Private Equity

Private Equity in Action: VMware “Unease” Continues

Broadcom’s acquisition of VMware is going just great! Unless you’re a customer
and apparently “confused”.

VMware has quickly become the poster child for why you must start migrating when
private equity comes amalgamating. You were warned.

Previous:

Perfidious Private Equity, Private Equity in Action: VMware Yet Again, Private
Equity in Action: VMware (Continued), Private Equity in Action: VMware, Private
Equity Will Buy and Ruin More Software Companies, Software Migration Alert:
VMware (Continued), Software Migration Alert: VMware

Related:

Broadcom-owned VMware kills the free version of ESXi virtualization software

Tik Tok Traitors: Maria Cantwell

Maria Cantwell (D-Washington D-West Taiwan?) seems to be key resistance to the
Tik Tok bill in the Senate. Contact her office. Maybe they’ll respond to you — I
didn’t warrant a response.

Previous:

Tik Tok’s Call to Arms: Useful Idiots, Fellow Travelers, Registered and
Unregistered Foreign Agents Unite!, Contact Your Senators

The Meme War with China is Going Just Great

(For the record, this is my first non-sarcastic use of “going just great”)

This is the war with West Taiwan we want to have:

Previous:

One, Two, Three, Four: We Want a Meme War


MORE
June 60
24.6.2024, 6:00SEI Blog (Feed)
5 Recommendations to Help Your Organization Manage Technical Debt
This SEI Blog post summarizes recommendations arising from an SEI study that
apply to the Department of Defense and other development organizations seeking
to analyze, manage, and reduce technical debt.

MORE
March 250
Today, 14:28Phil Windley's Technometria (Feed)
Decentralizing Energy

Summary: The Accidental Superpower discussed a post-Bretton Woods world where
easy, safe shipping cannot be assumed. In this "scared new world," renewables
change the dynamics of energy flows by moving production closer to energy use.

My wife, Lynne, recently gave me a copy of Peter Zeihan's book, The Accidental
Superpower: Ten Years On. The book was originally published in 2014, but Zeihan
has updated it by inserting chapters talking about what he got right in 2014,
what he got wrong, and why. The focus of the book is geopolitics—how geography
and demographics shapes the world order—and how Bretton Woods changed that in
significant ways. The book makes the case that so much of what made Bretton
Woods useful to the US and why the US engaged with the rest of the world for the
70 years following World War II is changing. As it changes the free trade system
enabled by Bretton Woods is also changing. This will have significant impact on
every country in the world.

Much of what changes has to do with energy. One of the things1 Zeihan got right
was his assertion that unlike much of the rest of the developed world, the US
doesn't need to import energy—specifically oil—we are a net energy importer.
This changes the dynamic wherein the US is willing to be the protector of
shipping lanes for the entire world. As a result, the future could see a US that
has the luxury of ignoring events in the Middle East, Ukraine, and elsewhere,
whereas Europe (to take just one example) cannot. The book is full of other
interesting predictions and conclusions just like this one. I encourage you to
read it if you find this as fascinating as I do.

Zeihan makes a big deal of shale oil production, which accounted for 66% of US
production in 2022. But as I read this, I was thinking about renewables. As I
wrote in 2020, I've gone in big on solar power at my house, love my EV, and have
replaced most things in the house (like the furnaces) with versions that run on
electricity.  I did this because it made my life easier and saves me money. The
fact that it's good for the environment is a bonus.

But, solar and wind are not just renewable, they also allow energy production to
be decentralized in ways oil and natural gas can't. Oil and natural gas deposits
are where they are. Some countries are blessed with them and the others have to
buy from those countries. And they're often far away, requiring shipping through
potentially hostile waters. But that's not true of renewables. They can usually
be built and located where ever the need is2. This changes geopolitical equation
in significant ways. Areas of the world that are not energy independent, like
Europe, are moving toward renewables too slowly to prevent future energy shocks.
The problem with renewables is that they're long-lead items—they take years to
plan and bring online.

Petroleum and Bretton Woods enabled the modern world, providing portable,
storable sources of energy that could easily and safely move to where ever it
was needed.3 If we are indeed at the end of the Bretton Woods era, the world is
in for significant changes as it adjusts to a life where free trade, and easy
access to petroleum-based energy, cannot be assumed. Moving energy production
closer to the places it's used is one strategy for dealing with this
world-altering disruption. Buckle up.

Notes There are other things that are important to the books overall conclusion
besides energy. I'm just cherry picking that because I was thinking about it.
For example, the US is largely self-sufficient from an overall import/export
standpoint. We don't import nearly as much as many other countries and could
replace what we do import relatively easily. It's not just renewables. Nuclear
power can also be located closer to demand than an oil deposit. I started my
career as a nuclear metallurgist, so I'm a fan. I think many countries are going
to be sorry they've closed nuclear plants and made them too hard to construct
profitably. The feats of engineering that have enabled these energy flows is
truly astounding.

Photo Credit: Oil Tanker at Sunset from Terski (Pixabay)

Tags: climate+change solar geopolitics energy


MORE
March 190
24.6.2024, 6:00SEI Blog (Feed)
API Security through Contract-Driven Programming
This blog post explores contract programming and specifically how that applies
to the building, maintenance, and security of APIs.

MORE
March 180
Today, 17:55ongoing by Tim Bray (Feed)
Bye, Allyson

She’s gone. She lived well. We’ll miss her.

We’ve known Ms McGrane since 2005, when she was a co-conspirator on the
wonderful little local Northern Voice blogging conference. We worked on other
stuff together and hung out now and then and carpooled to the Prairies once and
I can’t remember ever getting the slightest bit upset with her.

Here is a good-bye note from her partner Shane. If you want to leave a note
somewhere, leave it there.

Ally (rhymes with “valley”) was a fine dog-parent and a strong grant-writer and
a first-rate teacher and a connoisseur of fine cooking equipment and Canadian
football. If you’ve been to much in the way of Vancouver theatre and dance
events over the years, there’s a good chance that she produced the event or
secured its funding or educated the people who did those things.

I remember having coffee with her a couple years ago, she advising me on one of
my projects, laughing together at the Byzantine complexities of granting
bureaucracies and the childlike money-obliviousness of arts leaders and the
excellence of the coffee on that morning. Easy to be with.

Mesothelioma is a bitch; 8% 5-year survival rate, and there wasn’t that much
they could do for her by the time they got the diagnosis right. We visited her
last week and she was herself, cynical about her situation but it seemed more or
less at peace.

I won’t miss her as much as the dogs will, but there’s still a gap in my life.


MORE
March 180
22.6.2024, 19:00Platformonomics (Feed)
Platformonomics TGIF #44: March 15, 2024

Platformonomics TGIF is a weekly roll-up of links, comments on those links, and
perhaps a little too much tugging on my favorite threads.

Get PLATFORMONOMICS Updates By Email

Subscribe

No newsletter next week. But I hope to get the Clown Week update out before
that.

News Tik Tok’s Call to Arms: Useful Idiots, Fellow Travelers, Registered and
Unregistered Foreign Agents Unite!

I have been dutifully chronicling people and organizations who, intentionally or
unintentionally, advance the aims of the Chinese Communist Party in the West.
They are all now being mobilized for the Tik Tok battle.

The pending legislation gives Tik Tok the choice to abandon CCP control or
abandon doing business in America. The degree to which Tik Tok and the CCP (and
their speaking points are indistinguishable) have invoked a Code Red and gone to
the mattresses is a masterful case study for Google on how to muster for
existential battle (to cross the streams of some unrelated threads). They are
spraying cash across the political ecosystem and there are armies of mercenary
lobbyists and politicians gleeful to take it.

Noah Smith captures the situation: “This is a test of whether America is a real
nation, or an anarchistic playground for totalitarian empires.”

Lenin once observed of the lobbying industry: “The Capitalists will sell us the
rope with which we will hang them.”

I find the case for CCP control of Tik Tok and repeated misconduct persuasive.
And Tik Tok is banned in West Taiwan along with all other western social media
apps, so basic reciprocity is ample justification for this legislation.

I hope someone is collecting a comprehensive list of everyone taking coin from
Tik Tok and the CCP. http://www.TikTokTraitors.com looks available.

Contact Your Senators

The fate of the Tik Tok legislation will be decided in the Senate, where a a
lobbying onslaught has been unleashed.

For Seattle area readers, Washington Senator Maria Cantwell is at the center of
this battle as chair of the Senate Commerce Committee. Her former staff
represent Tik Tok, which raises questions about what influence they have with
their former boss.

Should you be interested in her position on the Tik Tok bill and how she is or
is not influenced by former staffers, you can Share Your Thoughts with her
office. Or contact other Senators.

Related:

TikTok ban faces uncertain fate in Senate; Washington state reps vote against
House bill

One, Two, Three, Four: We Want a Meme War

We are in an ideological war, not unlike the Cold War, with West Taiwan. The
Chinese Communist Party seeks to upend the global system and eliminate threats
to CCP rule, one of which they fear most is the very existence of alternative
forms of government. The CCP is not content to stay home and enjoy their
communist paradise, but aggressively meddle, coerce, and suborn across the
globe.

This war is a battle of ideals, particularly over the Western principles of
individual liberty, democratic rule, free markets and free expression. (I will
stipulate it is unclear if the West still believes in those principles, but
suggest they should).

Memes are the distilled essence of free expression, intrinsically a weapon of
free expression, and therefore an ideal medium for waging this ideological war.

There is nothing the CCP fears more than free expression. Yet they are the
thinnest-skinned aspiring great power the world has ever seen, so cannot help
but be drawn into battles on hostile territory. We, however, are a meme
juggernaut. A shitposting superpower.

Watching CCP propaganda arms straight-faced championing Western ideals like free
speech and fair competition is the battle we want to have.

Previous:

Xi Jinping Thought on Wall Street Running Dogs, The Perils of Working for a Cult
of Personality, Press Releases We’d Like to See, Best Nickname for Emmanuel
Macron, Winnie-the-Pooh: Lover of Honey, Champion of Freedom, Missing the
(Bamboo) Forest for the (Apple) Trees

Clown Week Preview: Oracle CAPEX

Oracle had a nice upward CAPEX curve going there for two years, exactly what you
want to see from a wannabe cloud provider. But that seems to be over. Stay tuned
for a deeper look at what is going on with Oracle’s CAPEX spend. As always, it
is hard to distinguish between what is cluelessness and what are lies.

It is tough when you take the free cash flow hit by ramping your CAPEX, yet
still find yourself falling ever further behind. A tiny solace is that at least
they’re not IBM.

Previous:

Follow the CAPEX: Keeping Up With The Clowns, Follow the CAPEX: Separating the
Clowns from the Clouds, Follow the CAPEX: Clown Watch, This Week in Securities
Fraud

EU Insanity: AI Regulatory Suicide

So impressive to see the future so clearly. And even more impressive given a
lack of proximity to the nascent technology being regulated.

Previous:

EU Insanity: Mistral Edition, Move Fast and Regulate Things (You Don’t
Understand), The EU Will Continue to Fiddle While Rome Burns, When “Move Fast
and Regulate Things” Breaks Down, AI Regulation: Move Fast and Regulate Things

Related:

Breton Says Atos’s Problems Emerged After He Quit to Join EU

EU Insanity: AI Energy Suicide

The US faces many challenges in expanding power generation and distribution and
getting back on the road to energy abundance. But at least we’re not Europe.
From a good Semianalysis piece on the AI Datacenter Energy Dilemma ($):

Some countries and regions such as the US will be able to respond flexibly with
a low electrical grid carbon intensity, low-cost fuel sources with supply
stability, while others such as Europe will be effectively handcuffed by
geopolitical realities and structural regulatory constraints on power. 

In Western Europe, electricity generation has been slowly declining, with a 5%
drop cumulatively over the past five years. 

Given Europe’s energy situation, the EU average industrial tariff reached $0.18
USD/kWh in 2022, with the UK at $0.235 USD/kWh and datacenter heavyweight
Ireland at $0.211 USD/kWh, nearly triple the electricity cost in the US. Like
Asia, Europe imports over 90% of its gas in the form of LNG, mainly sourced from
the Middle East (and also still from Russia, despite the ongoing war), so their
entire industrial base, not just Datacenters, is subject to geopolitical risk,
as most readers will vividly remember from the onset of the war in Ukraine.
Given the political and geopolitical realities, adding a massive amount of power
generation capacity to host the AI Datacenter boom in Europe would be very
challenging.

I made fun of Mistral when it was funded, and then made fun of the EU’s reaction
to Mistral wanting to actually build a business. But the French for a change not
being enthusiastic about the EU’s regulatory agenda makes sense. The French 1.)
have Mistral (even if it is just Llama 2) 2.) are strong at math and 3.) lucked
out that Putin invaded Ukraine, narrowly averting the shutdown of their nuclear
power plants and following the German path to energy suicide.

Previous:

Cloud Power Up, Behold the Nuclear Cloud, Nuclear Powered Cloud Data Centers Are
Happening, A Nuclear-Powered Cloud, When Nuclear-Powered Data Centers?

EU Insanity: The Kafka Singularity

Forget turtles, it is regulators all the way down. What if the EU itself ground
to a halt after getting caught in its own regulatory machinery?

Previous:

EU Insanity: Mistral Edition, Move Fast and Regulate Things (You Don’t
Understand), The EU Will Continue to Fiddle While Rome Burns, When “Move Fast
and Regulate Things” Breaks Down, AI Regulation: Move Fast and Regulate Things

Useful Idiots, Fellow Travelers and Unregistered Foreign Agents: Elon Musk Part
II

With a Tesla factory in West Taiwan, Elon is compromised on this and all other
issues relating to West Taiwan. He and his mouthpieces are misrepresenting the
Tik Tok legislation as a free speech issue, which is ironic yet totally
consistent.

Previous:

Useful Idiots, Fellow Travelers and Unregistered Foreign Agents: The Donald,
Jeff Yass, Microsoft Bing, DC Lobbyists Part III, Hollywood, DC Lobbyists Part
II, NBA, Warner Bros. Discovery, Germany, DC Lobbyists, Elon/Twitter, The Hugo
Awards, Harvard University, Wall Street, Apple


MORE
April 230
18.6.2024, 16:30All Things Distributed (Feed)
District heating: Using data centers to heat communities
An inside look at the Tallaght District Heating Scheme, where Heat Works is
using recycled heat from an AWS data center to warm a community in Dublin,
Ireland.

MORE
March 130
Today, 9:19mark nottingham (Feed)
There Are No Standards Police

It happens fairly often. Someone brings a proposal to a technical standards body
like the IETF and expects that just because it becomes an RFC, people will adopt
it. Or they’ll come across a requirement in an RFC and expect it to be enforced,
perhaps with some kind of punishment. Or they’ll get angry that people don’t pay
attention to an existing standard and do their own thing. This is so common that
there’s a ready response widely used by IETF people in these situations:

“There are no standards police.”

In other words, even if you do consider Internet standards to be a regulatory
force, there is no enforcement mechanism. One of their key characteristics is
that they’re voluntary. No one forces you to adopt them. No one can penalise you
for violating a MUST; you have to want to conform.

Of course, you can still feel compelled to do so. If an interoperability
standard gets broad adoption and everyone you want to communicate with expects
you to honour it, you don’t have many options. For example, if you want to have
a Web site, you need to interoperate with browsers; most of the time, they write
down what they do in standards documents, and so you’ll need to conform to them.

But that’s the successful path. For every HTTP or HTML or TCP, there are
hundreds of IETF RFCs, W3C Recommendations, and other standards documents that
haven’t caught on – presumably much to their authors’ dismay. Adopting and using
those documents was optional, and the market spoke: there wasn’t interest.

This aspect of the Internet’s standards has been critical to its success. If
people were forced to adopt a specification just because some body had blessed
it, it would place immense pressure on whatever process was used to create it.
The stakes would be high because the future of the Internet would be on the
line: businesses would play dirty; trolls would try to subvert the outcomes;
governments would try to steer the results.

Of course, all of those things already happen in Internet standards; it’s just
that the stakes are much lower.

So, voluntary adoption is a proving function – it means that not all of the
weight of getting things right is on the standardisation process, and that
process can be lighter than, for example, that used by the governments or the
United Nations (I’ll get back to that in a minute). That’s important, because it
turns out that it’s already incredibly difficult to create useful, successful,
secure, private, performant, scalable, architecturally aligned technical
specifications that change how the Internet works within all of the other
natural constraints encountered; it’s threading-the-needle kind of stuff. And we
need to be able to fail.

Historically, voluntary standards have been encouraged by governments in their
purchasing and competition policies - for example, OMB Circular A-119, EU
Regulation 1025/2012, and the EC guidelines on horizontal agreements. Standards
bodies are a ‘safe space’ where competitors can cooperate without risking
competition enforcement, so long as they follow a set of rules – and one of the
biggest rules is that adoption should be voluntary, not mandatory or coerced (at
least by those setting the standard).

But it’s no secret that the policy landscape for the Internet has changed
drastically. Now, there is increasing interest in using interoperability
standards as a mechanism to steer the Internet. Academics are diving deep into
the cultures and mechanisms of technical standards. Civil society folks are
coming to technical standards bodies and trying to figure out how to incorporate
human rights goals. Regulation is coming, and policy experts are trying to
figure out how to get involved too.

This influx has caused concern that that these relative newcomers are mistakenly
focusing on standards as a locus of power when, in fact, the power is expressed
in the adoption of a standardised technology. For example, Geoff Huston recently
wrote an opinion piece along these lines.

I have no doubt that some still come to the IETF and similar bodies with such
misapprehensions; we still have to remind people that ‘there are no standards
police’ on a regular basis. However, I suspect that at least the policy people
(including regulators) largely understand that it’s not that simple.

That’s because modern regulators are very aware that there are many influences
on a regulatory space. They want to learn about the other forces acting on their
target, as well as persuade and inform. Similarly, those who are involved in
policymaking are intensely aware of the diffuse nature of power. In short, their
world view is more sophisticated than people give them credit for.

(All that said, I’m still interested and a bit nervous to see what Global
Digital Compact contains when it becomes public.)

Another concern is that governments might try to influence Internet standards to
suit their purposes, and then exert pressure to make the results mandatory –
short circuiting the proving function of voluntary standards.

Avoiding that requires separating the legal requirement from the standards
effort, to give the latter a chance to fail. For example, MIMI may or may not
succeed in satisfying the DMA requirement for messaging interop. It is an
attempt to establish voluntary standards that, if successful in the market,
could satisfy legal regulatory requirements without using a preselecting
standards venue.

Of course, that pattern is not new – for example, accessibility work in the W3C
is the basis of many regulatory requirements now, but wasn’t considered (AFAIK)
by regulators until many years after its establishment.

Because of the newly intense focus on regulating technology, there’s likely to
be increasing pressure on such efforts: both the pace and volume of
standardisation will need to increase to meet the requirements that the
standards bodies want to attempt to address. I suspect aligning the timelines
and risk appetites of standards bodies and regulators are going to be some of
the biggest challenges we’ll face if we want more successes.

So right now I believe the best way forward is to create ‘rails’ for
interactions with legal regulators – e.g., improved communication, aligned
expectations, and ways for an effort to be declined or to fail without
disastrous consequences. Doing that will require some capacity building on the
parts of standards bodies, but no fundamental changes to their models or
decision-making processes.

This approach will not address everything. There are some areas where at least
some regulators and the Internet standards community are unlikely to agree.
Standards-based interoperability may not be realistically achievable in some
instances, because of how entrenched a proprietary solution is. Decentralising a
proprietary solution can face many pitfalls, and may be completely at odds with
a centralized solution that already has broad adoption. And, most fundamentally,
parties that are not inclined to cooperate can easily subvert a voluntary
consensus process.

However, if things are arranged so that when conforming to a voluntary consensus
standard that has seen wide review and market adoption is considered to be prima
facie evidence of conformance to a regulatory requirement, perhaps we do
sometimes have standards police, in the sense that legal requirements can be
used to help kickstart standards-based interoperability where it otherwise
wouldn’t get a chance to form.


MORE
March 130
24.6.2024, 6:00SEI Blog (Feed)
Example Case: Using DevSecOps to Redefine Minimum Viable Product
This SEI blog post, authored by SEI interns, describes their work on a
microservices-based software application, an accompanying DevSecOps pipeline,
and an expansion of the concept of minimum viable product to minimum viable
process.

MORE
March 120
Today, 17:55ongoing by Tim Bray (Feed)
Play My Music

When I’m away from home, I still want to listen to the music we have at home
(well, I can live without the LPs). We had well over a thousand CDs so that’s a
lot of music, 12,286 tracks ripped into Apple Lossless. Except for a few MP3s
from, well, never mind. This instalment of the De-Google Project is about ways
to do that with less Big-Tech involvement.

The former Google Play Music, now YouTube Music, allowed you to load your tunes
into the cloud and play them back wherever your phone or computer happened to
be. Except for it used to be easy to upload — just point the uploader at your
iTunes library — and now it’s hard, and then Google removed YouTube Music’s
shuffle-your-uploads feature from Android Auto. Also they fired a bunch of
YouTube Music contractors who were trying to unionize. So screw ’em.

I discovered three plausible ways to do this. First and most simply, dump the
tunes onto a USB drive; wherever you are in the world, you can usually plug one
in and play tunes from it.

Second, there’s Plex; you run a Plex server on one of your computers at home (in
our case a recent Mac Mini) which you point at music and video directories, and
it’ll serve them to clients on the Web or on phones or on platforms like WebOS
and Roku.

Also, it’ll serve your media to anywhere in the world, using UPnP to drill an
outgoing hole through your firewall. Obviously, this could make a
security-sensitive person nervous and does bother me a bit, because UPnP’s
history has featured some nasty vulnerabilities. I have a to-do to check whether
the version on my dumbass telco ISP router is reasonably safe. I believe that
Tailscale would offer a better security posture, but don’t want one more thing
to manage.

Finally, Apple Music can apparently do what YouTube Music does; let you upload
your tunes into the cloud and play them anywhere. But moving from one Big-Tech
provider to another doesn’t feel like progress.

Does it work?

Setting it up on Plex was a Just-Works experience. The process even reached out
through our modern Eero mesh to the old telco router and convinced it to set up
the appropriate UPnP voodoo. If you open the Plex server admin interface it
occasionally complains about a double-NAT situation but works anyhow.

Getting the USB working was kind of hilarious. First of all, I bought a 512G USB
stick. (My Mac says it only has 460GB, but what’s 50G between friends?) USB-A
because that’s what the car has. It took a couple of hours to copy all the music
onto it.

Then I plugged the USB stick into the car and it showed up instantly in the
“Sources” tab of the media player, but greyed out. I snickered when I noticed
that all the car infotainment menus were crawling and stuttering. Asking the
car’s mighty electronic brain to index that mountain of music was making it
sweat. Anyhow, after a few minutes, I could access the USB and now it works
fine, mostly.

By “mostly”, I mean that when I tell it to play music off the USB, it takes a
few seconds for the music to start, then a minute or more to get its shit
together and present a coherent picture of what it’s playing. And on one
occasion, the music player just randomly switched over to the radio. So I
suspect my inventory is pushing the poor little toy computer in the car pretty
hard. But once it’s going, the presentation is nice:

A few items to note here:

“Musick” is the name I gave the USB key.

That recording is Jesus’ Blood Never Failed Me Yet, a truly unique piece of work
by British composer Gavin Bryars. Opinions vary; I think it’s magical but it’s
one of the few pieces of music that I am absolutely forbidden to play anywhere
my wife can hear it.

The car software is way more flexible than Android Auto; this is just one of the
car’s three screens and there are a lot of options for distributing your music
and weather and maps and climate control across them.

Which is better?

It’s complicated. Obviously, the USB option doesn’t require any network
bandwidth. And I think the album-art presentation is nicer than Plex’s. (You can
see that here).

The audio quality is pretty well a wash. Plex is a little louder, I suspect them
of Loudness-War tactics, which is probably OK in a car with its inevitable
background noise. Plex also crossfades the song transitions, clever and pleasing
but really not essential.

Plex is really nice software and I feel a little guilty that I’m not sending
them any money. They do have a “Pro” level of service; must check it out.

Then of course Plex needs Android Auto. Which on the one hand I’m probably going
to be running a lot if I’m driving around town to appointments. But… Android
Auto is already a little shaky some days, not sure whether it’s crashing or the
car software is creaking or it’s just yet another lousy USB-C connection (I am
developing a real hate for that form factor).

Realistically, given that our car (a Jaguar I-Pace EV) wasn’t a big seller and
is five years old, can I really count on Google and Jaguar to do what it takes
to keep Android Auto running?

At this point I need to say a big “Thanks!” to everyone on Fedi/Mastodon who
gave me good advice on how to approach this problem.

Anyhow, as of now, we have two alternatives that work well. The De-Googling
march continues forward.


MORE
March 110
The De-Google Project

My family, like most, depends on a lot of online services. And again like most,
a lot of those services come from Big Tech giants in general and (in our case)
Google in particular. And like many people, we are becoming less comfortable
with that. So I’m going to try to be systematic about addressing the problem.
This post summarizes our dependencies and then I’ll post blog pieces about
updates as I work my way through the list. (The first is already posted, see
below.)

I’m calling this the “De-Google” project because they’re our chief supplier of
this stuff and it’s more euphonious than “De-BigTechInGeneral”.

NeedSupplierAlternatives Office Google Workspace ? Data sharing Dropbox ? Video
meetings Google Meet Jitsi, ? Maps Google Maps Magic Earth, Here, something
OSM-based Browser Apple Safari Firefox, ? Search Google Bing-based options Chat
Signal Photo editing Adobe Lightroom & Nik Capture One, Darktable, ? In-car
interface Google Android Auto Automaker software Play my music Plex, USB
Discover music Google YouTube Music Qobuz, Tidal, Deezer, Pandora, ? TV Prime,
Roku, Apple, Netflix, TSN, Sportsnet ?

The “Supplier” color suggests my feelings about what I’m using, with blue
standing for neutral.

Criteria

To replace the things that I’m unhappy with, I’m looking for some combination
of:

Open source

Not ad-supported

Not VC-funded

Not Google, Apple, Microsoft, or Amazon

Office

We’ve been using Gmail for a really long time and are used to it, and the
integration between mail and calendar and maps basically Just Works. The price
is OK but it keeps going up, and so do our data storage requirements, what with
all the cameras in the family. Finally, Google has stewardship of our lives and
are probably monetizing every keystroke. We’re getting a bit creeped out over
that.

I think that calendars and email are kind of joined at the hip, so we’d want a
provider that does both.

As for online docs, I will not be sorry to shake the dust of Google Drive and
Docs from my heels, I find them clumsy and am always having trouble finding
something that I know is in there.

Data sharing

Dropbox is OK, assuming you ignore all the other stuff it’s trying to sell you.
Maybe one of these years I should look at that other stuff and see if it’s a
candidate to replace one or two other services?

Video meetings

I dislike lots of things about Zoom and find Microsoft Teams a pool of pain, but
have been pretty happy with Google Meet. Nobody has to download or log into
anything and it seems to more or less Just Work. But I’d look at alternatives.

Maps

As I wrote in 2017, Google maps aggregate directions, reviews, descriptions,
phone numbers, and office hours. They are potentially a nuclear-powered monopoly
engine. I use Maps more and more; if I want to contact or interact with
something whose location I know, it’s way quicker to pull up Maps and click on
their listing than it is to use Google search and fight through all the ads and
spam.

The calendar integration is fabulous. If you have Android Auto and you’re going
to a meeting, pull up the calendar app and tap on the meeting and it drops you
right into directions.

The quality of the OpenStreetMap data is very good, but obviously they don’t
have the Directions functions. Who does? Obviously, Here does, and I was
enthused about it in 2019; but Android Auto’s music powers drew me back to
Google Maps. Aside from that, Magic Earth is trying, and their business model
seems acceptable, but the product was pretty rough-edged last time I tried it.

Browser

Safari is my daily driver. These days Chrome is starting to creep me out a bit;
just doesn’t feel like it’s on my side. Also, it’s no longer faster than the
competition. I’d like to shift over to Firefox one day when I have the energy

Then there are the Arcs and Braves and Vivaldis of this world, but I just
haven’t yet invested the time to figure out if one of these will do, and I do
not detect a wave of consensus out there.

By the way, DuckDuckGo has a browser, a shell over Safari on the Mac and Edge on
Windows. Lauren uses it a lot. Probably worth a closer look.

Search

The decline of Google Search is increasingly in everyone’s face. Once again, it
refuses to find things on this blog that I know are there.

Others in the family have already migrated to DuckDuckGo, and I now feel like an
old-school lagger for still not having migrated off Google. I wish there were
someone else taking a serious run at indexing the Web other than Bing — from yet
another tech giant — but here we are.

Lauren tells me to have a closer look at Ecosia, which seems very wholesome.

Chat

At the moment you will have to pry Signal out of my cold, dead, hands. You
should be using it too. ’Nuff said.

Photo editing

I pay my monthly tribute to Adobe, about whom my feelings aren’t as negative as
they are about the mega Tech Giants. I’d like not to pay so much, and I’d like
something that runs a little faster than Lightroom, and I’d like to support open
source. But… I really like Lightroom, and sometimes one absolutely needs
Photoshop, so I’m unlikely to prioritize this particular escape attempt.

In-car interface

Choices are limited. I see little point in migrating between Android Auto and
CarPlay, which leaves the software the auto maker installed. Which, in my
five-year-old Jaguar is… well, not bad actually. I think I could live with the
built-in maps and directions from Here, even with the British Received
Pronunciation’s butchery of North American place names.

But, I don’t know, we might stay with Android Auto. Check out this screenshot
from my car.

(Pardon the blurs and distortions.)

This is Android Auto displaying, as it normally does when I’m driving, maps and
music. By default, Google Maps and YouTube Music. But not here; on the right is
Plex, playing my own music stored on a Mac Mini at home.

On the left, it’s even more interesting: This is neither Google maps nor a
competitor; it’s Gaia GPS, the app I normally use to mark trail while
bushwhacking through Pacific Northwest rain forests. Somehow I fat-fingered it
into place either in the car or on my phone.

The lesson here is that (for the moment at least) Android Auto seems to be
genuinely neutral. It knows the general concepts of “apps that play music” and
“apps that are maps” and is happy to display whichever ones you want, not just
Google’s. (As a former Android geek who knows about Intents and Filters, I can
see how this works. Clever.)

So far, Android Auto doesn’t show ads, but I suppose it’s monetizing me by
harvesting traffic information to enrich its maps and I guess that’s a bargain I
can live with. I use that data myself when I want to go somewhere and there are
multiple routes and I can see which one is backed up by sewer work or whatever.

Discover music

I’ve been paying for YouTube Music since before it existed, and I’m genuinely
impressed with the way its algorithm fishes up new artists that it turns out I
really like. But just now Google laid off a bunch of YouTube Music “contractors”
(de facto, employees) who tried to organize a union, so screw ’em.

I haven’t investigated any of the alternatives in depth yet.

Play my music

In the decades where Compact Disks were the way to acquire music, I acquired a
lot. And ripped it. And pushed it up into Google’s musical cloud. And (until
recently) could shuffle my musical life on YouTube Music. But they removed that
feature from Android Auto, so screw ’em.

But I now have two good ways to do this. Check this out in Play My Music.

TV

The same gripe as everyone else: The streaming services have re-invented Cable
TV, which I only got around to dumping a couple of years ago. The right solution
is obvious: Pay-per-view at a reasonably low price, then the services could
compete on producing great shows that people will pay to see, rather than
sucking you into yet another subscription.

I suspect this column will stay red for quite a while. It’s amazing how much
business leaders hate simple business models where there’s a clean clear
one-time price for a product and customers have a clean clear choice who they
buy their products from.

The path forward

I don’t know if I’ll ever turn the center column all-green. And I don’t need to;
progress is progress. Anyhow, doing this sort of investigation is kind of fun.


MORE
March 110
22.6.2024, 19:00Platformonomics (Feed)
Platformonomics TGIF #43: March 8, 2024

Platformonomics TGIF is a weekly roll-up of links, comments on those links, and
perhaps a little too much tugging on my favorite threads.

Get PLATFORMONOMICS Updates By Email

Subscribe

Short edition this week, despite the insanity. But Clown Week cometh! (Think
Shark Week for CAPEX obsessives).

News Behold the Nuclear Cloud

Previous:

Cloud Power Up, Nuclear Powered Cloud Data Centers Are Happening, A
Nuclear-Powered Cloud, When Nuclear-Powered Data Centers?

Antitrust Incoherence: Spotify Edition



European Antitrust prioritizes protecting European firms. And they don’t have
many tech firms to protect…

Previous:

Antitrust Incoherence: Roomba Edition, A New Antitrust Doctrine

Useful Idiots, Fellow Travelers and Unregistered Foreign Agents: The Donald

This is going to make so many heads explode. Do tribal affiliations flip
overnight after Dear Leader weights in, reversing his prior stance?

Previous:

Useful Idiots, Fellow Travelers and Unregistered Foreign Agents: DC Lobbyists
Part III, Hollywood, DC Lobbyists Part II, NBA, Warner Bros. Discovery, Germany,
DC Lobbyists, Elon/Twitter, The Hugo Awards, Harvard University, Wall Street,
Apple

Useful Idiots, Fellow Travelers and Unregistered Foreign Agents: Jeff Yass



Nothing is more important than investor IRR.

Useful Idiots, Fellow Travelers and Unregistered Foreign Agents: Microsoft Bing

So much kowtowing. To what end?

Previous:

Best Nickname for Emmanuel Macron, Winnie-the-Pooh: Lover of Honey, Champion of
Freedom, Press Releases We’d Like to See


MORE
April 230
Today, 14:28Phil Windley's Technometria (Feed)
Identity Metasystems and Lessons from Building the Sovrin Foundation

I recently spoke with Riley Hughes of Trinsic on his Future of Identity podcast
about the birth of Sovrin Foundation, its inevitable growing pains,
self-sovereign identity, identity metasystems, and adoption. Give it a listen.

I'm grateful to Riley for having me on as a guest.

Tags: identity ssi podcasts


MORE
March 50
24.6.2024, 6:00SEI Blog (Feed)
10 Lessons in Security Operations and Incident Management
This post outlines 10 lessons learned from more than three decades of building
incident response and security teams throughout the globe.

MORE
March 50
22.6.2024, 19:00Platformonomics (Feed)
Platformonomics TGIF #42: March 1, 2024

Platformonomics TGIF is a weekly roll-up of links, comments on those links, and
perhaps a little too much tugging on my favorite threads.

Get PLATFORMONOMICS Updates By Email

Subscribe

News A CAPEX Obsessive’s Work is Never Done

Previous:

Follow the CAPEX: Cloud Table Stakes 2023 Retrospective, Follow the CAPEX:
Triangulating NVIDIA, more CAPEX

Google: Still Not Code Red?

We (yet) again ask if Google fully appreciates its predicament. They did manage
to get an anodyne and perfunctory email out a few days after the guardrails
self-immolation.

7 observations while reading the Google CEO email about Gemini:

⁃It gets off to a bad start with the use of “problematic” — a word used almost
exclusively by political activists, corporate bureaucracies, and the progressive
left. At a time when Google is being accused of being… https://t.co/xxPTl8hIwx

— Lulu Cheng Meservey (@lulumeservey) February 28, 2024

The search for Sundar’s successor has moved to the alumni network. But here’s my
neologism for the breakup option discussion:

Previous:

Google Goes Off the Guardrails, AI Marketing Hallucinations: Google Edition, The
Google Pivot Away from “AI-Safety First”, “Google has more than 160 people
working full time“, Google: Code Red Now?

Related:

“Corporate surrealism”, “…don’t reduce something by calling it a technical
problem when it’s really a broad-based company problem“

Google’s Cronkite Moment?

When you’ve lost Ben Thompson…

Related:

Cronkite moment

EU Insanity: Mistral Edition

I tried to summarize all the crazy assumptions, inconsistencies and hypocrisy
around EU responses to Mistral partnering with Microsoft, but just couldn’t. The
EU poses an immense cognitive and rhetorical challenge. I have entire posts on
the shelf because it is so hard to write prose that is commensurate with the
sheer level of insanity from the EU.

Previous:

Move Fast and Regulate Things (You Don’t Understand), The EU Will Continue to
Fiddle While Rome Burns, When “Move Fast and Regulate Things” Breaks Down, AI
Regulation: Move Fast and Regulate Things

EU Insanity: Performative Regulation has Consequences New Cloud Repatriation
Hypothesis

What if cloud repatriation really is happening, but is serverless?

Previous:

Cloud Repatriation Where Art Thou?, Cloud Repatriation: The Search Continues,
Cloud Repatriation: Still Not a Thing, Platformonomics Repatriation Index – Q1
2023: Surf’s Up?,

Private Equity in Action: VMware Yet Again

You mark up my portfolio, I’ll mark up yours…

Previous:

Perfidious Private Equity, Private Equity in Action: VMware (Continued), Private
Equity in Action: VMware, Private Equity Will Buy and Ruin More Software
Companies, Software Migration Alert: VMware (Continued), Software Migration
Alert: VMware

Related:



Everything is Super

Another day, another definition from the “supercloud” svengalis. At this precise
moment “supercloud” is currently defined as an “editorial event“, while throwing
“superchips” and “superapps” onto the pyre as they finally start to discover
that AI is the actual thing.

Previous:

“Supercloud”: Still Dead, “Supercloud” RIP

A Call for the Corporate Death Penalty

If Arthur Anderson got the corporate death penalty for its role in the Enron
fraud, the death penalty must be in consideration for serial malignancy McKinsey
(opioids crisis, insider trading, also Enron, other autocracies, hybrid cloud,
et al).

Previous:

McKinsey Karma?

Related:

McKinsey: Last Week Tonight with John Oliver

Useful Idiots, Fellow Travelers and Unregistered Foreign Agents: DC Lobbyists
Part III

Shaming Westerners who brazenly advance West Taiwan’s aims is working. So the
shaming will continue.

Previous:

Useful Idiots, Fellow Travelers and Unregistered Foreign Agents: DC Lobbyists
Part II, NBA, Warner Bros. Discovery, Germany, DC Lobbyists, Elon/Twitter, The
Hugo Awards, Harvard University, Wall Street, Apple

Useful Idiots, Fellow Travelers and Unregistered Foreign Agents: Hollywood

Previous:

Useful Idiots, Fellow Travelers and Unregistered Foreign Agents: DC Lobbyists
Part II, NBA, Warner Bros. Discovery, Germany, DC Lobbyists, Elon/Twitter, The
Hugo Awards, Harvard University, Wall Street, Apple

Related:

Band in China (South Park)


MORE
April 230
24.6.2024, 6:00SEI Blog (Feed)
CERT Releases 2 Tools to Assess Insider Risk
The average insider risk incident costs organizations more than $600,000. To
help organizations assess their insider risk programs, the SEI CERT Division has
released two tools available for download.

MORE
February 280
Today, 17:55ongoing by Tim Bray (Feed)
Money Bubble

I think I’m probably going to lose quite a lot of money in the next year or two.
It’s partly AI’s fault, but not mostly. Nonetheless I’m mostly going to write
about AI, because it intersects the technosphere, where I’ve lived for decades.

I’ve given up having a regular job. The family still has income but mostly we’re
harvesting our savings, built up over decades in a well-paid profession. Which
means that we are, willy-nilly, investors. And thus aware of the fever-dream
finance landscape that is InvestorWorld.

The Larger Bubble

Put in the simplest way: Things have been too good for too long in
InvestorWorld: low interest, high profits, the unending rocket rise of the
Big-Tech sector, now with AI afterburners. Wile E. Coyote hasn’t actually run
off the edge of the cliff yet, but there are just way more ways for things to go
wrong than right in the immediate future.

If you want to dive a little deeper, The Economist has a sharp (but paywalled)
take in Stockmarkets are booming. But the good times are unlikely to last. Their
argument is that profits are overvalued by investors because, in recent years,
they’ve always gone up. Mr Market ignores the fact that at least some of those
gleaming profits are artifacts of tax-slashing by right-wing governments.

That piece considers the observation that “Many investors hope that AI will ride
to the rescue” and is politely skeptical.

Popping the bubble

My own feelings aren’t polite; closer to Yep, you are living in a Nvidia-led
tech bubble by Brian Sozzi over at Yahoo! Finance.

Sozzi is fair, pointing out that this bubble feels different from the cannabis
and crypto crazes; among other things, chipmakers and cloud providers are
reporting big high-margin revenues for real actual products. But he hammers the
central point: What we’re seeing is FOMO-driven dumb money thrown at technology
by people who have no hope of understanding it. Just because everybody else is
and because the GPTs and image generators have cool demos. Sozzi has the
numbers, looking at valuations through standard old-as-dirt filters and shaking
his head at what he sees.

What’s going to happen, I’m pretty sure, is that AI/ML will, inevitably,
disappoint; in the financial sense I mean, probably doing some useful things,
maybe even a lot, but not generating the kind of profit explosions that you’d
need to justify the bubble. So it’ll pop, and my bet it is takes a bunch of the
finance world with it. As bad as 2008? Nobody knows, but it wouldn’t surprise
me.

The rest of this piece considers the issues facing AI/ML,  with the goal of
showing why I see it as a bubble-inflator and eventual bubble-popper.

First, a disclosure: I speak as an educated amateur. I’ve never gone much below
the surface of the technology, never constructed a model or built
model-processing software, or looked closely at the math. But I think the
discussion below still works.

What’s good about AI/ML

Spoiler: I’m not the kind of burn-it-with-fire skeptic that I became around
anything blockchain-flavored. It is clear that generative models manage to embed
significant parts of the structure of language, of code, of pictures, of many
things where that has previously not been the case. The understanding is
sufficient to reliably accomplish the objective: Produce plausible output.

I’ve read enough Chomsky to believe that facility with language is a defining
characteristic of intelligence. More than that, a necessary but not sufficient
ingredient. I dunno if anyone will build an AGI in my lifetime, but I am
confident that the task would remain beyond reach without the functions offered
by today’s generative models.

Furthermore, I’m super impressed by something nobody else seems to talk about:
Prompt parsing. Obviously, prompts are processed into a representation that
reliably sends the model-traversal logic down substantially the right paths. The
LLMbots of this world may regularly be crazy and/or just wrong, but they do
consistently if not correctly address the substance of the prompt. There is
seriously good natural-language engineering going on here that AI’s critics
aren’t paying enough attention to.

So I have no patience with those who scoff at today’s technology, accusing it
being a glorified Markov chain. Like the song says: Something’s happening here!
(What it is ain’t exactly clear.)

It helps that in the late teens I saw neural-net pattern-matching at work on
real-world problems from close up and developed serious respect for what that
technology can do; An example is EC2’s Predictive Auto Scaling (and gosh, it
looks like the competition has it too).

And recently, Adobe Lightroom has shipped a pretty awesome “Select Sky” feature.
It makes my M2 MacBook Pro think hard for a second or two, but I rarely see it
miss even an isolated scrap of sky off in the corner of the frame. It allows me,
in a picture like this, to make the sky’s brightness echo the water’s.

And of course I’ve heard about success stories in radiology and other
disciplines.

Thus, please don’t call me an “AI skeptic” or some such. There is a there there.

But…

Given that, why do I still think that the flood of money being thrown at this
tech is dumb, and that most of it will be lost? Partly just because of that
flood. When financial decision makers throw loads of money at things they don’t
understand, lots of it is always lost.

In the Venture-Capital business, that’s an understood part of the business
cycle; they’re looking to balance that out with a small number of 100x startup
wins. But when big old insurance companies and airlines and so on are piling in
and releasing effusive statements about building the company around some new
tech voodoo, the outcome, in my experience, is very rarely good.

But let’s be specific.

Meaning

As I said above, I think the human mind has a large and important
language-processing system. But that’s not all. It’s also a (slow,
poorly-understood) computer, with access to a medium-large database of facts and
recollections, an ultra-slow numeric processor, and facilities for estimation,
prediction, speculation, and invention. Let’s group all this stuff together and
call it “meaning”.

Have a look at Climbing towards NLU: On Meaning, Form, and Understanding in the
Age of Data by Emily Bender and Alexander Koller (2020). I don’t agree with all
of it, and it addresses an earlier generation of generative models, but it’s
very thought-provoking. It postulates the “Octopus Test”, a good variation on
the bad old Chinese-Room analogy. It talks usefully about how human language
acquisition works. A couple of quotes: “It is instructive to look at the past to
appreciate this question. Computational linguistics has gone through many
fashion cycles over the course of its history” and “In this paper, we have
argued that in contrast to some current hype, meaning cannot be learned from
form alone.”

I’m not saying these problems can’t be solved. Software systems can be equipped
with databases of facts, and who knows, perhaps some day estimation, prediction,
speculation, and invention. But it’s not going to be easy.

Difficulty

I think there’s a useful analogy between the narratives around AI and of
self-driving cars. As I write this, Apple has apparently decided that generative
AI is easier than shipping an autonomous car. I’m particularly sensitive to this
analogy because back around 2010, as the first self-driving prototypes were
coming into view, I predicted, loudly and in public, that this technology was
about to become ubiquitous and turn the economy inside out. Ouch.

There’s a pattern: The technologies that really do change the world tend to have
strings of successes, producing obvious benefits even in their earliest forms,
to the extent that geeks load them in the back doors of organizations just to
get shit done. As they say, “The CIO is the last to know.”

Contrast cryptocurrencies and blockchains, which limped along from year to year,
always promising a brilliant future, never doing anything useful. As to the
usefulness of self-driving technology, I still think it’s gonna get there, but
it’s surrounded by a cloud of litigation.

Anyhow, anybody who thinks that it’ll be easy to teach “meaning” (as I described
it above) to today’s generative AI is a fool, and you shouldn’t give them your
money.

Money and carbon

Another big problem we’re not talking about enough is the cost of generative AI.
Nature offers Generative AI’s environmental costs are soaring — and mostly
secret. In a Mastodon thread, @Quixoticgeek@social.v.st says We need to talk
about data centres, and includes a few hard and sobering numbers.

Short form: This shit is expensive, in dollars and in carbon load. Nvidia pulled
in $60.9 billion in 2023, up 126% from the previous year, and is heading for a
$100B/year run rate, while reporting a 75% margin.

Another thing these articles don’t mention is that building, deploying, and
running generative-AI systems requires significant effort from a small group of
people who now apparently constitute the world’s highest-paid cadre of
engineers. And good luck trying to hire one if you’re a mainstream company where
IT is a cost center.

All this means that for the technology to succeed, it not only has to do
something useful, but people and businesses will have to be ready to pay a
significantly high price for that something.

I’m not saying that there’s nothing that qualifies, but I am betting that it’s
not in ad-supported territory.

Also, it’s going to have to deal with pushback from unreasonable climate-change
resisters like, for example, me.

Anyhow…

I kind of flipped out, and was motivated to finish this blog piece, when I saw
this: “UK government wants to use AI to cut civil service jobs: Yes, you read
that right.” The idea — to have citizen input processed and responded to by an
LLM — is hideously toxic and broken; and usefully reveals the kind of thinking
that makes morally crippled leaders all across our system love this technology.

The road ahead looks bumpy from where I sit. And when the business community
wakes up and realizes that replacing people with shitty technology doesn’t show
up as a positive on the financials after you factor in the consequences of
customer rage, that’s when the hot air gushes out of the bubble.

It might not take big chunks of InvestorWorld with it. But I’m betting it does.


MORE
April 190
22.6.2024, 19:00Platformonomics (Feed)
Platformonomics TGIF #41: February 23, 2024

Platformonomics TGIF is a weekly roll-up of links, comments on those links, and
perhaps a little too much tugging on my favorite threads.

Get PLATFORMONOMICS Updates By Email

Subscribe

News Follow the CAPEX: Triangulating NVIDIA

I wrote a post about what we can learn from the intersection of NVIDIA’s revenue
and hypercloud CAPEX spending.

Previous:

Follow the CAPEX: Triangulating NVIDIA, more CAPEX

Related:

Jim Chanos on NVIDIA sales/hypercloud CAPEX

Google Goes Off the Guardrails

Generative AI is augmented reality, created by holding a mirror up to humanity,
and then “guardrailing” that reality into something altogether different (and
idealized).

Google has made Elon’s case in a way that Elon himself never could (and is still
unlikely to execute on successfully).

And yet another sign Google isn’t anywhere near a wartime footing. Pay-per-view
of this debacle’s postmortem would be huge.

The Tiananmen Square filter is particularly bizarre given Google has less
exposure to West Taiwan than most Big Tech companies.

Previous:

AI Marketing Hallucinations: Google Edition, The Google Pivot Away from
“AI-Safety First”, “Google has more than 160 people working full time“, Google:
Code Red Now?

Does the New York Times Editorial Page Ever Talk to the Business Side of the
House?

New York Times slogan: “All the Hypocrisy About Tech That’s Fit to Print!”

Previous:

Living in a Glass House: The New York Times On Tech

Useful Idiots, Fellow Travelers and Unregistered Foreign Agents: DC Lobbyists
Part II

Shaming Westerners who brazenly advance West Taiwan’s aims may work. So the
shaming will continue.

Previous:

Useful Idiots, Fellow Travelers and Unregistered Foreign Agents: DC Lobbyists,
Elon/Twitter, The Hugo Awards, Harvard University, Wall Street, Apple

Useful Idiots, Fellow Travelers and Unregistered Foreign Agents: NBA

Previous:

Press Releases We’d Like to See: Daryl Morey, Trey Parker and Matt Stone Awarded
the 2020 Nobel Peace Prize

Useful Idiots, Fellow Travelers and Unregistered Foreign Agents: Warner Bros.
Discovery

A sanitized Xinjiang travelogue no less.

Useful Idiots, Fellow Travelers and Unregistered Foreign Agents: Germany

Previous:

Collapsing Economic Models: German Edition, EU Rethinking Its Stance on
Mercantilism, German Auto Industry Reaches Hail Mary Stage, How Do You Say
“Burning Platform” in German?, Volkswagen’s Death Throes Continue, The Fate of
the European Economy: Automotive Edition, Europe: Investing in Slave Labor


MORE
April 230
Follow the CAPEX: Triangulating NVIDIA

Previous CAPEX obsessing

Lets compare NVIDIA’s Data Center segment revenue to the CAPEX spending of the
hyperclouds (Amazon, Google, Microsoft, henceforth AGM) plus Meta over the last
couple years (AGMM collectively). We’d like to know much of NVIDIA’s revenue is
from the hyperclouds (and honorary hypercloud sidekick Meta) and how much of
hypercloud CAPEX is going to NVIDIA for AI infrastructure.

Above we see the AI liftoff in NVIDIA’s Data Center business in the last year.
ChatGPT launched five quarters ago (November 30, 2022) while the NVIDIA H100
shipped “in the fall” of 2022.

NVIDIA’s fiscal year ends in January, but below we will compare their quarters
with the overlapping calendar quarters of their big customers, so remember the
NVIDIA revenues go a month later than the hyperclouds.

The chart above compares all-up, corporate level CAPEX spend of AGM (so includes
Amazon’s #bonkers pandemic spending on logistics infrastructure). NVIDIA’s total
Data Center revenues hit 49% of AGM all-up CAPEX in Q4.

Now we compare AGM’s estimated spending on data center CAPEX (based on
proprietary Platformonomics analysis aka, in proper analyst tradition, a guess).
NVIDIA Data Center revenue is at 81% of that estimated AGM data center spending.
It is hard to imagine NVIDIA getting half of hypercloud CAPEX, so that suggests
NVIDIA is spreading out the GPUs across lots of customers.

Next we add Meta to the all-up corporate spend picture. I don’t usually track
Meta as they don’t have a cloud (and likely won’t due to a recurring tendency to
pull the rug out from under developers, which is a bad look for aspiring
platforms), but they do spend a lot on CAPEX1 and more recently have gone all-in
on GPUs (though it is really unclear what they are doing with them beyond
training Llama — perhaps LLM-powered search?). That takes NVIDIA Data Center
revenue down to 40% of AGMM CAPEX in Q4.

Finally, we’ll add our proprietary guess of data center infrastructure spending
by AGM plus Meta. NVIDIA Data Center revenue is about 65%. Again, that suggests
they’re spreading the GPUs well beyond the hyperclouds and Meta.

None of the above tells us anything about what the hyperclouds and Meta are
spending, collectively or individually, on AI infrastructure. NVIDIA obviously
has a lot of other customers (enterprises, universities, nation states, et al.)
and are over-allocating scarce supply to boutique GPU clouds2 (e.g. CoreWeave,
Llama Labs, Oracle) in hopes of building leverage over the hyperclouds (who
happen to be doing their own competitive silicon).

NVIDIA did say in their conference call:

“In the fourth quarter, large cloud providers represented more than half of our
data center revenue, supporting both internal workloads and external public
cloud customers.”

So over $9 billion of their revenue went to the hyperclouds (and they may
generously include other not-so-hyper providers in that bucket, i.e. what we
call “clowns“).

Microsoft and NVIDIA

Much more interesting is what NVIDIA says about customer concentration in their
SEC filings. Our assumption is that Microsoft is NVIDIA’s largest customer. From
these disclosures, we can construct the following:

Microsoft was likely responsible for 19% of NVIDIA’s revenue in FY24 and 22% in
the fourth quarter. This suggests Microsoft spent $11.58 billion with NVIDIA in
NVIDIA’s FY24. Microsoft also buys (a few) GPUs for Surface devices so it isn’t
entirely AI infrastructure, but it is close.

And if Microsoft spent almost $5 billion with NVIDIA in Q4, that leaves ~$4
billion to split between Amazon, Google, and perhaps other “large cloud
providers”. It is unclear if Meta is in NVIDIA’s “large cloud provider”
category, but if so, it would suggest Amazon and Google are getting very small
allocations.

A second customer gets called out in Q3 as accounting for 13% of revenue in Q3
and 10% of revenue for the first three quarters of the year ($2.4B and $3.9B
respectively). This is probably Meta (too early for Amazon who took a long time
to make peace with NVIDIA and Google is dividing their investment between NVIDIA
and their own TPUs). Customer two didn’t warrant a breakout in any other
quarter.

If we switch to Microsoft’s calendar (so the NVIDIA numbers are off by a month),
we can look at spend with NVIDIA vs. Microsoft’s total CAPEX spend. It peaks to
43% by Q4! The mad scramble for GPUs is accelerating!

$11.8 billions tops my prediction of how much Microsoft spent on generative AI
infrastructure from just a couple weeks ago:

Microsoft is reputed to be the largest customer for both NVIDIA (Q4 2022, Q2
2023, 2H 2023) and AMD (plus is doing its own AI silicon). The abrupt 4% bump in
CAPEX as a percentage of revenue, after a steady 13-14% for years prior, is the
best proxy for the incremental AI spending in 2023. That suggests an incremental
AI-driven spend of about $9 billion, or 22% of overall CAPEX.

That means CAPEX intensity for the rest of Microsoft’s (non-AI) cloud
infrastructure is actually declining, even as overall Azure cloud revenue grew
by 30% (to which they attributed 6 points to AI). Either they’re stretching the
infrastructure/not investing as far ahead of demand or there really is something
to the extensions of server lifetimes (which I still dismiss as accounting
hijinks downstream from raw CAPEX investments).

Microsoft’s $11.58 billion spend translates at list prices to about 385,000
H100s (on top of whatever they bought in 2022). Presumably you also get a unit
discount when you write a three comma check. There is also some (required)
networking gear in there too3, but that just muddles a nice big number.

NVIDIA CAPEX

NVIDIA’s own CAPEX is measly, just $1.07 billion for the fiscal year, and down
42% from the prior year. Less than 2% of revenue. Fabless indeed.

What have I missed here?

Seventh largest in the CAPEX Extended Universe I track, just behind TSMC. ︎Who
may be deadbeats? ︎In the conference call, NVIDIA said a couple things about
networking which is also in the Data Center segment and complements the GPUs in
AI infrastructure: “Compute revenue grew more than 5x and networking revenue
tripled from last year.” and “Networking exceeded a $13 billion annualized
revenue run rate. Our end-to-end networking solutions define modern AI data
centers. Our InfiniBand solutions grew more than 5x year on year.” [$13 billion
run rate equates to $3.25 billion in Q4, which is about 18% of NVIDIA’s Data
Center business]. ︎

MORE
February 230
24.6.2024, 6:00SEI Blog (Feed)
OpenAI Collaboration Yields 14 Recommendations for Evaluating LLMs for
Cybersecurity
This SEI Blog post summarizes 14 recommendations to help assessors accurately
evaluate LLM cybersecurity capabilities.

MORE
February 210
22.6.2024, 19:00Platformonomics (Feed)
Platformonomics TGIF #40: February 16, 2024

Platformonomics TGIF is a weekly roll-up of links, comments on those links, and
perhaps a little too much tugging on my favorite threads.

Get Updates By Email

Subscribe

News CAPEX Everywhere



The Hamas data center under UN headquarters in Gaza had eight racks. In Oracle
parlance, that is four regions.

Previous:

Follow the CAPEX: Keeping Up With The Clowns

Cloud Repatriation Where Art Thou?

With Q4 2023 numbers in, the Platformonomics Repatriation Index continues its
decline (i.e. cloud prevails). This index’s construction is extremely favorable
for cloud repatriation, yet we still don’t see it. Digital Realty Trust and
Equinix should be huge beneficiaries from any cloud repatriation. Even when we
include their “other revenue” and only compare them against AWS, the slowest
growing cloud, we still don’t see any signs of repatriation. The data center
companies’ 2024 guidance is for single digit growth, so they don’t seem to see
it either.

They’re not getting any AI tailwind either, which is a topic for another week,
but here’s a glimpse of the future:

Data centers are not the salvation for the beleaguered commercial real estate
market, even if that is the superficial conventional wisdom amongst investors.

Previous:

Cloud Repatriation: The Search Continues, Cloud Repatriation: Still Not a Thing,
Platformonomics Repatriation Index – Q1 2023: Surf’s Up?, The Inexorable
Attraction Between Terrible Things: Cloud Repatriation and Private Equity,
Private Equity in Action: Data Centers

SEC Embraces Platformonomics ClownWatch, Warns Amazon and IBM

Previous:

Introducing Platformonomics ClownWatch, IBM and the Art of Misleading Investors,
This Week in Securities Fraud, Clown CAPEX, IBM’s “Cloud” Business (or Lack
Thereof), IBM: How Much Longer Do the Good Times Last?, The AWS Generative AI
Soap Opera, The Amazon AI Inferiority Complex

Media Wants More Sources Revealed

The never-ending hypocrisy in media coverage of technology never fails to
entertain.

Previous:

Living in a Glass House: The New York Times On Tech

Useful Idiots, Fellow Travelers and Unregistered Foreign Agents: DC Lobbyists

They’re registered foreign agents, which is even worse. Aid and Comfort.

Previous:

Useful Idiots, Fellow Travelers and Unregistered Foreign Agents: Harvard
University, Useful Idiots, Fellow Travelers and Unregistered Foreign Agents:
Wall Street, Useful Idiots, Fellow Travelers and Unregistered Foreign Agents:
Apple

Related:

Lawmakers weigh blacklist for firms lobbying for Chinese military-linked
companies

Useful Idiots, Fellow Travelers and Unregistered Foreign Agents: Elon/Twitter

Elon is a very compromised geopolitical actor.

Previous:

Useful Idiots, Fellow Travelers and Unregistered Foreign Agents: Harvard
University, Useful Idiots, Fellow Travelers and Unregistered Foreign Agents:
Wall Street, Useful Idiots, Fellow Travelers and Unregistered Foreign Agents:
Apple

Useful Idiots, Fellow Travelers and Unregistered Foreign Agents: The Hugo Awards



Western self-censorship on behalf of West Taiwan is such a bad look. Yet
pervasive.

Previous:

Useful Idiots, Fellow Travelers and Unregistered Foreign Agents: Harvard
University, Useful Idiots, Fellow Travelers and Unregistered Foreign Agents:
Wall Street, Useful Idiots, Fellow Travelers and Unregistered Foreign Agents:
Apple


MORE
March 60
Today, 17:55ongoing by Tim Bray (Feed)
Social Photos

I like taking pictures, and I like sharing pictures wherever I hang out online.
A problem with this is knowing that the pictures will very rarely look as good
in other people’s browsers and apps as they do to me in Lightroom on a big
bright 4K screen. Thus this piece, a basic investigation of how photos are
processed and transformed on Mastodon, Bluesky, and Threads.

I was never that much of an Instagram poster; Insta does a good job of taking
your crappy phone pix and juicing them up with filters so they look way better.
That’s irrelevant to me, because not only do I like taking pictures, I like
polishing them with Lightroom and Silver Efex and so on. So with a few
exceptions, everything I want to share gets pulled onto my Mac and edited before
I share it. And once I’ve done that, why would I post pictures anywhere but
where I have my normal conversations?

The picture

Here it is:

Taken with a Pixel 7 out an airplane window somewhere just west of Havre,
Montana. It seems like there are two layers of clouds at the left of the picture
but if you look closely, the lower one I think is the Rocky Mountains in the
distance.

That’s a big picture, both in its subject and raw size: The Pixel version, after
editing, is 3814x2290. Also it has a lot of fine detail, and rewards zooming in.
When I post it, I’d like some sense of the bigness to come across, and when
tapped to enlarge, I’d like it to wow people a little, especially those
fortunate enough to be looking at big screens. And I’d like it to be at least OK
on your phone.

Normally, pictures here in the blog are limited to max 720x720 in the column of
text, and the larger version you get by clicking to 1440x960. But in this case,
if you click you get a 2558x1536 version, the objective being that that’ll be
big enough to fill almost any screen it gets viewed on.

Methodology and apology

The question I want to investigate is, “which platforms are going to make my
pictures look good?” But I haven’t really figured out yet how to do that. To
start with, what kind of picture is going to do best as a metric to judge the
processing quality?

Anyhow, I picked this one and posted it to Mastodon, Bluesky, and Threads, and
here gather data about the results. But hey, why not tap those links on whatever
device you’re using right now and see what you think about how the picture looks
there?

The columns are:

Bytes: the size of the photo as downloaded.

WxH: width and height, in pixels.

“Q”: the JPG quality, as reported by Imagemagick’s identify --verbose. The
quotes are there because I’m not sure how to interpret, or even whether it’s any
use at all.

BytesWxH“Q” Original16715142558 153694 Blog form6249611440  86594 BlueskyAndroid
FS302972864  663 BlueskyAndroid mini42410345  345 BlueskyWeb FS5363452000 120180
BlueskyWeb mini1123351000  60180 MastodonWeb FS15551112558 153690 MastodonWeb
mini86374619  37290 PhanpyWeb FS15551112558 153690 PhanpyWeb mini86374619  37290
ThreadsWeb FS8880672160 129790 ThreadsWeb mini8880672160 129790

Note that each of the posts included not one but two pictures, because I was
also interested in how the platforms allocated screen space. The platforms
typically have two display modes, “mini”, as shown in the feed, and “FS” for
Full Size, what you get when you click on the picture.

Original/Blog form

I think that ideally, I’d like each platform’s presentation of the picture, when
you click on it, to have the same number of pixels and for each pixel to have
the same color value, as in my original.

Bluesky

First of all are numbers from the Android app, but please don’t take them
seriously. The process of extracting them from the Pixel’s screen and getting
them onto my Mac involved multiple irritating steps, each one of which may have
damaged the bits. So I didn’t repeat the exercise for the other platforms. They
are mostly here to encourage me, should I pursue this further, to find a good
clean way to extract this information.

I do note, however, that the “mini” form in the Bluesky Android feed really
crushes those poor little pictures down and, for this particular picture, offers
no suggestion that it’s big.

The Web version of Bluesky does not preserve my pixels, but coerces the size
down to 2K and 1K width in FS and mini versions.

Mastodon and Phanpy

Phanpy is an alternate client for Mastodon; I think it’s very good and it’s my
daily driver. The table reveals that, in this case, the alternate client pulls
in the same images as the official Web client, which is good.

It also reveals that Mastodon preserves the picture’s dimensions, but obviously
reprocesses it somehow, because the photos grow (somewhat) smaller. I wish they
didn’t do that. It’s open-source, I should peek in and see what they actually
do.

Phanpy does a better job of actually showing the pictures in-feed than the
official Mastodon client, and both are nicer than Bluesky.

Threads

I had difficulty, because the Threads Web client is a tangly JavaScript fever
dream, so it’s really hard to get at the underlying photos, but my efforts
suggested that it uses the same picture for the “mini” and “FS” versions, just
getting the browser to scale them down.

Furthermore, Threads doesn’t want pictures to be more than 2160 pixels wide.

Maybe I’m wrong

Because the experimental work was manual and thus highly prone to fumblefingers
and brain farts. If you think that any of these numbers are wrong, you may be
right; please yell at me.

I hesitate to offer a conclusion because this is, as noted at the top, the first
steps in what could be a large and interesting research project, one that I
probably don’t have the expertise to conduct. But, here are a few anyhow.

First, they all do a pretty good job. Second, none of them actually offer an
opportunity to view my bits exactly as uploaded, which I think they should.
Third, client designers should follow Phanpy’s lead in figuring out how to make
better use of screen real-estate to highlight images.

What Lewis Carroll Said

And I quote: “What is the use of a book,” thought Alice, “without pictures or
conversations?”


MORE
February 210
Today, 14:28Phil Windley's Technometria (Feed)
Zero Trust with Zero Data

Summary: The physical world is full of zero trust examples, but they gather
attributes for the access control decisions in a very different way than we're
used to online.

Presenting your ID to buy beer is used so often as an example of how verifiable
credentials work that it's cliche. Cliche or not, there's another aspect of
using an ID to buy beer that I want to focus on: it's an excellent example of
zero trust

Zero Trust operates on a simple, yet powerful principle: "assume breach." In a
world where network boundaries are increasingly porous and cyber threats are
more evasive than ever, the Zero Trust model centers around the notion that no
one, whether internal or external, should be inherently trusted. This approach
mandates continuous verification, strict access controls, and
micro-segmentation, ensuring that every user and device proves their legitimacy
before gaining access to sensitive resources. If we assume breach, then the only
strategy that can protect the corporate network, infrastructure, applications,
and people is to authorize every access. From Zero Trust
Referenced 2024-02-09T08:25:55-0500

The real world is full of zero trust examples. When we're controlling access to
something in the physical world—beer, a movie, a boarding gate, points in a
loyalty program, prescription drugs, and so on—we almost invariably use a zero
trust model. We authorize every access. This isn't surprising, the physical
world is remarkably decentralized and there aren't many natural boundaries to
exploit and artificial boundaries are expensive and inconvenient.

The other thing that's interesting about zero trust in the physical world is
that authorization is also usually done using Zero Data. Zero data is a name
StJohn Deakin gave to the concept of using data gathered just in time to make
authorization and other decisions rather than relying on great stores of data.
There are obvious security benefits from storing less data, but zero data also
offers significantly greater convenience for people and organizations alike. To
top all that off, it can save money by reducing the number of partner
integrations (i.e., far fewer federations) and enable applications that have far
greater scale.

Let's examine these benefits in the scenario I opened with. Imagine that instead
of using a credential (e.g., driver's license) to prove your age when buying
beer, we ran convenience stores like a web app. Before you could shop, you'd
have to register an account. And if you wanted to buy beer, the company would
have to proof the identity of the person to ensure they're over 21. Now when you
buy beer at the store, you'd log in so the system could use your stored
attributes to ensure you were allowed to buy beer.

This scenario is still zero trust, but not zero data. And it's ludicrous to
imagine anyone would put up with it, but we do it everyday online. I don't know
about you, but I'm comforted to know that every convenience store I visit
doesn't have a store of all kinds of information about me in an account
somewhere. Zero data stores less data that can be exploited by hackers (or the
companies we trust with it).

The benefit of scale is obvious as well. In a zero data, zero trust scenario we
don't have to have long-term transactional relationships with every store,
movie, restaurant, and barber shop we visit. They don't have to maintain
federation relationships with numerous identity providers. There are places
where the ability to scale zero trust really matters. For example, it's
impossible for every hospital to have a relationship with every other hospital
for purposes of authorizing access for medical personal who move or need
temporary access. Similarly, airline personal move between numerous airports and
need access to various facilities at airports.

Finally, the integration burden with zero trust with zero data is much lower.
The convenience store selling beer doesn't have to have an integration with any
other system to check your ID. The attributes are self-contained in a
tamper-evident package with built-in biometric authentication. Even more
important, no legal agreement or prior coordination is needed. Lower integration
burden reduces the prerequisites for implementing zero trust.

How do we build zero data, zero trust systems? By using verifiable credentials
to transfer attributes about their subject in a way that is decentralized and
yet trustworthy. Zero data aligns our online existence more closely with our
real-world interactions, fostering new methods of communication while decreasing
the challenges and risks associated with amassing, storing, and utilizing vast
amounts of data.

Just-in-time, zero data, attribute transfer can make many zero trust scenarios
more realizable because it's more flexible. Zero trust with zero data,
facilitated by verifiable credentials, represents a pivotal transition in how
digital identity is used in authorization decisions. By minimizing centralized
data storage and emphasizing cryptographic verifiability, this approach aims to
address the prevalent challenges in data management, security, and user trust.
By allowing online interactions to more faithfully follow established patterns
of transferring trust from the physical world, zero trust with zero data
promotes better security with increased convenience and lower cost. What's not
to like?

Photo Credit: We ID Everyone from DALL-E (Public Domain) DALL-E apparently
thinks a six-pack has 8 bottles but this was the best of several attempts.
Here's the prompt: Produce a photo-realistic image of a convenience store clerk.
She's behind the counter and there's a six pack of beer on the counter. Behind
her, clearly visible, is a sign that says "We I.D. Everyone" .

Tags: identity ssi zero+trust zero+data authorization verifiable+credentials


MORE
February 150
18.6.2024, 16:30All Things Distributed (Feed)
What I've been reading since re:Invent
After a busy conference season, I've taken some time to catch up on reading and
make a dent in the pile of books on my nightstand. Here's what I've started,
finished, and picked up since re:Invent.

MORE
February 130
24.6.2024, 6:00SEI Blog (Feed)
Using ChatGPT to Analyze Your Code? Not So Fast
This blog post explores the efficacy of ChatGPT 3.5 in identifying errors in
software code.

MORE
February 130
22.6.2024, 19:00Platformonomics (Feed)
Platformonomics TGIF #39: February 9, 2024

Platformonomics TGIF is a weekly roll-up of links, comments on those links, and
perhaps a little too much tugging on my favorite threads.

Get Updates By Email

Subscribe

CAPEX Week is over! It is a huge step down, but on to the Superbowl!

News Follow the CAPEX: Cloud Table Stakes 2023 Retrospective

I published my annual look at hypercloud CAPEX spending this week. The best part
of publishing is all the people who come out of the woodwork with great
perspective to add. The worst part is all the jokes I think of right after
publishing.

The scale of hypercloud CAPEX investment is immense. But apparently not immense
enough:

Previous:

Follow the CAPEX: Cloud Table Stakes 2023 Retrospective

Introducing Platformonomics ClownWatch

CAPEX spending helped us separate the clowns from the clouds.

CAPEX is also a great lens for generative AI, which requires even more expensive
infrastructure (thanks NVIDIA!). Once again, we can contrast flowery rhetoric
with hard investment.

IBM (no surprise) is the first member of the Platformonomics ClownWatch list of
AI poseurs. IBM is running the same playbook they used for cloud: chant the
buzzwords, cut actual CAPEX spending, talk up irrelevant products, hope to book
some consulting business from less discerning enterprise customers, and try not
to go to jail for securities fraud.

Amazon CEO Andy Jassy said “every single business” at Amazon had “multiple”
generative AI initiatives. Except, it seems, the people who build out the AWS
infrastructure. In the middle of the AI boom where GPUs are the most coveted
items on Earth, AWS CAPEX investment was down 10% in 2023 (even as the business
grew by 13%). The disconnect between Amazon’s incessant AI chatter and shrinking
investment requires us to put AWS on negative watch for potential addition to
the ClownWatch AI list. Stay tuned.

Previous:

IBM and the Art of Misleading Investors, This Week in Securities Fraud, Clown
CAPEX, IBM’s “Cloud” Business (or Lack Thereof), IBM: How Much Longer Do the
Good Times Last?, The AWS Generative AI Soap Opera, The Amazon AI Inferiority
Complex

Viewer Mail #1

Q: Was the cloud infrastructure build-out a ZIRP phenomenon? AWS seems to have
ended debt financing about the time interest rates started to normalize.

A: I don’t think so. The cloud is a real business funded from cash flows and
still growing, which is the antithesis of ZIRP phenomena. Meta and Microsoft are
still using finance leases to fund server purchases. And in the case of Amazon,
I think they just have more cash compared to the early days (and they have more
cash because they’ve run out of things to invest in, including, it seems, AWS
infrastructure).

Previous:

The Great Startup Reset: Why Founders Should Prepare for Lower Valuations (ZIRP
phenomena)

Server Depreciation

I’m writing too much a lot about depreciation. Traditionally my view is gross
investment is what is interesting as that corresponds to the growth in
infrastructure. Depreciation is just an after-the-fact bean counter adjustment.
But depreciation adds up to real money (or at least real accounting adjustments)
when you have tens of millions of servers and extend their useful lives by 50%
(from four years to six). Moore’s Law needs to get back in gear or the
accountants will win this battle.

Previous:

BREAKING: Depreciation Schedules, Depreciation Schedules Meet Moore’s Law,
Follow the CAPEX: Cloud Table Stakes 2023 Retrospective

Software Migration Alert: Everbridge

As you know, when private equity comes amalgamating, it is time to start
migrating.

Previous:

Perfidious Private Equity, Private Equity in Action: VMware (Continued), Private
Equity in Action: VMware, Private Equity Will Buy and Ruin More Software
Companies, Software Migration Alert: VMware (Continued), Software Migration
Alert: VMware, Software Migration Alert: Rover, Software Migration Alert: Avid
Technology, Software Migration Alert: EngageSmart, This Week’s Software
Migration Alerts

So You Want to Build an AI Company: Episode 2

Their administrative bloat is your opportunity.

Previous:

So You Want to Build an AI Company


MORE
March 60
Follow the CAPEX: Cloud Table Stakes 2023 Retrospective

Tl;dr: Cloud CAPEX is once again both exciting and revealing!

Previous retrospectives: 2016, 2017, 2018, 2019, 2020, 2021, 2022 plus
earlier/other CAPEX musings.

Capital expenditures (CAPEX) are old school. You acquire some abiding apparatus
around which you build a business. It is the antithesis of the “asset light”
avoidance of hard work or the financial shenanigans that characterize many
trendy business models. Buy some (typically unglamorous and/or unwieldy)
machinery and earn back your investment over time.

The CAPEX that begets cloud infrastructure has been mostly uneventful in recent
years, ever since it helped us separate the clouds from the clowns (the cloud
pretenders who talked a good game but didn’t put real money where their mouths
were). But times change! Cloud CAPEX is again both exciting and revealing thanks
to generative AI.

Cloud CAPEX is fascinating at multiple levels. The sheer magnitude of hypercloud
CAPEX spending is breathtaking and glorious in and of itself. The hypercloud
trio, who I quaintly think of as software companies, are marshaling matter
across a variety of businesses (not just cloud infrastructure) on the same scale
as the world’s very largest metal-benders.  

Cloud infrastructure is a multi-hundred-billion-dollar industry, with the
majority of IT spend still an opportunity ahead of them. Hypercloud CAPEX not
only tells us what is happening in the market, but also what the providers think
will happen, as they must invest ahead of revenue.

And now generative AI is disrupting the cloud infrastructure business, toppling
architectures, customer priorities and industry pecking orders. Cloud CAPEX
gives perspective on how AI platitudes are turning into GPU purchases.

There is lots of good analysis about who has the most GPUs or which silicon
might challenge NVIDIA. As fun (and interesting) as that is, that is not our
purpose here. We’ll stick to what the cloud infrastructure lens reveals about
both cloud and AI investment. We’ll try to stick to the data and save the really
speculative conclusions for later. And we’ll even resist (almost) making fun of
the rise of GPUs as an asset class (because everything should be
financialized!).

Standard disclaimer: Unless otherwise indicated, the reported numbers are the
companies’ total CAPEX spend, not just cloud infrastructure, so includes land,
office buildings, campus redevelopments, warehouses, panopticonvenience stores,
manufacturing tooling, self-driving cars, delivery vehicles, flying machines,
flying delivery machines, satellite constellations, hardware that both is and is
not required for quantum computing, and – what should be the absolute top
priority for Congressional hearings – the still missing-in-action Google space
elevator. The numbers include finance leases for both Amazon and Microsoft, as
well as build-to-suit leases for Amazon (the leases are debt instruments used to
finance specific CAPEX expenditures, namely servers and buildings).

Get Updates By Email

Subscribe

Now the numbers!

SUMMARY

The three hypercloud companies – Amazon, Google, and Microsoft – collectively
spent over $127 billion on CAPEX in 2023, flat from 2022. That is company-wide
CAPEX. Their combined cloud infrastructure CAPEX was on the order of $80 billion
(we don’t get clean breakouts from Google or Microsoft).

Amazon’s (relative) CAPEX austerity continues, as the company spent a measly
$53.7 billion, a decline of 20%, following a 9% decline in 2022. Unlike last
year, it wasn’t just retail fulfillment retrenchment. AWS infrastructure
spending also declined for the first time ever. Given how much Amazon talks
about generative AI, it is a very odd year to cut AWS CAPEX.

Google eked out a 2% increase to $32.3 billion. Their much-anticipated 2022
server fleet refresh still hasn’t happened. There may be a glimmer of an AI
infrastructure uptick in the fourth quarter, but otherwise nothing that screams
“Code Red”.

Microsoft CAPEX grew 45% to $41.2 billion. Microsoft is patient zero in the mad
scramble for GPUs with its OpenAI partnership. OpenAI is nothing without GPUs,
and those GPUs are in Azure. Microsoft crossed the chasm from training to
inference at scale, serving hundreds of millions of monthly users. Microsoft’s
AI investments are impossible to miss.

Amazon still spends more than any other (non-Chinese1) company in the world,
while Google and Microsoft stand amongst the biggest CAPEX spenders on the
planet.

The three hypercloud companies’ cumulative CAPEX spend since 2000 is over $820
billion, with $250 billion of that spending in the last two years. Even in the
event of continued sobriety at Amazon, that number should easily pass a trillion
dollars invested in 2025.

Looking at CAPEX as a percentage of revenue shows Amazon’s decline back to 2015
levels, the fizzling out of Google’s quadrennial server refresh cycle, and
Microsoft’s generative AI spike.

This year we’ll review the hypercloud companies in reverse alphabetic order.

Microsoft

Microsoft has gone from perennially the most boring to the most interesting of
our CAPEX contestants. After being smoothly monotonic since 2016, and very
consistent as a percentage of revenue, Microsoft’s CAPEX jumped by 45% (+$12.8
billion) in 2023 (ChatGPT was introduced November 30, 2022).

FUN FACT: Microsoft spent more on CAPEX in 2023 than Oracle has in its entire
history.

The company said calendar fourth quarter CAPEX spending was actually
“lower-than-expected due to delivery for a third-party capacity contract
shifting from [fiscal year] Q2 to Q3” (NVIDIA or AMD?) and they expect “capital
expenditures to increase materially” in 2024. In June the company sent the
rarely-fielded CFO/CTO tag team to warn Wall Street of its coming CAPEX binge,
introduce the concept of software COGS, and assure them it would all pay off.

Microsoft is reputed to be the largest customer for both NVIDIA (Q4 2022, Q2
2023, 2H 2023) and AMD (plus is doing its own AI silicon). The abrupt 4% bump in
CAPEX as a percentage of revenue, after a steady 13-14% for years prior, is the
best proxy for the incremental AI spending in 2023. That suggests an incremental
AI-driven spend of about $9 billion, or 22% of overall CAPEX.

But this number understates Microsoft’s mad scramble for GPUs. The company is
also spending billions to rent capacity from various boutique GPU clouds
(including CoreWeave, Lambda Labs, and Oracle). That matrix multiplication
machinery shows up as OPEX, not CAPEX.

Microsoft even identifies GPU availability as a risk factor in its 10-K:

Our datacenters depend on the availability of permitted and buildable land,
predictable energy, networking supplies, and servers, including graphics
processing units (“GPUs”) and other components.

Now they just have to turn all this investment into profitable business. And
profit margin expectations at Microsoft are very high…

Google

Google has been a disappointment to CAPEX obsessives for the last two years. We
were hoping for their fifth quadrennial server upgrade cycle to hit in 2022 with
CAPEX spiking back to at least 17% of revenue, thereby boosting annual CAPEX to
over $48 billion.

Instead, we got an anemic bump from 2021’s nine-year low of 10% of revenue all
the way up to 11% in 2022 (it was still 28% dollar growth to $31.5 billion in
CAPEX given their ever-growing revenue base). In 2023 Google CAPEX nudged up 2%
to $32.3 billion.

We remain hopeful that despite the bean counters seemingly winning the battle
between depreciation schedules and a withering Moore’s Law, that fifth refresh
cycle is merely delayed as the useful accounting life of servers has been
extended from four years to six years.

It is amusing that on the list of “discipline” and “efficiency” initiatives
Google felt worth surfacing to Wall Street is improved “machine utilization”.
Presumably that means there are material savings to be had (they also announced
they were eliminating staplers2). One can easily imagine individual development
teams having (and having forgotten about) tens of thousands of servers in their
couch cushions3. So there is a scenario where Google’s infrastructure was
(vastly?) overbuilt and some of it can be recycled to meet growth demands.

FUN FACT: Google spent more on CAPEX in 2023 than IBM has over the last decade.

On the AI front, Google faces the most interesting strategic quandary of the
hyperclouds. Google invented transformers and has long led in AI research.
Letting others capitalize on their inventions would (merely) merit a Fumbling
the Future sequel. But to the degree generative AI poses a real and disruptive
threat to Google’s search franchise (one of the best businesses ever), their
response is both incredibly important and incredibly closely scrutinized.

That response, as Google plays strategic defense for the first time, has been
muddled. The CEO dedicated time to denying he invoked a “Code Red” response to
ChatGPT (he certainly should have invoked a “Code Red”, so why deny it?)4. They
are (appropriately) trying to downplay any disruption to search, but a
competitive version of Gemini is late. Questions are coming from inside the
house about management’s ability to transition the company from peacetime to
wartime.

Infrastructure should be an immense advantage for Google as they respond to the
“Code Red”. But that 2023 bump in CAPEX of 2% doesn’t even keep up with
inflation (though IBM has been triumphant about that level of “growth”). So
there was no infrastructure “Code Red” in 2023, and the contrast with Microsoft
is stark.

But if we drill down to the quarterly trend, Google’s CAPEX did grow
sequentially through 2023, with over a third of their annual spend coming in the
fourth quarter. We expect that will continue in 2024, with Google’s guidance of
“notably larger” CAPEX. The sleeping giant is at least stirring, but they’re
still at least a year behind Microsoft.

In terms of the implications of AI’s rise for the cloud computing business, I’ve
argued that Google Cloud is a hobby for Google, well down the CEO’s list of
priorities. If Google’s core franchise is at risk, I think hobby status becomes
even more pronounced, as (strategic) attention is all you have5.

Amazon

The biggest cloud CAPEX news from 2023 is the first ever decline in AWS
infrastructure investment. Amazon’s overall CAPEX spending fell 20% to $53.7
billion (still the biggest corporate CAPEX spend ex-China), while AWS spending
fell 10% to $24.8 billion. AWS was outspent on infrastructure this year by
Microsoft and perhaps also Google.

FUN FACT: Amazon in austerity mode still spent more on CAPEX in 2023 than all
three US mobile operators combined (ATT, T-Mobile, Verizon)

There are two broad dynamics in play here. The first is regime change (both
financial and in leadership) at Amazon, combined with a lot of previous excess
(some called it #bonkers but “CAP-EXcess” would have been good too) that needed
to be “optimized”.

For a long time (call it the Bezos Doctrine), Amazon tried to reinvest every
single penny back into the business and minimize accounting profits. By 2015,
they couldn’t reinvest cash flow fast enough and the company began to show
growing profits and free cash flow (despite their very best intentions).

The pandemic interrupts this trend, as Amazon revenues soared with people stuck
at home with nothing to do but shop, while the company simultaneously mounted
one last throwback #bonkersbonkers retail infrastructure investment blowout in
2021 (they invested $73.7 billion in CAPEX to double the size of their
fulfillment network and build their own UPS-sized transportation network,
overbuilding just as the pandemic ended, a binge from which they are still
cleaning up “optimizing”).

But now Amazon is inexorably reverting to that pre-pandemic profit curve. Amazon
is now “a profit deal”, which means costs, including CAPEX, get actual scrutiny.

New CEO Andy Jassy has embraced financial optimization with gusto (broad
layoffs, asking why Alexa exists much less is losing $5 billion a year, adding
“rinky dink” ads to Prime Video, etc.), and has not exempted his baby AWS. This
regime change probably marks the calcification of Amazon’s footprint, as once
you become a profit optimizer, it is very hard to deviate from that path to make
big, bold, new investments. But beyond whacking the money-losing projects, they
still need to keep the franchise retail and cloud businesses growing and
profitable. And both of those require massive CAPEX.

Beyond the regime level changes, 2023’s material decline in AWS CAPEX is quite
alarming for what it says about both the existing cloud business and how they
are faring in the AI race.

AWS has been reinvesting ~35% of revenue in recent years in cloud
infrastructure. If you squint at the chart below, you’ll see the rate of
investment roughly turning into a similar level of revenue growth the following
year. But that relationship has broken down in the last two years, as revenue
growth has slowed to 12-13%. That suggests that AWS is overbuilt relative to
recent revenue growth.

But a surplus of older gear doesn’t help with generative AI. One observation
about Amazon generally, and AWS in particular, is they don’t tend to talk about
things unless they’re behind. And they’ve been running their mouth non-stop
about generative AI, so cutting billions in CAPEX spend alongside that
rhetorical frenzy is quite a surprise.

Amazon and AWS face a host of challenges around generative AI. Despite all the
rhetoric, they’re way behind, and this is a space where catching up is
especially hard. Their Titan LLM is missing in action (and on the wrong side of
their own “there is no compression algorithm for experience” quip). They had to
license Anthropic for Amazon’s own internal use, though the announcement was
dressed up to look like an endorsement of AWS’s generative AI infrastructure.

But worse, their existing infrastructure may actually be a hindrance. It has
been argued AWS’s infrastructure is “poorly adapted” for the generative AI
world. And they clearly got sideways with NVIDIA, which means their H100
shipments are “significantly lower than their share of the public cloud”.

It looks like Amazon tried to push NVIDIA around, as they’re accustomed to
doing, but lost. The power dynamics have changed and NVIDIA now has the upper
hand6. Amazon resisted for months, but eventually capitulated, and had to host
NVIDIA’s DGX Cloud (and even say nice things about it despite really not wanting
to do it) to get more GPUs.

But beyond product and vendor travails, slashing CAPEX by 10% in the middle of
the AI race is an enormous red flag (and different than a Code Red). It makes
Amazon’s incessant claims to generative AI leadership all the harder to stomach.
As a result, we are adding AWS to the negative watch list, putting them at risk
of being added to the “AI clown” list (where they would join initial inductee
IBM). If Google is at least a year behind Microsoft, AWS is even further behind.
And their guidance doesn’t suggest they have a clear CAPEX plan for 2024:

CapEx will go up in 2024. I’m not giving a number today, but we do — we’re still
working through plans for the year, but we do expect CapEx to rise as we add
capacity in AWS for region expansions, but primarily the work we’re doing with
generative AI projects.

As a final observation for the most hard core of CAPEX obsessives, AWS’s use of
finance and built-to-suit leases have dropped to almost nothing (less than half
a percent of AWS CAPEX in 2023). This is notable as they were ~80% as recently
as 2019. It was also interesting to hear the CFO say, “We define our capital
investments as a combination of CapEx plus equipment finance leases.” So no
respect for build-to-suit leases.

What’s Next?

There are a couple things on the CAPEX agenda:

Triangulate the cloud CAPEX numbers against NVIDIA’s numbers when they report in
a couple weeks. Add Meta to the CAPEX club. They’re not a hypercloud (and aren’t
likely to be given their developer DNA, or lack thereof), but they spend almost
as much as the hyperclouds on CAPEX ($28.1 billion in 2023). And they’re
gobbling GPUs. How much of that CAPEX supports Facebook/Instagram/Whatsapp vs
metaverse/hardware vs. new things? Are they building generative AI-powered
search? Check in on the clown car race amongst the CAPEX pretenders (after
Oracle announces earnings).

Please comment below (or contact me privately) with what I got wrong here and
what else I should know about!



Because I neither believe nor want to look up the Chinese CAPEX numbers ︎Perhaps
Google, with all its AI experience, considers staples a bigger risk than the
more traditional existential risk of paper clips? ︎My bet is there are
individual hypercloud dev team environments with more servers than IBM Cloud.
︎“Schrödinger’s Code Red” would be a good application for Google’s quantum
computer. ︎It is probably a bad sign when a joke needs an explanatory footnote,
but that was an “Attention is All You Need” reference. I also considered mashing
that up with “You Can’t Always Get What You Want“. ︎To the point where NVIDIA
are trying to revive the vertically integrated computing model of yore, spanning
chips, systems, software and cloud services. But that is a topic for another
day. ︎

MORE
March 60
Today, 14:28Phil Windley's Technometria (Feed)
Acceptance Networks for Self-Sovereign Identity

Summary: We can't have broad adoption of verifiable credentials until we find a
way to scale their presentation by providing tooling that credential verifiers
can use to reduce their risk and gain confidence in the facts presented to them.

When I hand a merchant in London a piece of plastic that I got from a bank in
Utah to make a purchase, a tiny miracle happens. Despite the fact that the
merchant has never met me before and has no knowledge of my bank, she blithely
allows me to walk out of the store with hundreds of dollars of merchandise,
confident that she will receive payment. I emphasized the word confident in the
last sentence because it's core to understanding what's happened. In the past,
these kinds of transactions required that the merchant trust me or my bank. But
in the modern world, trust has been replaced by confidence.

We often mix these concepts up and I'm as guilty as anyone. But trust always
involves an element of risk, whereas confidence does not. These are not binary,
but rather represent a spectrum. In the scenario I paint above, the merchant is
still taking some risk, but it's very small. Technology, processes, and legal
agreements have come together to squeeze out risk. The result is a financial
system where the risk is so small that banks, merchants, and consumers alike
have confidence that they will not be cheated. There's a name in the financial
services industry for the network that reduces risk so that trust can be
replaced with confidence: an acceptance network.

Acceptance Networks

An acceptance network is the network of merchants or service providers that
accept a particular form of payment, usually credit or debit cards, from a
particular issuer or payment network. The term refers to a broad ecosystem that
facilitates these transactions, including point-of-sale terminals, online
payment gateways, and other infrastructure. Each component of the acceptance
network plays a crucial role in ensuring that transactions are processed
efficiently, securely, and accurately. This drives out risk and increases
confidence. Acceptance networks are foundational components of modern payment
ecosystems and are essential to the seamless functioning of digital financial
transactions. Visa, Mastercard, American Express, and Discover are all examples
of acceptance networks.

Before the advent of acceptance networks, credit was a spotty thing with each
large merchant issuing it's own proprietary credit card—good only at that
merchant. My mom and dad had wallets full of cards for JC Penney, Sears,
Chevron, Texaco, and so on. Sears trusted its card. Chevron trusted its card.
But it was impossible to use a Chevron card at Sears. They had limited means to
verify if it was real and no way to clear the funds so that Chevron could pay
Sears for the transaction.

That scenario is similar to the state of digital identity today. We have
identity providers (IdPs) like Google and Apple who control a closed ecosystem
of relying parties (with a lot of overlap). These relying parties trust these
large IdPs to authenticate the people who use their services. They limit their
risk by only using IdPs they're familiar with and only accepting the (usually)
self-asserted attributes from the IdP that don't involve much risk. Beyond that
they must verify everything themselves.

Fixing this requires the equivalent of an acceptance network for digital
identity. When we launched Sovrin Foundation and the Sovrin network1 in 2016, we
were building an acceptance network for digital identity, even though we didn't
use that term to describe it. Our goal was to create a system of protocols,
processes, technology and governance that would reduce the risk of
self-sovereign identity and increase confidence in an identity system that let
the subjects present verifiable credentials that carried reliable attributes
from many sources.

I've written previously about identity metasystems that provide a framework for
how identity transactions happen. Individual identity systems are built
according to the architecture and protocols of the metasystem. Acceptance
networks are an instantiation of the metasystem for a particular set of users
and types of transactions. A metasystem for self-sovereign identity might have
several acceptance networks operating in it to facilitate the operation of
specific identity systems.

Problems an Acceptance Network Can Solve

To understand why an acceptance network is necessary to reduce risk and increase
confidence in identity transactions, let's explore the gaps that exist without
it. The following diagram shows the now familiar triangle of verifiable
credential exchange. In this figure, issuers issue credentials to holders who
may or may not be the subject of the credentials. The holder presents
cryptographic proofs that assert the value of relevant attributes using one of
more of the credentials that they hold. The verifier verifies the proof and uses
the attributes.

Verifiable Credential Exchange (click to enlarge)

Let's explore what it means for the verifier to verify the proof. The verifier
wants to know a number of things about the credential presentation:

Were the credentials issued to the entity making the presentation? Have any of
the credentials been tampered with? Have any of the credentials been revoked?
What are the schema for the credentials (to understand the data in them)? Who
issued the credentials in the proof?

The first four of these can be done cryptographically to provide confidence in
the attestation. The technology behind the credential presentation is all that's
necessary. They can be automated as part of the exchange. For example, the proof
can contain pointers (e.g., DIDs) to the credential definitions. These could
contain public keys for the credential and references to schema.

The last one—who issued the credential—is not a technical matter. To see why,
imagine that Alice (as holder and subject) has been issued a credential from her
university (the issuer) giving information about her educational experiences
there. She's applying for a job and wants to present the credential to a
prospective employer (the verifier). How does the employer know that Alice
didn't just make the credential herself or buy it from a diploma mill?

Knowing who issued the credential is not something that can be done solely with
technology (although it can help). The employer in this scenario wants more than
an identifier for the issuer. And they want to know that the public key really
does belong to the university. In short, the employer wants to resolve the
identifier to other information that tells them something about the university
and the credential. There are lots of ways to do that—people have been doing
this sort of thing for centuries: states keep registries of businesses
(universities are businesses), accreditation organizations keep registries of
schools they've accredited, the Department of Education has registries of
various institutions of higher education in the US, and so on.

The employer could make use of these by building its own database of university
identifiers it trusts. And every time a new one shows up, they could investigate
and add it to their registry (or not)2. But going back to the magic of the
credit card scenario that I opened this article with, if every merchant had to
keep their own registry of banks, the experience wouldn't be magical for me or
the merchant. The financial acceptance network makes it easy for the merchant to
have confidence that they'll be paid because they have not only technology, but
processes, protocols, governance, and legal agreements that make the
verification process automatable.

Acceptance Networks for Digital Identity

For some use cases, keeping your own registry of the issuers you trust works.
But for many, it's just too much work and makes it difficult to make use of a
variety of credentials. This kind of "localized trust" is unwieldy in an
identity system that might involve millions of issuers and identifiers and
credentials for billions or even trillions of subjects. I've written extensively
about identity metasystems and what they provide to help bridge the gap. This
one, on how metasystems help provide life-like identity for digital systems is
perhaps the most comprehensive. Acceptance networks implement metasystems.

An acceptance network for digital identity must have a number of important
properties, including the following:

Credentials are decentralized and contextual—There is no central authority for
all credentials. Every party can be an issuer, a holder (identity owner), or a
verifier. Verifiable credentials can be adapted to any country, any industry,
any community, or any set of trust relationships.

Credential issuers decide on what data is contained in their credentials—Anyone
can create a credential schema for their use case. Anyone can create a
credential definition based on any of these schemas.

Verifiers make their own trust decisions about which credentials to
accept—There's no central authority who determines what credentials are
important or which are used for what purpose. The acceptance network supplies
the technical underpinnings for credential exchange and support protocols for
automating the verification of credential issuers.

Credential verifiers don't need to have any specific technical, contractual, or
commercial relationship with credential issuers—Verifiers do not need to contact
issuers to perform verification.

Credential holders are free to choose which credentials to carry and what
information to disclose—People and organizations are in control of the
credentials they hold (just as they are with physical credentials) and determine
what to share with whom.

You may be thinking "but these are mostly about decentralized decision making."
While it would be easier to imagine the acceptance network as a big directory,
that solution can't possible support all the different ways people and
organizations might want to use credentials. That doesn't mean an acceptance
network couldn't be run by a single organization, like some financial services
networks. Just that it has to support a variety of credential ecosystems running
common protocols. I also think that there will be more than one and most issuers
and verifiers will be part of several (again, like in financial services).

Structure of an Acceptance Network

One of the things we can take away from the architecture of financial services
acceptance networks is that they are built in layers. No one has thought more
about how this can work than Drummond Reed and the Trust Over IP Foundation
(ToIP).3 This figure, from ToIP, shows how such a stack works.

Trust Over IP Stack (click to enlarge)

The layers build on each other to provide something the lower level didn't.
Layer 1 is the foundational functionality, like DID methods. Layer 2 builds on
that to support creating digital relationships with anyone. Layer 3 uses those
relationships to effect credential exchange. Layer 4 is the ecosystems that say
things about the issuers for different use cases. The dual stack emphasizes the
need for governance at every layer.

The acceptance network specifies the accepted protocols and technologies. The
acceptance network also supports ecosystems, providing governance models and
technology. The acceptance network is involved at each layer. Here are some
examples of things an acceptance network might do at each layer:

Layer 1—limit the allowed DID methods and certify them. Layer 2—require that
wallets and agents using the network support specific versions of the DIDComm
protocol. Provide a certification framework for wallet and agent vendors for
security and interoperability. Layer 3—require specific versions of the exchange
protocols. Participate in protocol development. Provide a certification
framework for specific implementations to aid with security and
interoperability. Layer 4—support the formation, certification, and discovery of
credential ecosystem providers. Govern what is required to be a certified
ecosystem provider and provide models for acceptable ecosystem governance.

As part of it's overall governance of the ecosystem, the acceptance network also
provides model legal agreements for and between the various participants, trust
mark rights (think of the Visa logo), and drives a uniform user experience.

The following diagram shows the credential exchange from the preceding figure
with an acceptance network providing support to the verifier so that it can have
confidence in the data the issuer has supplied through the holder.

Acceptance Network in Operation (click to enlarge)

Credential issuers who know their credential might be widely used would join one
or more acceptance networks. They agree to follow the rules and regulations in
the governance framework of the acceptance network. The acceptance network
issues a credential to them that they can use to prove they are a member.4 The
acceptance network maintains a registry—likely a registry of registries—that
verifiers can use to discover information about the issuer of a credential that
has been presented to them.

Using an Acceptance Network

Returning to our previous scenario, Alice holds a credential issued by her
university. She presents it to a prospective employer who wants to know that the
credential is from an accredited university. Alice's university has been
accredited by an accreditation organization5. They have followed their process
for accrediting Alice's university and issued it a credential. They have also
added the university to their registry. The university and the accrediting
organization are members of an acceptance network. The employer's systems know
to automatically query the acceptance network when it received a credential
proof from a issuer it does not know. Doing so provides the assurance that the
issuer is legitimate. It could also provide information about the accreditation
status of the university. This information reduces the risk that the employer
would otherwise bear.

In this scenario, the employer is trusting the processes and structure of the
acceptance network. The employer must decide which acceptance networks to use.
This is much more scalable than having to make these determinations for every
credential issuer. The acceptance network has allowed the verification process
to scale and made the overall use of verifiable credentials easier and less
risky.

A Note on Implementation

This discussion of acceptance networks has undoubtedly brought images to your
mind about how it is structured or how to build one. The comparison to financial
services acceptance networks points to a network run by an organization. And the
term registry brings to mind a database of some kind. Why these are certainly
possibilities, I think it's also possible to imagine more decentralized
solutions. For example, the registry could be a distributed ledger or
blockchain. The governance is likely most easily done by an organization, but
there are other options like a decentralized autonomous organization (DAO). The
scenario I described above illustrates a federated system where certifying
authorities for specific ecosystems determine their own methods, processes, and
requirements, but link their registry to that of the acceptance network.

Conclusion

As I mentioned above, we've been solving the problem of how to know which
institutions to trust for centuries. We have ways of knowing whether a
university is accredited, whether a bank is real, whether a company is actually
registered and what its reputation is. What is missing is an easy way to make
use of this information digitally so that processes for reducing risk can be
automated. Acceptance networks rationalize the process and provide the needed
tooling to automate these checks. They reduce the many-to-many problem that
exists when each verifier has to determine whether to trust each issuer with a
more scalable many-to-several system. Acceptance networks allow credential
presentation to scale by providing the needed infrastructure for giving
verifiers confidence in the facts that holders present to them.

Notes You can see in the linked post how we used trust to describe what we were
building, even as we were reducing risk and inspiring confidence. Note that this
investigation could make use of technology. Knowing the universities name, they
could look up a well known location on the universities web site to find the
identifier. They could use PKI (digital certificates) to be sure they're talking
to the right place. They could look up the university in an online registry of
accredited universities. Trust over IP isn't the only one working on this. Marie
Wallace of Accenture and Stephen Wilson of Lockstep Partners have been writing
about this idea. Note that there could be different levels or types of members
who perform different roles in the ecosystem and make different agreements. An
example is the Northwest Commission on Colleges and Universities.

Photo Credit: Data flowing over networks from DALL-e

Tags: ssi identity decentralization


MORE
February 70
1.5.2024, 16:56James Governor's Monkchips (Feed)
Why you should attend the State of Open Conference 2024 in London this week

This week it’s The State of Open Conference 2024 at The Brewery, London. It was
great last year. I heartily recommend you attend. As I said on twitter at the
time:

The UK now has its own OSCON. the event will happen again and will go from
strength to strength.

Attendees and speakers were a who’s who of open source, open hardware and open
culture people generally. The open data track was particularly lively. There
were so many of my friends there, it really felt like my people had all
congregated in London for the day. The event also felt very inclusive, in terms
of both speakers, but also attendees. It reflected London’s rich diversity.

The speaker list is extremely impressive again this year.

But the real reason I think SOOCon24 is so important is the focus on policy,
governance and open source sustainability. Open source is under a great deal of
pressure right now. VCs are encouraging their portfolio companies to adopt
“business source” licenses, which are not actually open source. Why does this
matter? As my colleague Stephen O’Grady argues:

A world in which non-compete licensing grows at the expense of open source is
problematic enough. A world in which vendors blur the definition of open source
such that regular users can no longer differentiate between the two is much,
much worse.

Pedantic as it may seem, then, the question of whether something is actually
open source really does matter, as those who would redefine the term will find
out if they get their way.

This movement has also bled into the current AI explosion. What is “Open” AI?
That’s something we need to work out – and major market players are casually
calling things open source, which frankly aren’t. Another area of governance and
policy under scrutiny is regulation of AI – we can’t just leave this as the era
of “You Only Live Once.” Controls will be necessary, and governments are
scrambling to put them in place. At SOOCon24 the organisation behind the
conference Open UK will be capturing opinions and data to feed back the UK
government about regulation going forward. I believe we’re going to see AI Bill
of Materials requirements regulated at national level.

It’s a pivotal time, and these discussions are vitally important – that’s why
they need a home. We’re literally talking about the economic foundations of the
digital economy, the means of production which have served us pretty well these
past couple of decades, and the opportunities for making and learning which have
made tech such a transformative success. Authors and creators need stable
foundations to work on. Copyright and licensing matters. Back to Stephen:

Instead of the embarrassment of riches of open source projects we have today
that developers may take up and use for whatever they choose without reference
or restriction, we’d have a world dominated by projects carrying varying,
conflicting usage restrictions that would render the licenses incompatible with
one another and not usable by some.

I am glad Amanda Brock and team are pulling this event together, for all of the
reasons outlined above, and I look forward to seeing you there. I believe there
are a few tickets available.

If you’re interested in AI and prompt engineering, and all of the craft,
sustainability and social angles, you should also check out my conference Monki
Gras 2024: Prompting Craft. March 14th and 15th, Shoreditch London. Tix here.


MORE
February 50
24.6.2024, 6:00SEI Blog (Feed)
Applying the SEI SBOM Framework
This SEI Blog post examines ways you can leverage your software bill of
materials (SBOM) data, using the SEI SBOM Framework, to improve your software
security and inform your supply chain risk management.

MORE
February 60
22.6.2024, 19:00Platformonomics (Feed)
Platformonomics TGIF #38: February 2, 2024

Platformonomics TGIF is a weekly roll-up of links, comments on those links, and
perhaps a little too much tugging on my favorite threads.

Get Updates By Email

Subscribe

You’re all no doubt recovering from this week’s CAPEX festivities.

News



Follow the GPUs: Cloud CAPEX

We’re wrapping up the Superbowl for CAPEX obsessives, when the hypercloud
companies reveal their annual infrastructure investment. After a few boring
years, CAPEX is back to the center of the world as generative AI requires new
(and expensive) infrastructure. As in the early days of cloud, CAPEX is a great
tell on who is actually investing in AI and who is just bloviating (obligatory
notice: you can’t spell bloviating without A and I). I’ll publish a full
breakdown in the next couple days, but will leave you with a few tastes here.

Collectively, Amazon, Google and Microsoft’s corporate CAPEX spending (so more
than just cloud infrastructure) was nearly unchanged from 2022, while their
cumulative CAPEX spending this century passed $820 billion. But Amazon continues
to drop from their #bonkers pandemic spend, while Google was up just 2% and
Microsoft up 45% (GPUs!):

Amazon’s new-found restraint expanded beyond the retail business this year, with
AWS CAPEX spending declining for the first time ever, even as AWS revenue grew
by 13%:

Stay tuned for the full annual report.

Previous:

CAPEX

Antitrust Incoherence: Roomba Edition

Another scrapped acquisition underscores the utter incoherence of antitrust
policy today. It should be obvious we need a coherent doctrine that gives
everyone a sense of where the lines are on the playing field.

Roomba was a middling, non-strategic acquisition in a competitive market, but
nicely illustrates the insanity of the different doctrines in play today:

European Antitrust (EU) prioritizes protecting European firms (like this Bosch
robotic vacuum cleaner boasting of its “five ways vacuuming just became more
fun”!) and raising revenue from American tech firms (after all, a continent
can’t live just on wine and cheese). Hipster Antitrust (US FTC) says no to all
Big Tech acquisitions, because Big Tech is Bad! (The hipster antitrust white
whale is the Instagram acquisition, and some day, they fear, some other
acquisition might also be successful). The middlemen oligopolies at the heart of
our insanely bloated and inefficient healthcare system? Evidently not an
antitrust problem. VC Antitrust says yes to all acquisitions because, dammit,
nothing is more important than VC returns. (I have yet to hear a coherent
antitrust argument for approving Adobe’s failed acquisition of Figma, which runs
afoul of traditional antitrust. And note the FTC didn’t actually move against
the Figma acquisition, providing yet another huge indictment of their
priorities).

Needless to say, none of these are good doctrines.

And note private equity gets credit for putting Roomba into “a debt
straightjacket” after the deal was scrapped. When private equity comes
amalgamating, it is always time to start migrating.

Previous:

A New Antitrust Doctrine

Private Equity in Action: Data Centers

“This is Blackstone Inc.’s bet on the AI revolution.”

Just putting a marker down on this one.

Previous:

Private Equity in Action: Media Division, Private Equity in Action: VMware
(Continued), Private Equity in Action: VMware, Private Equity Will Buy and Ruin
More Software Companies, Software Migration Alert: VMware, Software Migration
Alert: VMware (Continued), Perfidious Private Equity

IBM and the Art of Misleading Investors

IBM opened their Q4 earnings release with:

“In the fourth quarter, we grew revenue in all of our segments, driven by
continued adoption of our hybrid cloud and AI offerings. Client demand for AI is
accelerating and our book of business for watsonx and generative AI roughly
doubled from the third to the fourth quarter,” said Arvind Krishna, IBM chairman
and chief executive officer.

You’d almost think IBM was an AI company and seeing real revenue from AI. But
when asked to “elaborate specifically on exactly what the book of business
means”, much bobbing and weaving ensues:

So Toni, on the AI book of business, this is not all revenue in the quarter. I
would just begin with that statement to set it straight. At this stage, we
wanted to start looking at what is our momentum, what is the sentiment from our
clients. So we went to a measure that is more reflective of, I’ll use the word
signings. What is the commitment the clients are making to us? Consulting is
straightforward. It is the signings. Consulting signings are anywhere from 12 to
24 months on average is how much time they play out over there. And on software,
it’s what they’re committing to. And we are using SaaS ACV. So it’s a 12-month
commitment, which is typical for as a service as well as, since we do offer our
portfolio both ways as license or as a service, it includes the license piece as
well. Now over a long-term, let’s call it a couple of years or more, yes, the
book of business should turn into an amount of revenue in a quarter, but that’s
going to take a bit of time to catch up. But we felt that this gives the better
indicator right now of what is our traction and what is our acceleration in that
part of the business.

IBM is running the same playbook they ran for cloud. Chant the buzzwords and
hope investors won’t notice IBM isn’t actually investing and don’t have
competitive products. As with cloud, IBM’s AI results are not something they
actually report to the SEC, yet lead with in investor communications.

Previous:

This Week in Securities Fraud, IBM’s “Cloud” Business (or Lack Thereof), IBM:
How Much Longer Do the Good Times Last?

Move Fast and Regulate Things (You Don’t Understand)

“The first law of this kind”. Indeed.

Previous:

The EU Will Continue to Fiddle While Rome Burns

The GPU Chip Race

I spoke to the New York Times about the big tech companies creating their own
GPUs to get out from under NVIDIA’s thumb. In my defense, this is a rare topic
where the Times doesn’t have a conflict of interest in their tech coverage.

Previous:

GPUs are CAPEX; CAPEX is Increasingly GPUs, Living in a Glass House: The New
York Times On Tech, Living in a Glass House: What is the New York Times?

“Supercloud”: Still Dead

I’m not sure if this is an attempted exit strategy (finally!) or yet another
demonstration that “supercloud” has always been a definitional random walk with
a dash of Tourette’s Syndrome (more likely), but the sultans of “supercloud” are
taking a victory lap for a “prediction” that their architectural baby is just
CloudFlare. And if CloudFlare had a good year, so too must have “supercloud”!
Never mind that the majority of prior definitions explicitly excluded the idea
of CloudFlare being the Platonic form of “supercloud”. But they do seem to have
finally noticed that generative AI is a thing so hopefully they’ll move onto
that and dial down the supercloudifragilisticexpialidociousness.

Previous:

“Supercloud” RIP

Trade Surplus Region Complains Trade Deficit Region is “Protectionist”

Mercantilist nations with surplus manufacturing capacity have concerns about
where they will be able to dump their surplus.

Previous:

The EU Will Continue to Fiddle While Rome Burns

Related:

European Union Trade & Investment Summary, Trade Wars Are Class Wars

Useful Idiots, Fellow Travelers and Unregistered Foreign Agents: Harvard
University

Henceforth we shall refer to China as West Taiwan.

Previous:

Useful Idiots, Fellow Travelers and Unregistered Foreign Agents: Wall Street,
Useful Idiots, Fellow Travelers and Unregistered Foreign Agents: Apple

With Twitter engagement circling the drain, please comment here.


MORE
March 60
Today, 17:55ongoing by Tim Bray (Feed)
Vision 1993

I’ve plowed through the first wave of AVP (Apple Vision Pro) reviews, and it
seems pretty obvious that, at the current price and form factor, it’s not gonna
be a best-seller. But I remain a strong believer in Augmented Reality (the AVP
is VR not AR, for the moment). As I was diving into the reviews, a little voice
in the back of my head kept saying “I once read about what this is trying to
be.”

What I was remembering was Virtual Light, a 1993 novel by William Gibson,
allegedly set in 2006. It paints a clear picture of a future that includes AVP’s
descendants. So I re-read it. Maybe looking back is the way to look forward.

But first…

I wanted to say: It’s a terrific book! If you haven’t read it you might really
like it. I hadn’t in years and I sure enjoyed the re-read. The people in it are
charming, and it’s built around a fabulous physical artifact that drives the
plot. No, I don’t mean AR goggles, I mean San Francisco’s Bay Bridge, which in
Virtual Light and two subsequent novels, has been wrecked by an earthquake and
become a huge countercultural shantytown, one of the coolest venues Gibson has
ever invented, and that’s a strong statement. Also, protagonists Chevette and
Rydell are two of his best characters; another strong statement.

Anyhow, I don’t think it’s much of a spoiler to say that the AR devices I’m
writing about, despite being what the title refers to, are peripheral to the
plot. It turns out that one such device contains information that’s secret
enough to attract hired killers, skip tracers, and crooked Homicide cops to
recover it when it gets stolen; plenty of plot fuel right there.

Quoting

Here are a few out-takes from the book describing the titular technology.

Quote:

Nothing in it but a pair of sunglasses, expensive-looking but so dark she hadn’t
even been able to see through them last night.

Quote:

…she took that case out.

You couldn’t tell what it was made of, and that meant expensive. Something dark
gray, like the lead in a pencil, thin as the shell of one of those eggs, but you
could probably drive a truck over it… She’d figured out how you opened it the
night before: finger here, thumb there. It opened. No catch or anything, no
spring… Inside was like black suede, but it gave like foam under your finger.

Those glasses, nested there. Big and black. Like that Orbison in the poster… She
pulled them from the black suede… They bothered her … they weighed too much. Way
too heavy for what they were, even with the big earpieces. The frames looked as
though they’d been carved from slabs of graphite.

She put them on. Black. Solid black.

“Katharine Hepburn.” Skinner said.

Quote:

Warbaby wore a black Stetson set dead level on his head, the brim turned up all
the way around, and glasses with heavy black frames. Clear lenses, windowpane
plain.

Quote:

“You date you some architects, some brain-surgeons, you’d know what those are…
Those VL glasses. Virtual light.”

“They expensive, Sammy Sal?”

“Shit, yes. ’Bout as much as a Japanese car… Got these little EMP-drivers around
the lenses, work your optic nerves direct. Friend of mine, he’d bring a pair
home from the office where he worked. Landscape architects. Put ’em on, you go
out walking, everything looks normal, but every plant you see, every tree,
there’s this little label hanging there, what its name is. Latin under that…”

Quote (at a crime scene with Warbaby and Freddie):

Rydell noticed the weight as he slid them on. Pitch black. Then there was a
stutter of soft fuzzy ball-lightning, like what you saw when you rubbed your
eyes in the dark, and he was looking at Warbaby. Just behind Warbaby, hung on
some invisible wall, were words, numbers, bright yellow. They came into focus as
he looked at them, somehow losing Warbaby, and he saw that they were forensic
stats.

“Or,” Freddie said, “you can just be here now —”

And the bed was back, sodden with blood, the man’s soft, heavy corpse splayed
out like a frog. That thing beneath his chin, blue-black, bulbous.

Rydell’s stomach heaved, bile rose in his throat, and then a naked woman rolled
up from another bed, in a different room, her hair like silver in some
impossible moonlight—

Rydell yanked the glasses off…

Quote:

“Here. Check it out.” He put them on her.

She was facing the city when he did it. Financial district… “Fuck a duck,” she
said, those towers blooming there, buildings bigger than anything, a stone
regular grid of them, marching in from the hills. Each one maybe four blocks at
the base, rising straight and featureless to spreading screens likke the
colander she used to steam vegetables. Then Chinese writing filled the sky.

Hmmm…

What does Gibson’s 30-year-old vision teach us?

The devices are still heavier than you’d like, but light enough to wear all the
time out in the real world.

Still expensive.

They look super-cool.

They are transparent while in use.

You can use them to show pictures or share information the way you would today
by handing over a phone or tablet.

How you get information into them was as un-solved in 1993 as it is today.

But the real core value is the “A” in “AR” — augmenting an aspect of the real
world that you’re looking at. Even if only by hanging text labels on it.

For me, that last point is at the center of everything. I want to be in a park
at night and see fiery snakes climbing all the trees. I want to walk into a
big-box store and have a huge glowing balloon appear over the Baking Supplies. I
want floating labels to attach to all the different parts of the machine I’m
trying to fix.

Watching TV, by yourself, on a huge screen, is not the future. Augmenting
reality is.

The AVP? Some of its tech constitutes necessary but far from sufficient steps on
the way from here to that 1993 vision.


MORE
March 280
DC Again

For the second time this winter I’ve been to Washington and back. Herewith
pictures and feelings. Everyone’s seen plenty of photos of The Capital City Of
The United States so I’ve tried for fresh views. My feelings are more mainstream
than my pictures, I’m pretty sure; the stories DC tells aren’t subtle.

Apologies

I know folks in the region and I have standing offers from people whose taste I
trust to introduce me to others whose company they’re sure I’d enjoy. I ignored
those people and offers. Sorry about that, and I have an excuse. This trip was
another workin’ for The Man episode (the man being Uncle Sam). Specifically, I
was there for a seven-hour deposition by attorneys representing Meta. I felt
this to be a Performance, one that mattered and therefore worthy of my full
attention. So I needed to get centered, be in the zone. To walk and think.

Don’t ask me how the deposition went, I dunno. When you’re speaking at a
conference or rally or whatever, if the audience laughs at your jokes you know
it’s probably going well. But you don’t throw jokes at defendants’ lawyers.

That’s all I’m gonna say. Once again, I hope there’ll be lurid rear-view blog
fodder once this is over.

The Wharf

This trip, like the last, took me to DC’s Southwest Waterfront district. Only
this time, I visited the waterfront. Here’s the view looking out.

Or, I should properly say, The Wharf, as reflected below.

It’s a walkway along the Potomac, several blocks long, backed by a solid wall of
restaurants and bars. The first evening I went down there, it was a
freakishly-warm 24°C and this Canadian enjoyed the hell out of a
restaurant-patio dinner. An extremely expensive dinner. I shouldn’t complain,
because I was also happy walking along the river’s edge, and I made two
different stranger couples happy by offering to take pictures of them in front
of one waterfront vista or another. Few smiles are cheaper.

So, I recommend the Wharf. But, it’s really expensive. Which highlights a
problem I’m starting to have with Washington DC. Like my hometown, there’s far,
far too much money there.

Bad vibe

After I flew in, pretty late one evening, I hit the hotel bar for pizza and
beer. Pretty weird: Two big TVs, one Fox and one CNN. A certain class of
American White Male Person, generally bulgy and prosperous, vibing more on the
Fox side, talking (I eavesdropped) about football in Atlanta suburbs and being
the agent for George Lazenby and how they’d had to grovel before a powerful
woman named Barbara because they’d done something stupid and their career was
over unless Barbara absolved them. They all had Good Relationships with
Important People who could Make Things Happen. If they hadn’t wanted to be
eavesdropped they wouldn’t have talked so loud.

Museums

The day before the Big Event I decided to tourist — there are all those museums,
plus I wanted to get physically tired so that I’d sleep well. I started at the
Hirschhorn because I love ambitious contemporary art showplaces. Granted, when I
visit them, usually at least half the work seems weak or irrelevant or deranged.
I don’t care, I admire the ambition, a contemporary artist has to try going
somewhere nobody’s ever gone before and of course the failure rate is high.

These places often have big rooms featuring artists trying to do big things,
visually or intellectually or both. It’s just a special kind of space, and when
I leave a big contemporary gallery, the outside world seems sepia-toned, free of
sharp mental edges.

None of the current exhibits really shook my grip on reality, which is what good
contemporary art is supposed to do, but I enjoyed my visit; here are a couple of
snaps.

Above: Cupboard by Simone Leigh.

Below: From Wheel of Life by Jessica Diamond.

I hit the National Archive museum: Disappointing, unless you regard America’s
founding documents as sacred texts. In and of themselves, they’re not much to
look at.

Finally, the Museum of the American Indian. I found the collection a little
thin, albeit with a clear-eyed view of all the lies and thefts and betrayals.
But, that name…

The city

Washington is obviously kind of a company town and these days, most Federal
departments only ask people to show their faces one day a week, usually Tuesday.
It also suffers from the urban-American sin of being car-optimized, shot through
with broad rivers of asphalt. Which, except for Tuesdays, are now pretty well
empty. You can cross one of these six-lane behemoths more or less anywhere, any
time. I understand the Mayor is furiously lobbying the Feds to arm-twist the
civil servants back into the offices but it doesn’t look like it’s working.

The National Mall

It’s the big grass strip between the Capitol and the Washington Monument, dotted
with Smithsonia and, on a chilly damp weekend day, hosting several flag-football
games, well-organized and offering flashes of real talent. Heart-warming, I
thought. Also a temporary white structure with music coming out of it and a sign
saying “David’s Tent”. Hey, random free music is an opportunity and the door was
open. It turned out to be a Christian worship joint and there was this frowning
dude alone on stage strumming and singing: “holy, holy, holy” over and over. I
only stayed inside for a minute or two and that was a lot of holy holies. There
were two people in the front row of the front section, and as for the rest…

They need a better product manager.

Eventually I ended up on the Capitol steps, the same ones the Trumpers stormed
that January 6th. Lots and lots of those sectional fences that failed so badly
were still queued up there. I hoisted an end of one with one arm, no strain. I’m
not sure what they’re for, actually.

Anyhow, here’s the view down the Mall from there. You might want to enlarge it.

This picture started out sort of dreamy and abstract, then I bashed the hell out
of it with Lightroom and Silver Efex. I claim it looks like what I saw, only
more so.

Wandering back

To my hotel I mean. The route went by an endless cavalcade of Congressional
office buildings. Here’s one: check out the flag deployments, I guess we know
where those particular congresspeople stand.

By this time, I’d touristed for enough hours that my feet were pretty sore and I
was finding the serried office-block ranks sort of oppressive. Also, I was damn
hungry, and then my life was saved by My Own Pizza, entirely unpretentious,
selling good pies for a good price. All the customers were tourists with kids or
local blue-collar workers, which in DC means Black. Way better vibe than my
hotel bar.

On the last trudge of the day, this:

Asleep on a vent that I guess offers warm air. It’s raining.

These buildings host the wealthiest organization that has ever existed on this
planet.


MORE
February 30
22.6.2024, 19:00Platformonomics (Feed)
Platformonomics TGIF #37: January 26, 2024

Platformonomics TGIF is a weekly roll-up of links, comments on those links, and
perhaps a little too much tugging on my favorite threads.

Get Updates By Email

Subscribe

You no doubt are in final preparations for next week. I certainly am.

News Private Equity in Action: Media Division

Private equity relentlessly demonstrates, despite all their claims to the
contrary, they are not very good operators of the businesses they buy. I’m sure
their financial modeling is impeccable, but their business and product strategy
intuitions are laughable (“hey, we can lay off even more employees and just use
ChatGPT for our media business”).

We’ve said it before, and will no doubt say it many times again, but when
private equity comes amalgamating, it is time to start migrating. You don’t want
to rely on anything private equity is running into the ground.

Previous:

Private Equity in Action: VMware (Continued), Private Equity in Action: VMware,
Private Equity Will Buy and Ruin More Software Companies, Software Migration
Alert: VMware, Software Migration Alert: VMware (Continued), Perfidious Private
Equity

This Week in Securities Fraud

As we prepare for the single biggest week of the year (Cloud CAPEX Week, coming
next week)!, lets cleanse our palates by checking in on one of our favorite
CAPEX clowns (i.e. non-spenders of CAPEX).

IBM beat (their own) earnings guidance and posted barn-burner revenue growth of
2.2% for 2023 (inflation was 3.4%). The market went wild and IBM stock is at a
ten year high (though still short of my generational call to sell IBM at $209
because they were missing the cloud transition. They have underperformed every
benchmark since).

IBM is running their cloud computing playbook again for AI. They talk up AI
without making any real investment, and hope investors can’t distinguish them
from real companies that are actually investing. It worked for them in the last
24 hours, but as with cloud computing, it is hard to see their slippery words
competing in the long term against hundreds of billions of dollars of real
CAPEX.

(Boy does the Twitter clown emoji suck).

IBM’s total CAPEX and CAPEX as a percentage of revenue are at record lows. Where
are IBM’s GPUs? Selling busloads of offshore consultants who implement other
companies’ AI technology is probably not what investors think they’re getting
with “IBM the AI company”.

Previous:

Follow the CAPEX: Keeping Up With The Clowns, IBM’s Lost Decade, IBM
Misadventures

Useful Idiots, Fellow Travelers and Unregistered Foreign Agents: Wall Street

The Chinese Communist Party has decreed henceforth improved investor confidence
in the Chinese economy, a revival of animal spirits, and a bullish outlook for
the Chinese stock market. (Communist cadres, admittedly, may lack a nuanced
appreciation of how authoritarian edicts affect a market already suffering from
too much authoritarianism).

The CCP’s bestest allies in the whole wide world, Wall Street, are of course
rallying in support of the party proclamation.

Previous:

Useful Idiots, Fellow Travelers and Unregistered Foreign Agents: Apple, Xi
Jinping Thought on Wall Street Running Dogs, The Perils of Working for a Cult of
Personality, Press Releases We’d Like to See, Best Nickname for Emmanuel Macron,
Winnie-the-Pooh: Lover of Honey, Champion of Freedom, Missing the (Bamboo)
Forest for the (Apple) Trees

With Twitter engagement circling the drain, please comment here!


MORE
March 60
24.6.2024, 6:00SEI Blog (Feed)
The Top 10 Skills CISOs Need in 2024
This post outlines the top 10 skills that CISOs need in 2024 and beyond.

MORE
January 310
10 Benefits and 10 Challenges of Applying Large Language Models to DoD Software
Acquisition
This post presents 10 benefits and 10 challenges of applying LLMs to the
software acquisition process and suggests specific use cases where generative AI
can provide value.

MORE
January 220
Today, 17:55ongoing by Tim Bray (Feed)
Lofree Flow

Back in 2008, I complained here about Apple keyboards. Apple still hasn’t
addressed my complaint (hard to believe, I know). So, 15 years later, as in late
2023, I picked up a Lofree “Flow”, which I gather had a Kickstarter and
advertises itself as “The smoothest mechanical keyboard”. I’ve been using it
long enough to offer opinions.

The problem with Apple keyboards is illustrated by the picture below; The Lofree
is in the middle.

They’re all wireless. While the small Apple product at the bottom is really old,
today’s equivalent has the same problems: There are no Home, End, PgUp, and PgDn
keys, and the arrow keys are annoyingly small. The larger Apple at the top
wastes space on the accountants’ number pad.

The LoFree is small enough and has the keys I want.

Why?

There are several places where I work regularly, and none of them feature a
large desk, to the extent that if I have one or more of coffee, printed
document, camera, or whatever else on the desk, things get crowded. So small is
good.

Also, I use the navigation keys but never the number grid.

Also, those last few years at AWS a few of the engineers near where I sat had
fancy keyboards with multiple colors and would argue about key-feel options.

Now, I generally like Apple keyboards and managed to avoid buying any of the
infamous “butterfly” models. But there was this rave review in The Verge and I
was trying to avoid some boring work that I was supposed to do, and there was a
Black Friday sale, so here I am, my fingers now laying down a mellow rhythm
section behind the creation of the prose melody you are now reading.

What’s good

The feel and the sound. I got the tactile “Phantom” keys rather than the linear
“Ghost” option, so I can’t report on the latter, but I don’t play twitchy games
and the consensus seems to be that tactile is good for a fast touch typist. I’m
not as blindingly fast as I used to be but I’m still north of 100wpm. It may be
relevant that I have big fat farmer’s fingers and have always hit keyboards
pretty hard.

I’m struggling a bit for vocabulary to describe how this thing feels. I’ll
settle for “eager” — the keys feel like they’re working with me to inject these
blogwords into the noösphere.

As for the sound, it’s just delicious; low but distinct, and to my ears,
euphonious.

Obvious problems

You might as well work with the thing plugged in, because the battery isn’t big
enough and (perhaps to compensate) the Bluetooth goes to sleep quickly and (when
unplugged) wakes up slowly.

The backlighting is lousy; it illuminates the edges of the keys adequately, but
the letterforms only weakly. (Which for a touch typist is usually not a
problem.) The “ambient” lighting is ridiculous; think of the flames on Wayne and
Garth’s Mirthmobile. And if you have the lights turned up it runs the battery
down even faster.

I should mention that the enthusiast community is annoyed that they have trouble
(on some systems) customizing the keyboard layout, and that the selection of key
options is limited. My modestly-customized Mac keyboard layout just worked.

Mobility

My biggest problem is that switching back and forth between the Lofree and an
Apple keyboards makes my fingers clumsy for enough minutes to be annoying. Easy
enough, just carry the Lofree around. It’s not big and despite the
metal-not-plastic frame, is acceptably light. Except for, there isn’t a carrying
case on offer. Which feels like a real miss.

Success?

Oh, definitely. Typing on it makes me happy. None of the problems interfere much
given the way I use it. I guess the nav buttons are a little small but my
fingers are starting to know where they are. Those big fat arrow keys are great.
No, I’m not going to tumble over the edge onto the slippery slope of keyboard
obsession. But this first step turned out fine.


MORE
February 50
22.6.2024, 19:00Platformonomics (Feed)
Platformonomics TGIF #36: January 19, 2024

Platformonomics TGIF is a weekly roll-up of links, comments on those links, and
perhaps a little too much tugging on my favorite threads.

Get Updates By Email

Subscribe

Things still seem a little slow early in 2024. Not sure how much of that is the
Davos reality distortion black hole and how much is people (wisely) saving their
energy for the upcoming annual CAPEX numbers.

News Private Equity in Action: VMware (Continued)

VMware has become a page-by-page read-along of the private equity playbook.
There is no investment in future features or products; the franchise products
shall be milked. Always remember: “When private equity comes amalgamating, it is
time to start migrating”.

Previous:

Private Equity in Action: VMware, Private Equity Will Buy and Ruin More Software
Companies, Software Migration Alert: VMware, Software Migration Alert: VMware
(Continued), Perfidious Private Equity

Whatever Happened to the AI Safety People?

Will calls for airstrikes soon be extended to every AGI-capable script kiddie?
The magic words “open source” trump safety-ism? Where has the hysteria gone?

Previous:

Everyone In Silicon Valley Wants To Be Henry Kissinger, Existential Risk Now Has
a Precise Threshold, You Can’t Spell Alien Invasion Without A and I, Existential
AI Risk + Twitter-level Nuance

Fiddling While Rome Burns: Robotic Vacuum Cleaner Edition

Despite facing actual (as opposed to the AI safety-ism kind) existential
threats, the EU sees a future selling wine, cheese, and robotic vacuum cleaners.

Previous:

The EU Will Continue to Fiddle While Rome Burns

Fiddling While Rome Builds: AI Regulatory Speaking Points Edition

If you chant “competitive advantage” enough, it will actually become so, right?
Magic PR words trump actual regulations?

Previous:

The EU Will Continue to Fiddle While Rome Burns

So You Want to Build an AI Company

The blue part seems like a good place to start. There is a similar chart for
education.

With Twitter engagement circling the drain, please comment here!


MORE
March 60
Today, 17:55ongoing by Tim Bray (Feed)
Mourning Google

On March 15, 2010, I started a new job at Google. The fourteen years since that
day feel like a century. The title of my announcement was Now A No-Evil Zone
and, OK, I can hear the laughing from ten timezones away. I tried, then, to be
restrained, but there are hardly words to describe how happy and excited I was.
I had escaped from the accretion disk the former Sun Microsystems was forming
around Oracle, that blackest of holes. And Google, in 2010, was the coolest
place in the world to work.

Let me quote myself from a little bit further into that piece, on the subject of
Google: “I’m sure that tendrils of stupidity and evil are even now finding
interstitial breeding grounds whence they will emerge to cause grief.” Well,
yeah.

This is in my mind these days as I’m on a retired-Googlers mailing list where
the current round of layoffs is under discussion and, well, it really seems like
the joy has well and truly departed the Googleplex.

But they did.

(The Alphabet “Investor Relations” page

is also its home page.)

It’s not just Google

The last two decades of my career featured the arcing then crashing of popular
regard for Big Tech. It’s hard to believe now, the years when those lovably
nerdy Bay Area kids were leading humanity to a brighter, better-lit future; our
leaders were lionized and when people found out you actually worked for Google,
their eyes widened and you could feel the focus.

These days, Big Tech features in hostile congressional hearings, mass layoffs,
and messy antitrust litigation. It offers few experiences that can be
uncritically enjoyed.

While I was inside the Rooms Where It Happened, it was actually pretty hard to
notice the public trust in our work auguring into the mountainside of alienation
and cynicism. It’s not that I think the companies are the problem, it’s the
machineries and imperatives of Late Capitalism, which for a while we foolishly
thought Internet companies could route around.

“Ten blue links”

I remember the dismissive phase well: Ten blue links was boring, it was the
past, it was not what people wanted. They want answers to their questions,
complete and correct, so much more wholesome than an abbreviated sampling of the
General Internet Uproar. And that was partly right: When I type in “-12C in F”
or “population of vietnam” I just want a number.

But those Ten Blue Links surfaced by the PageRank-that-was had a special magic.
I found them intensely human, a reflection of the voices populating what remains
of the Web, the only platform without a vendor. This was true when I was there
and I said so, but was laughed at.

And now, in Anno Domini 2024, Google has lost its edge in search. There are
plenty of things it can’t find. There are compelling alternatives. To me this
feels like a big inflection point, because around the stumbling feet of the Big
Tech dinosaurs, the Web’s mammals, agile and flexible, still scurry. They
exhibit creative energy and strongly-flavored voices, and those voices still
sometimes find and reinforce each other without being sock puppets of
shareholder-value-focused private empires.

Psychopaths

For my money, that was the center of Google’s problem. Larry and Sergey were
smart guys who recognized they didn’t know shit about corporateness and quickly
got into a pattern of hiring and empowering psychotic pricks who were presumably
“good at business”. Not gonna talk about some of the things I saw because these
people are wealthy and litigious.

But I do have a question.

What to use?

Among Google products, I mean. These days, when I use Google Search or Chrome or
Maps I just don’t feel like they’re on my side. And maybe that’s not
unreasonable; after all, I’m not paying for them. Problem is, the best
alternatives aren’t obvious.

For now, here’s the direction I think I’m going: Use Chrome for Google stuff:
Maps, Calendar, Docs, Translate. Safari and Firefox for non-Google stuff; they
ain’t perfect but I think they’re better aligned with my interests.

Our family company is still on Google Workspace or whatever it is they call
Dasher these days: Mail, Contacts, Photos, Calendar, Meet. It’s OK. We pay for
it and the price is sane. I don’t feel like it’s looking for ways to monetize
each keystroke. I’d totally consider a less-scary alternative.

I fear the combination of Google Maps and Reviews because it stinks of monopoly.
But I use Maps anyhow in my car via Android Auto because it’s nicely integrated
with YouTube Music (which I like) and Google Calendar. For a while I used the
Here.com maps and liked them a lot. I guess I could listen to YouTube over
Bluetooth.

Did I mention Android? I can’t stop using it, because I used to work in that
building and because I decline to use iOS; If I wrote code for it I might not be
able to give it away. And I carry Pixel phones, because I love the cameras.
Having said that, hearing Andy Rubin’s name still makes my gut clench.

I love YouTube because I end most evenings, after everyone’s gone to bed, with a
live musical performance by someone wonderful. But enshittification is creeping
in at the edges.

That cafe

In 2012 I moved from Android to Google’s Identity group. It happened to be in
the same buildings as Google+, at a time when Google was definitely putting all
its wood behind that arrow. Larry and Sergey’s offices were there too (not a
coincidence). There was a major fringe benefit: Access to the Cloud Café.

It was ethereal — OK, pretentious — almost beyond belief. Almost entirely
vegetarian, rare plants hand-gathered by Zen monks and assembled into jewel-like
little platelets-full that probably strengthened eleven different biochemical
subsystems just by existing. And the desserts were beyond divine. Admittedly,
sometimes when I left, my Norwegian-farmer metabolism grumbled a bit about not
having had any proper food, but still.

It was wonderful. It was absurd. And I got a $90K bonus that year because
Google+ hit its numbers.

It’s over, I think. It’s OK to miss it.


MORE
January 190
24.6.2024, 6:00SEI Blog (Feed)
The Latest Work from the SEI
This SEI Blog post summarizes some recent publications from the SEI in the areas
of supply chain risk management, technical debt, large language models, quantum
computing, acquisition, and trustworthiness in AI systems.

MORE
January 150
22.6.2024, 19:00Platformonomics (Feed)
Platformonomics TGIF #35: January 12, 2024

Platformonomics TGIF is a weekly roll-up of links, comments on those links, and
perhaps a little too much tugging on my favorite threads.

SUBSCRIBE

Get Updates By Email

Subscribe

Slowish week. No doubt everyone is busy preparing for the release of annual
CAPEX numbers in a couple weeks.

News Private Equity in Action: VMware

First layoffs. Then price increases. Now scaling back the long tail of customers
to “600 customers who will struggle to change suppliers“. This is the private
equity playbook and why we keep chanting “when private equity comes
amalgamating, it is time to start migrating”.

Previous:

Software Migration Alert: VMware, Software Migration Alert: VMware (Continued),
Perfidious Private Equity

How Many Models? Better Late Than Never

I don’t want to live in a world where someone else is making the world a better
place. I want to live in a world where someone else is making my jokes come
true.

Previous:

Press Releases We’d Like to See: Iceland Embraces the Bitcoin Economy

Everyone In Silicon Valley Wants To Be Henry Kissinger

It’s not just a job, it’s existential!

I’m sure they won’t get played on this…

Previous:

Existential Risk Now Has a Precise Threshold, You Can’t Spell Alien Invasion
Without A and I, Existential AI Risk + Twitter-level Nuance

The European China Shock

The “China Shock” had a huge impact on US manufacturing with major economic,
political and social consequences. The German auto industry, which is the
cornerstone of the European economy (along with wine, cheese and megalomaniacal
regulators), is next on the receiving end, with similar consequences likely.

Previous:

The EU Will Continue to Fiddle While Rome Burns, The EU’s Reckoning Approaches,
Collapsing Economic Models: German Edition, EU Rethinking Its Stance on
Mercantilism, German Auto Industry Reaches Hail Mary Stage, How Do You Say
“Burning Platform” in German?, Volkswagen’s Death Throes Continue, The Fate of
the European Economy: Automotive Edition, Europe: Investing in Slave Labor

With Twitter engagement circling the drain, please comment here!


MORE
March 60
24.6.2024, 6:00SEI Blog (Feed)
The Top 10 Blog Posts of 2023
Every January on the SEI Blog, we present the 10 most-visited posts of the
previous year. This year’s top 10 highlights our work in zero trust, large
language models, DevSecOps, and artificial intelligence.

MORE
January 90
22.6.2024, 19:00Platformonomics (Feed)
Platformonomics TGIF #34: January 5, 2024

Platformonomics TGIF is a weekly roll-up of links, comments on those links, and
perhaps a little too much tugging on my favorite threads.

Get Updates By Email

Subscribe

We’re back and Happy New Year! We’ll start the year with some predictions about
what will (and will not) happen in 2024 (at least as it will be chronicled
here).

We Will Talk About AI

Incessantly.

We will fixate on the trajectory of generative AI. Will frontier models continue
to improve or plateau? Almost every question of how the industry unfolds is
downstream of that dynamic.

Companies that don’t lead in AI but feel like it is their birthright to lead
based on good work decades ago will beg to differ (often using early-in-the-race
metaphors or blah-blah-blah “enterprise” blah-blah-blah).

More than a few investors still see AI as a revolutionary new technology cycle
rebooting everything that lets them invest like its 2021 (due diligence, schmoo
diligence). We will question their judgment, especially those who securitize
GPUs

The megalomaniacally self-important, Luddities and the media (but then I repeat
myself) will continue to frame generative AI as a scary existential issue, while
the self-appointed hall monitor industrial complex continues their power grab.

AI discourse will continue to be framed through the lens of dystopian science
fiction movies (which is an odd choice for “guess the next word” technology).

Whatever the hype, you can’t spell CAPEX without AI AI is a CAPEX-driven
phenomenon, so following the CAPEX means following the GPUs.

Cleanup on the SaaS Aisle Will Continue

We still have too many SaaS companies on the shelves. The reduction in inventory
will continue. We just have to hope there are more cleanup participants than
private equity.

Private Equity Will Buy and Ruin More Software Companies

As Larry Ellison (perhaps apocryphally) says, every ecosystem needs a scavenger.
After years of excess, the software industry is in dire need of a thorough
scavenging. In theory this is what private equity says they do: fix broken
and/or overcapitalized businesses. And private equity loves software.

The model is straightforward: buy a software company using a lot of debt, slash
expenses (engineers are expensive!), jack up prices and then hope to flip it
before customers can migrate.

The problem is there is an ocean of private equity money (~$5 trillion) chasing
too few relevant deals, and, more problematically, private equity doesn’t
actually know how to fix businesses, especially software businesses. The list of
software that got better when private equity got involved is a very, very short
list (let me know if you can think of any). Private equity will no doubt get the
money, but customers end up holding the bag.

Hence our slogan “when private equity comes amalgamating, it is time to start
migrating” (rhyming remains difficult).

VMware is quickly becoming the poster child for private equity buyouts. Layoffs.
Price increases. And more private equity software deals are coming. Alteryx.
DocuSign. We’ll be issuing software migration alerts as private equity deals
happen.

The EU Will Continue to Fiddle While Rome Burns

Even as the complete bankruptcy of their economic, energy and security
strategies has been revealed to all over the last two years. the EU still dreams
of being a “regulatory superpower”. Infused with cookie consent banner
confidence, the EU has turned its attention to AI.

Despite prioritizing speed over comprehension, the EU have failed in their
stated goal to “move fast and regulate things”. China, the UK and US all beat
the EU across the regulatory finish line and now even the French (pause and let
that sink in) think they’ve gone too far.

Her defence of the AI Act comes after Macron argued the legislation risks
leaving European tech companies lagging behind those based in the US and China.

“We can decide to regulate much faster and much stronger than our major
competitors,” the French leader said earlier this month. “But we will regulate
things that we will no longer produce or invent. This is never a good idea.”

Those comments were seen as potentially setting up a new fight over the EU’s new
regulatory regime for AI which is considered among the strictest to be proposed
anywhere in the world.

Though the AI Act was agreed this month, the law still needs to be ratified by
member states in the coming weeks. France, alongside Germany and Italy, are in
early discussions about seeking alterations or preventing the law from being
passed.

(Remember, it is only dirigisme if you’re in France. Anywhere else it is just
sparkling industrial policy).

Meanwhile, the EU’s economic model is collapsing. The German auto industry, the
engine of the EU economy, is in dire trouble. Everybody loves mercantilism until
you run into a bigger mercantalist (hello China!). Deindustrialization is deeply
corrosive. Populism just getting started in Europe. Brexit may turn out to be
the geopolitical analogue of the Accenture partners who got kicked out of
Anderson Consulting just before Anderson got the death penalty for enabling
Enron.

A continent can’t live solely on fines of American tech companies.

Copyright Holders Will Advance Self-Serving Theories

The New York Times wants to replace Disney as the face of copyright maximalism.
So we’ll learn about fair use, the copyright clause of the Constitution (and why
the EU is yet again irrelevant in setting technology norms), the underlying
social contract behind copyright, rent-seeking, try to find any actual and
material damages, why we all stand on the shoulders of giants, and how a few
sands of grain on a beach tend to overstate their importance.

The 6600 tokens I’ve contributed to the LLM training corpus are my infinitesimal
contribution to an immeasurably valuable canon of human thought and expression
(and my best shot at immortality).

Debt, Deglobalization, Demographics and Despotism Will Dominate

Technology has had a great run on center stage, but there are other forces at
play. I’m thinking of calling them the Four Dorsemen, but obviously am open to
other suggestions given that is a terrible term.

Things That Will Not Be

Some things won’t change: cloud repatriation, “supercloud” and metaverse real
estate still won’t be things in 2024.

With Twitter engagement circling the drain, please comment here!


MORE
January 60
Today, 17:55ongoing by Tim Bray (Feed)
Five Electric-Jag Years

I got a note from Jaguar advising that my free five-year “InControl Remote and
Protect” subscription was expiring and would be $99/year (Canadian) going
forward. That’s right, this month is five years since I picked up our 2019
Jaguar I-Pace and joined the EV tribe. Thus this (final?) visit to the Jaguar
Diary series.

$99, you say?!

Yeah. What you get is a mobile app (and Web site) that does useful things,
including:

Turn on the climate control to warm or chill the car while you’re having
breakfast. This is a big deal in Canada.

Make it honk so you can find it in a big parkade.

Check whether you remembered to lock it, and do so if you didn’t.

Keep a diary of all your trips, which is nice and I guess super-useful if you
expense your driving.

Since several of these require that Jaguar operate servers (hey, on AWS in
Dublin) I guess I shouldn’t mind paying. Feels a little pricey but hey, I bought
a Jag so I shouldn’t care? And I wouldn’t, except for they’re almost certainly
harvesting data from the car like crazy and selling it. Which I call
double-dipping.

What about the car?

I have not regretted buying it for a single second of those five years. It’s
comfy, faster than strictly necessary, reliable, practical, and cheap to run,
even with that extra $99/year. Go read the other diary entries for that stuff. I
guess the only thing to add is that, five years in, it still feels pretty well
new. It’s fun to drive. The battery seems to hold about the same number of
kilometres.

But #WarOnCars!

These days, my social-media feed contains many people who point out that Cars
Are Bad and the spaces humans live in should be optimized for humans, and you
Really Shouldn’t Like Cars.

And I agree, mostly. I totally want to choke cars out of the spaces we live in,
to create a fifteen-minute city. Simultaneously, I like cars for their own sake,
for their engineering and aesthetics, for the joy of being able to go from my
front door to anywhere else on the continent while seated comfortably, listening
to good music.

Yes, those viewpoints are in conflict but so what. People, statistically, also
like alcohol and nicotine and speeding and adultery and Reality TV and junk
food. And can agree, abstractly, that indeed, those things are bad. It all comes
down to protect me from what I want.

There are two problems: First, the entire western edge of North America was
built around sprawl and highways. Second, Late Capitalism’s egregious inequality
has arranged that it’s hard for most people to afford to live near their jobs,
or even where the public transit is good.

So yeah, as we develop our cities, let’s exclude cars to the extent possible.
And let’s do something about the economy too. I will vote for regulations that
restrict my use of my car. And to the #WarOnCars troops: I’m on your side, but
don’t tell me I shouldn’t like them.

To start with, here’s a free idea: There should be a charge for occupying city
space with your car, and it should depend heavily on the car’s size and weight.

EV trends

I suspect that central Vancouver has one of North America’s highest EV
densities. On our block, our side of the street, are a Nissan Leaf, a Tesla, a
Hyundai Kona EV, and me. And there’s frequently a Rivian parked out front, don’t
know where it lives. In fact I’m starting to see a lot of Rivians. If you need a
truck, the Rivian looks like a nice one, but I don’t think people in this
neighborhood do.

When I bought the Jag I was worried it was like buying a PC in the Nineties;
wait six months and there’ll be something way better. I got to test that
hypothesis because a good friend just picked up a Hyundai Ioniq 6, 2023 World
Car of the Year. We went for a drive and yeah, the state of the art has
improved.

The Ioniq is, and feels, lighter than the Jag. It charges a lot faster. Its
Android Auto implementation is nicer. The dashboard graphics are gracefully
space-age. The shifter is clever and intuitive. No, it doesn’t have the Jag’s
raw power or silky way through turns. But on balance, it is really a whole lot
of car for the money. I’d buy one.

The most important electric vehicles aren’t going to be personal automobiles.
They’re going to be the buses and trains that eventually come to fill in the
public-transit grid and turn cars into rarely-needed luxury options. They’re
going to be the trucks that are currently a huge source of carbon loading.

Cars will be with us for a while. But they should be lighter and smaller and
fewer. And electric.


MORE
January 60
Happy New Year!

Customarily, on this day we go for a walk by the sea. This year “we” was just
me, because Post-Covid. I have pictures; subdued pictures, it was that sort of
day. Herewith a few of those, and year-end ramblings on optimism, AI/ML,
cameras, and social media.

Optimism?

The climate catastrophe is gonna get worse before it starts getting better. But
I see rays of light that might illuminate 2024. I really don’t think the
Americans are going to elect That Guy again. I think unemployment will stay low
and worker power will increase correspondingly. I think there’s a significant
chance we get a vaccine that actually stops Covid transmission, as opposed to
today’s, which mostly just moderate its effects (still important of course). I
think the health problems in my immediate family will improve a bit — Lauren is
showing early signs of recovery from Post-Covid.

Did I mention the climate catastrophe? I hope our political leaders come to
their senses, get out from under the Carbon Hegemony, and do the necessary
things that will necessarily bankrupt much of the Petroleum sector. If they
don’t, I think it near-inevitable that some of those defending the planet’s
future will discard their commitment to non-violence. There is nothing people
won’t do to protect their children.

Weirdly, and assuming that our species’ self-inflicted climate-disaster injuries
aren’t fatal, there is an upside. This situation falsifies the central premise
of Late Capitalism: That continued unending growth is essential, or even
desirable. Particularly in the inevitable case where declining birthrates become
pervasively global and there are fewer people each year needing goods and
services.

Put another way: Jeff Bezos can blow off “Day 2” as much as he wants. But the
Day-1 growth-at-all-costs dogma isn’t sustainable or even survivable. Day 2 is
inevitable, and we need to learn how to make it good.

AI/ML

I’ve spent my working life in software, and AI seems to be the only thing anyone
wants to talk about. I’m not smart enough to know where this tech ends up
fitting in. Also, I’m arrogant: I don’t think anyone else is smart enough
either. It’s painfully obvious that we’re in the midst of a bubble; just watch
the VC froth. A lot of that money is going to the same place as the billions
they gave to the crypto-bros.

I do have a recommendation: Assuming you’re busy and have a lot of input, don’t
waste time reading “future of AI” predictions. Nobody knows.

Photography

When I went down to the sea, I took a Fujifilm X-Cam with the Samyang 135mm/F2
bolted on, and of course my Pixel. All but one of these pictures are
Fuji/Samyang. This is my chance to opine, once again, that the best way to use a
camera that’s not a phone is to strap a difficult and opinionated lens on it
then follow where it leads.

I’m not that optimistic about the future of “real” cameras. When you watch the
reviews from passionate camera-philes like Chris and Jordan over at Petapixel,
you realize that, to the extent that newer cameras are better, the improvement
is at the margins; for example, shooting elite athletes at a distance. All
modern cameras take great pictures, most times. This notably includes the one in
your phone; but its lens is the opposite of difficult.

2023 saw two real steps forward in camera technology: C2PA and global shutter.
Global shutter is cool but of interest to only a few, and C2PA’s impact is
strictly on the structure of belief; the technology itself is boring. Neither
will help you get a better cat picture.

Social media

Most of my 2023 hours were filled by family health issues, my expert-witness gig
with Uncle Sam, and what comes after Twitter. On the latter, I know only one
thing for sure: That privately-owned centralized social media has not worked,
will not work, can not work. Decentralized federation is the only sane path
forward and we are right now making that up as we go along. Am I crazy to think
that few things matter more than the forces that shape the broader human
conversation?

I know that if you’ve been reading me at all, you’ve heard this enough, but
forgive me, it’s too late to stop now: I think the member-owned social-media
co-op we’re building at CoSocial offers a plausible glimpse of a resilient, fun,
billionaire-proof social-media future. I’m an old guy but I’m as excited as a
kid about this path forward.

This space

2024 will be this blog’s 21st year of operation. No month has had fewer than
three pieces and my hope is to do as well or better going forward. Not writing
feels like not breathing.

During the latter half of 2023, I produced two large legal submissions in
connection with my Uncle-Sam gig, well over 400 pages in total. They are stuffed
full of confidential information and will be read by only a small handful of
people, then forgotten. That makes me sad, but producing them was still fun.

Back in 2018, I wrote enough Song of the Day pieces to fill a half-year’s days,
and enjoyed it a lot. Most of the material was old; I was mining my own personal
musical journey. Nothing wrong with that.

But these last few years, I’ve been listening to lots of new music and, once
Uncle Sam has settled down, I’ll play a variation: Music of the Day. Most of it
will be new-ish or at least new to me. And up front, I confess that most of that
is driven by YouTube Music guessing what Tim might like. So don’t let anyone
tell you I’m against AI in the general case.

Happy 2024, and good luck!

We’re going to need it.


MORE
January 30
Planes, Cameras, Capitals

Since Covid started I haven’t traveled, except for a short hops to visit Mom in
Saskatchewan. But I spent Monday and Tuesday in Washington DC workin’ for The
Man (the man being Uncle Sam) and came away with notes on planes and hotels and
cameras and people.

First, camera news

The rest of this piece is miscellaneous travel notes, aimed at people who are
interested in Washington DC or the travel experience. But you might want to skim
through it anyhow for the pictures. As usual, I traveled with a camera bag
containing an excellent Fujifilm and a variety of lenses. I never broke it out.
The results from the Pixel 7 are just unreasonably good.

See the moon? This picture should hardly even be possible. It was dusk, the
contrast was stark, the moon fetching but faint, and the sky very dark. It only
took a little Lightrooming to prettify the sky and sharpen a bit.

DC

I’ve always liked DC and this time I was in a new neighborhood, Southwest
Waterfront, which is pretty nice even with too many large Federal agencies.

A residential street in Southwest Waterfront, near the alley in the previous
picture. It has a bit of a fakey “HDR” look but that’s actually what it looked
like, the traditional streetlights work well with the stone frontage. I posted
these two pictures, and another nearby, on Mastodon as a quiz, challenging
people to identify the city. One frighteningly-smart person did.

There are cities that feel unwelcoming, but DC isn’t one of them. On the way
from the airport to my hotel via (excellent!) public transit I got lost twice,
consulted a taxi dispatcher and random cop for help. I had a shopping mission
where I needed guidance from big-store staff. I ate a couple of solo dinners,
tucked into the corners of large-ish establishments at unpopular times to
minimize plague risk. Nobody made me feel like I was bothering them, everyone
was happy to help, and in the empty-ish restaurants, staff dropped by to just
shoot the shit a bit.

100% of the people I’m talking about were Black. Just a coincidence?

By the way, I heartily recommend Walter’s Sports Bar — I figured that since I
was in the Nation’s Capital, the most American possible thing to have would be a
burger in a sports bar. The place is comfy, the food is fine, and, as noted, the
people are nice.

What happened?

On this sort of trip, I’d normally write up the payload — the meat of the
meetings and the shape of the work. But it is, as they say, “the subject of
current litigation.” I promise, assuming I remain of sound mind after this is
over, to share as much as I legally can. It should entertain.

Plague protection

Given that each episode of Covid incurs a 5-10% chance of progress to some
flavor of Long Covid, which my wife is already suffering from, and that the top
medical recommendation for Long Covid sufferers is “Don’t get Covid again”, I
was extra ultra paranoid about the possibility of infection. So I wore a KN95 in
all public indoor spaces, to the extent possible, which means except when I was
actually putting food or drink in my mouth. As I write this, I won’t know for a
few more days whether or not it worked.

Other people? not so much. The entire population has decided to “put Covid
behind them” I guess. I can’t begin to understand how they see the trade-offs. I
don’t think I’m crazy and it’d be ungenerous to think that everyone else is.

Enviro-sinner

I can’t write about this without addressing the environmental issues. A handy
online flight-emissions calculator suggests this little jaunt worsened the
global CO2 problem by 1.4 metric tons. Ouch.

I could point out that this level of sin is a tiny pale shadow of what I
committed in my decades of jet-setting on behalf of one employer or another, or
that my trip, in support of presumably-worthy high-stakes litigation, was of
benefit to the species. Doesn’t really make it OK. I am now lobbying The Man to
make future meetings virtual rather than physical.

Planes

Meh. Like many things in life, the experience on offer has declined over the
decades, as the “business class” experience has simultaneously improved; a
symptom of our society’s current insanely-dysfunctional level of inequality.
Since I have a vast inventory of points left over from my jet-set phase, I
upgraded both ways. Shoot me. Or, shoot out the window.

These pictures are miraculous, shot through a not-terribly-clear airplane window
with backing light trying to reflect off it. This is the first time the
“computational photography” power of a Pixel has scared me a bit, because the
picture actually looks quite a bit better than what my eyes saw. And that was
before I juiced it up with Lightroom. Hmmm.

I simultaneously miss travel and have enjoyed the past few years of not fighting
airlines and airports; my long-held belief is that this is an industry that
hates its customers. I have reserved a special level of hate for Toronto’s
Terminal 1.

Thus, you can imagine my surprise when my rollie busted a wheel in DC and I
decided to check it through to Vancouver, and they just checked it through! No
need to do the recover-and-recheck in Toronto. And also, you can come in from
the US and go out again on domestic without passing through security. I guess I
must tip my hat to my old enemy.

Oh, and (noted reluctantly) more good news. The video system on one leg of the
flight was Android-based (telltale boot screen) and wow, it was miles better
than any previous in-flight entertainment experience. I knew about Android Auto,
but I guess there’s now Android Airplane. So the pilots can say “OK Google, land
on runway 23”? I hope it doesn’t need a reliable USB-C connection.

Postmodern hotel

How can a hotel be interesting? This chain I’d never heard of that I’m not gonna
name had rooms right across the street from work and the reviews were good and
the price was good. I’m not naming them because I think their concept is sound
but they haven’t nailed the execution yet. The idea is that the rooms are
teeny-tiny (like I’ve experienced in Japan) but on the ground level there’s this
huge sprawling “Living Room” with lots of desks and sofas and tables and a big
TV lounge and a bar with decent basic food. The decor is breathlessly hip
everywhere.

I watched Monday Night Football there, enjoying the company of a
Mexican-American family who were crushed when Seattle pulled off that absurd
last-minute win over Philadelphia; tears were shed on Jalen-Hurts-branded
jerseys.

Also, excellent breakfast.

I think they might be onto something, except for I had a hard time sleeping
because the duvet was too thick and the pillows were too thick, so I was cooking
and my neck was hurting.

I filled out the feedback form and got an email from a real person, so maybe
it’ll get better. I mean, the core competence of a hotel has to be getting you a
good night’s sleep, so this isn’t an excusable miss. Anyhow, it’s years since
I’ve stayed at a hotel that wasn’t a boring beige cookie-cutter.

Travel!

It’s what humans do; go read Bruce Chatwin. I’ve missed it. But, to be enjoyed
in moderation.


MORE
December 230
Today, 9:19mark nottingham (Feed)
RFC 9518 - What Can Internet Standards Do About Centralisation?

RFC 9518: Centralization, Decentralization, and Internet Standards has been
published after more than two years of review, discussion, and revision.

It’s no secret that most people have been increasingly concerned about Internet
centralization over the last decade or so. Having one party (or a small number
of them) with a choke hold over any important part of the Internet is counter to
its nature: as a ‘network of networks’, the Internet is about fostering
relationships between peers, not allowing power to accrue to a few.

As I’ve discussed previously, Internet standards bodies like the IETF and W3C
can be seen as a kind of regulator, in that they constrain the behaviour of
others. So it’s natural to wonder whether they can help avoid or mitigate
Internet centralization.

I started drafting a document that explored these issues when I was a member of
the Internet Architecture Board. That eventually became
draft-nottingham-avoiding-internet-centralization, which became an Independent
Stream RFC today.

But it was a long journey. I started this work optimistic that standards could
make a difference, in part because Internet standards bodies are (among many
things) communities of people who are deeply invested in the success of the
Internet, with a set of shared end user-focused values.

That optimism was quickly tempered. After digging into the mechanisms that we
have available, the way that the markets work, and the incentives on the various
actors, it became apparent that it was unrealistic to expect that standards
documents – which of course don’t have any intrinsic power or authority if no
one implements them – are up to the task of controlling centralization.

Furthermore, centralization is inherently difficult to eradicate: while you can
reduce or remove some forms of it, it has a habit of popping up elsewhere.

That doesn’t mean that standards bodies should ignore centralization, or that
there isn’t anything they can do to improve the state of the world regarding it
(the RFC explores several); rather, that we should not expect standards to be
sufficient to effectively address it on their own.

You can read the RFC for the full details. It covers what centralization is, how
it can be both beneficial and harmful, the decentralization strategies we
typically use to control it, and finally what Internet standards bodies can do.

One final note: I’d be much less satisfied with the result if I hadn’t had the
excellent reviews that Eliot Lear (the Independent Submissions Editor) sourced
from Geoff Huston and Milton Mueller. Many thanks to them and everyone else who
contributed.


MORE
May 20
24.6.2024, 6:00SEI Blog (Feed)
Acquisition Archetypes Seen in the Wild, DevSecOps Edition: Clinging to the Old
Ways
This SEI blog post draws on SEI experiences conducting independent technical
assessments to examine problems common to disparate acquisition programs. It
also provides recommendations for recovering from these problems and preventing
them from recurring.

MORE
December 180
Today, 17:55ongoing by Tim Bray (Feed)
Recent Music Books

Seems that for the past few months most of the books I’ve read have been sci-fi
(which I wrote up here) or about music. Herewith notes on four of the latter.
The subjects are Philip Glass, John Cale, Cuba, and getting loaded.

Glass

Words Without Music is Philip Glass’s autobiography. I should start by saying
that I’m a big fan, bought lots of albums back in the day. For anyone unfamiliar
with the flavor, I recommend the following: Open up your favorite streaming
service and say “Play Mishima by Philip Glass”. You’ll know pretty quick whether
you like it. If you do, you won’t be alone; I don’t have hard data but I think
Phil is probably the best-selling practitioner of “New Music”, i.e. contemporary
stuff that isn’t Pop music.

I caught a live concert too, Phil and the Philip Glass Ensemble, in which he
plays but doesn’t lead; (Early editions of the Ensemble included Steve Reich on
keys.) They performed the Koyaanisqatsi sound-track while the movie played. What
a sound!

Any Glass fan is going to want to read this, I think. But it’s not problem-free.
First: He talks tons about his process and his high-level thinking about putting
his music together, but zero about its music-theory basis. By the way, one
reason he’s so prolific is that he doesn’t have to write a Master Score as he
composes: He just holds the music in his head and writes out the individual
parts, one by one. I saw a documentary once which showed him writing musical
lines at a comfortable handwriting speed, saying with an amused expression “This
part, you see, this part goes pretty fast.”

Second, it’s a depressing reminder of how deeply the Reagan-Thatcher dogma has
savaged the fabric of our society. Glass was a bright kid in a middle-class
family who went from an elite public school to an elite college with
Nobel-Laureate teachers, then to an elite music school, then to Paris to study
at the feet of Nadia Boulanger. The Paris section, by the way, is an astonishing
read; Whatever you think of his music, the man studied his craft with heroic
intensity. Then he settled in New York and by, moving furniture and driving
taxi, earned enough to rent a loft and compose furiously, turning the world of
“New Music” inside out. Try to pull that off today. Everything’s been
financialized and efficiency-maximized and there’s little space left for variant
shoots of any art form to thrive then twist them sideways.

Oh, another irritant: Most of the book is written in an extremely transparent
flat-aspect style, which gets out of the way and I respect, and I was thinking
“good solid prose, it’s a pity Phil doesn’t try to reach back and bring it a
little bit, like he does in his music.” And then in the very last chapter, he
does. Here’s a paragraph, relating his reaction when asked about what it felt
like when composing one of his big pieces:

I don’t know … Because I’m not sure that I am there at that moment. The ordinary
witness has been lost  — the artist Philip has robbed the daily Philip of his
ability to see himself. That’s very clearly what happens when people say “I
wrote it in a dream,” or “I don’t know where the music came from. … All they’re
really saying is “I don’t remember how I did it,” and they make up an outside
source. But the real source is not any of those things. It’s a process that the
artist has learned. He has tricked himself into gaining that extra attention
that he needed to do the work.

I kinda wished that Phil had let loose some of that writing energy on more parts
of the book. Whatever, it’s a valuable piece of the historical record.

Cale

John Cale, a Welshman, is a founding member of the Velvet Underground and one of
life’s Really Interesting People. Disclosure: His album Sabotage/Live is central
to the way I think of music: Live is better. Loud is better. Terrifying is best.
I stage-managed one of the concerts on the tour that album showcases; that show
went severely off the rails in a way that gives me a distant echo of PTSD all
these decades later. I may write about it some year.

Anyhow, I’m here to write about What’s Welsh For Zen, a sort of autobiography,
in large format with plenty of photography and art splashed across all the
pages. I wanted to read it and discovered that it’s only available used and for
hundreds of dollars. So I took it out of the public library for free; isn’t it
great to be a member of a civilization?

Anyhow, Cale was a member of the same white-hot New York art scene that Phil
Glass was, only with Lou Reed and Andy Warhol and lots of drugs. Boy, does John
ever hate Lou. But he invests a lot of words in explaining what kind of a person
Lou is and what it’s like to work with him. Or try to. Whatever, they produced
some pretty fucking wonderful music together. Nobody ever said making art had to
be fun.

Cale’s portrayal of Warhol is way kinder. What I notice, looking back all these
decades, is that everyone I’ve read writing about Warhol seems to be describing
a different person. Warhol managed the Underground for a while and it’s pretty
clear that some of the ideas he brought to their oeuvre, in particular
performance practices, have become integral to popular music at large.

Anyhow, Cale plays a significant but not leading role in the story of how Rock
music became what it is today. I’m glad that he put all this stuff on the
record.

Getting loaded

In Vancouver’s library system you go online, you find the book you want, you put
it on hold, and after a while you get an email telling you it’s arrived at your
local branch. Whenever I bike over there I go look at the featured-books shelves
and occasionally pick one up.

When I was picking up the Phil Glass book I noticed Too Late To Stop Now: More
Rock’n’Roll War Stories by Allan Jones, who spent decades in the pop-music
journalism profession, back when you could make a decent living in journalism of
many flavors. This is, um, extremely lightweight, mostly tales of the author
ingesting massive quantities of alcohol and cocaine and weed while hanging out
with Rock Stars you might have heard of. Parts of it work well for an oldster
like me because we remember the rage and excitement in the air when the New Wave
swept away Prog Rock. Boy, does Mr Jones hate anything even remotely
prog-related and, to be fair, those geezers kinda brought it on themselves.
Anyhow, what saves it is that he’s consistently funny.

There’s a connection: One of the segments covers an extended conversation with
John Cale, which re-iterates the awfulness of working with Lou Reed, but in this
matter Cale comes off better through a third-party pen than his own. Anyhow, if
you were listening to music in the late Seventies this has a good chance of
going deep on some artist you really cared about. And it might open your ears to
something new; in my case, to the work of Roy Harper, whom I’d heard of but not
actually heard, and now I have and am glad of it.

Cuba

What happened was, I was watching a YouTube of a live concert by Rhiannon
Giddens, which I recommend doing because she is by the way totally a goddess.
She tends to chat at the audience a bit between songs, and in connection with
something I totally forget, she recommended Cuba and Its Music: From the First
Drums to the Mambo, by Ned Sublette. I thought it sounded interesting. During
the depths of Covid’s first wave, I took a few months of private Afro-Cuban
music lessons via Zoom; bought a nice pair of Congas, even. I recommend this for
anyone who’s even a bit musically literate and interested in rhythm. Just
learning about the Clave rhythm and Bell pattern, and a bit of son and rumba
lore, changed the way I think about rhythms. If you’re ever doing rhythm in a
jam, throwing in a bit of clave feel will make everything magically funkier.

So, I recommend this book, but… wow, it’s huge. It’s “only” 688 pages, the
listing says (I read it in the Kindle app on my tablet) but it feels twice as
long; took me weeks to finish. It is monumentally thorough; by the time you’re
finished, you will be damn erudite on the history and culture of Cuba from the
Pre-Columbian era through to the Fifties. It’s advertised as being “Part 1” but
was published in 2004, so I’m not sure about Part 2.

The Spanish organized-crime mob — I’m talking about Columbus and his
successors — pretty well wiped out Cuba’s Indigenous population (although I
learned in writing this that there are people who to this day claim to be Taíno
and preserve that culture). Thus, there remain four input streams that
intertwined to create Cuba’s remarkably-rich mix of language and religion and
especially music:

Spanish, which, given the dates, included strains of…

Moorish, being the still-strong influence of Al-Andalus.

Then there were the enslaved Africans; they were not just one people, but from a
variety of nations, and brought along with them two main currents of language
and musical culture:

The first, from West Africa’s southern coast — think of the range from Côte
d'Ivoire to Benin, which was itself complex, including flavors from the
coastline and then up into the Sahel, where once again there was a Muslim/Arabic
influence, and last but definitely not least,

The Congo, a general term here for southern Africa‘s eastern shore: Equatorial
Guinea down to Angola.

In Cuba, the Church was less efficient in its proselytizing than elsewhere in
the New World, and thus the African religions, and even bits and pieces of the
languages, survive and have a hefty influence on some of Cuba’s musics. Yes,
that’s musics, plural; it’s not just one thing at all.

Anyhow, to get the beginnings of a feel for the feel, type “Arsenio Rodriguez”
or “Tito Puente” into your favorite music source and see what happens.

Am I ever happy to have read this book, and if I don’t stop now, this blog piece
will start to inherit its punishing length and complexity. I’ve probably said
enough for you to make a good guess whether it’s for you.

Meta-music

I already have a couple of shelves of books on music, a high proportion being
biographical, and reading these outings reminded me of why. Books on music is a
habit I recommend.


MORE
December 130
24.6.2024, 6:00SEI Blog (Feed)
Applying Generative AI to Software Engineering: Navigating Ethical and
Educational Landscapes
The post explores the application of generative AI in software engineering with
an emphasis on the ethical and educational landscapes.

MORE
December 120
Today, 17:55ongoing by Tim Bray (Feed)
Recent Sci-Fi

Seems that for the past few months most of the books I’ve read have been sci-fi
or about music (which I wrote up here). Herewith notes on the most recent
sci-fi/fantasy. Featuring Emily Tesh, Martha Wells, John Scalzi, Sandra Newman,
Vajra Chandrasekera, Arkady Martine, and P. Djèlí Clark.

Used to be I’d put Amazon links in when I mentioned books, and route a few
commission bucks into my Amazon account. No longer. While I’m not a fanatic, if
I can do something using a Big Tech or an alternative, I’m going to be trying
the alternative. Alternatives are good.

Twisty space opera

Some Desperate Glory is the debut from Emily Tesh and I expect great things from
her, because this is pretty great. Let’s see, we have an earnest young trainee
at a fascist outer-space military academy, but her Faith In The Mission is
slipping. Wait… I’m hearing echoes from plenty of old-and-not-that-great sci-fi
pulp.

And yeah, the first 40% of the book, while it’s sharply-written and keeps you
turning pages, is treading a pretty well-worn path. Then it goes sideways. More
than once. I mean the sideways paths are reasonably well-trodden too but still
surprising and clever and entertaining.

I can’t help noticing that there’s all this good space opera written by
lesbians? Not obvious from first principles why that should be, but OK.

Murderbot and…

I suspect everyone knows about Murderbot now. The latest is System Collapse and
if you’re going to read it, which I recommend, it’s probably a good idea to to
go back and re-read Network Effect first; the new book is an extremely immediate
sequel and it will probably help to refresh your mind on who all these people
and bots are. System Collapse is a little shorter and lighter in weight, which
is OK. If you like Murderbot you’ll like it fine.

Then there’s Ms Wells’ Witch King, a huge, complicated saga of epic conflict
between somewhat-divine characters. And they’re all interesting enough
characters. But I don’t know, they just didn’t grab me the way the Murderbot
cast does. Witch King is really well-crafted. The combat scenes are explosive.
I’ll have to give another of Ms Wells’ non-Murderbot works a try, but this one
didn’t work that well for me.

Scalzi fun

During the course of the last year I read The Kaiju Preservation Society and
Starter Villain, both by John Scalzi. These are blasts of pure fun, with
razor-sharp dialog, jaw-dropping set pieces, and lots of laughs. Nothing here
will change your life but I’m pretty sure you won’t regret dropping a few bucks
on either. If only for the dolphins with the attitude problem.

1984++

Just like every other member of the world’s Anglophone population, 1984 was
compulsory one of my high-school years and yeah, one doesn’t forget it, admires
it even, but I wonder if anyone really likes it? Hardly a pleasant experience.
Anyhow, Julia, by Sandra Newman, builds around Winston Smith’s story from his
lover’s point of view. It’s a bigger book than the original and you won’t forget
it either, whether you actually like it that much. It’s compelling, and the
perspective Ms Newman gets in backing off a bit from the hyperfocus on Winston’s
world, in considering Airstrip One and Oceania a little more widely, is entirely
convincing and pushes your mind in different directions.

Julia is a more interesting person than Winston and her experience is more
intensely felt. I liked parts of it. I won’t read it again. It attaches a sort
of coda to the story after its natural ending that didn’t really work for me,
but I can’t think of what I’d replace it with. If 1984 made an impression on
you, this will quite likely leave you with one of equal or greater strength.

Bright doors

Vajra Chandrasekera is from Colombo, Sri Lanka. In the admirable Jo Walton’s
September 2023 Reading List she got all excited about Chandrasekera’s The Saint
of Bright Doors. She says “Everything is slightly too big and too bright, and
details keep piling up and slipping out of control, and it’s all stirred
together with a dash of Kafka—but in a good way.” Can’t disagree. Fetter, who
casts no shadow, leaves his sorcerous murderous mother to find his fortune in
the Big City…

It’s as colorful as anything I’ve read in years. I really enjoyed that, although
the cleverly-constructed world is perhaps a bit more interesting than the story
inhabiting it. It’s, uh, not space opera and not romance and not fantasy and not
1984 but its author has clearly drunk from all those wells.

More lesbian space opera!

A Memory Called Empire, by Arkady Martine, won the 2020 Best-Novel Hugo so it’s
hardly obscure. I liked it, and its sequel A Desolation Called Peace (Hugo
2022), a whole lot. It’s got a Galactic Empire and, like most space operas set
in one of those, you get to meet the Galactic Emperor and his family and court.
In fact, the first book spends most of its time right there. Which is fine,
because our plucky heroine Mahit Dzmare and the people she gets to know are
plenty of fun and there’s plenty of action, and an Aztec undercurrent.

The first book is slightly lacking in space battles and aliens and so on, but
the second solves both those problems in the first few pages, and it’s a good
space battle and they’re excellent aliens. And then romance, and spicy sex, and
sordid politics, and, well, not boring at all. It’s called the Teixcalaan Empire
series and I’m pretty sure there will be more.

I enjoyed Ms Martine’s work enough that I also picked up Rose/House, which is
shorter, and beautifully constructed, and defies description. There’s this house
in the desert, there’s a corpse in it, and the house is really hard to get into.
I think it’s fair to say it’s a stronger piece of work than the imperial space
operas. But I’ll still totally snap up the next Teixcalaan-Empire book.

Dealing with Djinn

When I looked up P. Djèlí Clark after reading A Master of Djinn, I discovered
that he’s super-political; issues of class and race and imperialism are all over
his Web site — he’s a Black American. They’re all over the book too, now that I
think about it, but so smoothly and well-worked-in that I never noticed while
reading it.

The book is a detective story, with Djinn, set in “Egypt in the 1920s”; I put
that in quotes because the 1920’s weren’t like that, and neither is Egypt.
Anyhow, it’s action-packed; maybe a little over-written, there are places where
it could do with a little less atmosphere and a little more dialog and action.
But the Djinn are great and so’s the heroine and it’s all very satisfying in the
end.

Books are good

I find it very supportive to my currently-somewhat-assailed mental health to
drag my eyes away from the global clusterfuck surrounding us and sit in a soft
chair looking on words on pages. You might too?


MORE
December 130
22.6.2024, 19:00Platformonomics (Feed)
Platformonomics TGIF #33: December 8, 2023

Platformonomics TGIF is a weekly roll-up of links, comments on those links, and
perhaps a little too much tugging on my favorite threads.

This is the last issue of the year. May your new year be filled with
non-performative competence, pluralism, reason, universalism and vigor (a
contrarian hope to be sure).

News AI Marketing Hallucinations: AWS Edition

With generative AI, Amazon finds themselves on the wrong end of their favorite
expression that “there is no compression algorithm for experience”. (And the
evidence is growing that there is a ton of work that goes into model quality
after the training is done, and sees benefits from scale, which makes it ever
harder for laggards to catch up).

Amazon not only over-promised and under-delivered with their generative AI
efforts at re:invent, they even managed to raise the expectations bar to claim
only AWS could deliver some new nirvana of enterprise LLM capabilities. The
immediate and egregious failure to do so badly hurts their credibility, not just
for their insatiable desire to be seen as the generative AI leader, but also
across all the areas where they actually do lead.

Previous:

The AWS Generative AI Soap Opera

AI Marketing Hallucinations: Google Edition

Google seems to have taken the AWS re:invent marketing approach as a challenge
instead of a cautionary tale. Unlike Amazon, Google has largely managed to keep
their generative AI insecurities to themselves (and as the original inventor of
the transformer with a two hundred billion dollar advertising business, they
have a lot more to be insecure about).

But as we hit a year since Google’s CEO may or may not have declared “Code Red”
in response to ChatGPT (we did learn Google definitively has “Code Yellow“),
they clearly felt compelled to do something. They announced their Gemini model,
accompanied by two sneaky marketing moves where they got caught. One was a faked
demo of their multi-modal capabilities and the other was comparing GPT-4
performance to the Gemini Ultra model which isn’t available.

Previous:

The Google Pivot Away from “AI-Safety First”, “Google has more than 160 people
working full time“

What Comes After Titanium?

The Apple reseller (aka the phone companies) TV ads are amusing as the only
thing they can find to say about the latest iPhones is “Titanium”.

Year End Reflections On Some Favorite Threads Cloud CAPEX – I will update the
numbers after the clouds report in January, but CAPEX is all about the GPUs
right now. And on the clown CAPEX front, it is almost time to call a winner in
the clown car race. Cloud Repatriation – still not a thing. Sorry VCs who picked
this as an investing theme. “SuperCloud” – has hit its two-year anniversary of
not being a thing. Congratulations!!!

Previous:

Follow the CAPEX: Cloud Table Stakes 2022 Retrospective, Follow the CAPEX:
Keeping Up With The Clowns, Platformonomics Repatriation Index, “Supercloud” RIP

AI Alliance: Wah-Wah, Wah-Wah

Beyond the edgy white-on-black press release, what exactly should we expect from
this august alliance beyond platitudes? The fact IBM is still pretending Red Hat
is somehow an independent entity participating on their own accord suggests a
greater focus on numbers than substance.

Software Migration Alert: VMware (Continued)

When private equity comes amalgamating, it is time to start migrating. Why?
Because private equity ruins software companies.

VMware CEO Raghu Raghuram has left the building. He will be missed. So too
apparently has the security talent, which also will be missed. As I’ve
previously observed, private equity really struggles to model cybersecurity in
their spreadsheets:

There is one area where private equity-owned software companies are showing
uncharacteristic leadership: they’re at the epicenter of some of the largest
cybersecurity breaches.

Previous:

Software Migration Alert: VMware, Perfidious Private Equity

McKinsey Karma?

The John Oliver episode doesn’t cover all of McKinsey’s misdeeds, but does an
excellent (and entertaining) job on the ones it does cover.

Previous:

2020 Platformonomics Blog Retrospective

Related:

McKinsey: Last Week Tonight with John Oliver


MORE
December 90
24.6.2024, 6:00SEI Blog (Feed)
Creating a Large Language Model Application Using Gradio
This post explains how to build a large language model across three primary use
cases: basic question-and-answer, question-and-answer over documents, and
document summarization.

MORE
December 40
22.6.2024, 19:00Platformonomics (Feed)
Platformonomics TGIF #32: December 1, 2023

Welcome to the New and Improved Platformonomics.

Subscribe to get this weekly newsletter and other posts delivered via email.

The OpenAI Soap Opera

Good thing I was off last week and didn’t add to all the effervescent interim
takes on this saga. What a clown show.

My guess is season 2 of everyone’s favorite new hit show, Governance, will focus
on Anthropic. The dramatic tension rests on whether they still want to be the
doomer-i-est of AI companies (they left OpenAI seeing it as completely cavalier
about AI risks), or if the OpenAI board absurdity cured them of that.

Previous: Culture Clash Conjecture: Amazon and Anthropic

The AI Doomer Soap Opera

Even Nick Bostrom, the father of the paperclip maximizer, “is worrying he may
have yelled “Terminator!” in a crowded theater.”

Previous: Why Not Cut Out the Middleman and Just Name AI Safety Rules After
Dystopian Science Fiction Movies?, Existential Risk Now Has a Precise Threshold,
You Can’t Spell Alien Invasion Without A and I, Existential AI Risk +
Twitter-level Nuance

The AWS Generative AI Soap Opera

ChatGPT launched a year ago on November 30th, right in the middle of AWS’s
annual re:invent developer conference. AWS has been talking incessantly about
generative AI ever since, but because their release cycle is wrapped around
re:invent, we’ve been waiting for this last week to see if their bits match
their words. The short answer is no, and re:invent was another episode of
talking way too much about generative AI relative to what was delivered. The
claim to be “reinventing generative AI” with a bunch of late and me-too
announcements just hurts their credibility. We got a lot of “three-tier
strategy” arm-waving (note there is no actual foundation model tier) and
glorification of fondue and cuckoo clocks (i.e. ye old Switzerland strategy).
I’m not alone in being underwhelmed:

It is totally fine to say you’re just going to be an infrastructure provider for
generative AI (especially if that is what you’re doing). But claims to
leadership in all things when you’re manifestly not competitive (or even
present) don’t help your cause. And the more you talk about generative AI, the
more you distract from your palpable plumbing prowess, where you are the leader
and what the vast majority of your customers care most about. In other words,
don’t overindex on your weaknesses, yet here we are.

There were also a bunch of IBM vibes, between the repeated reminders of
long-past glories and vague claims that “only we can provide enterprise
capabilities”, like IBM made about cloud computing (except for the having an
actual cloud part).

In terms of the announcements, the Q “spokesmodel” looks to be a RAG chatbot of
AWS documentation, presumably on Anthropic (and the q about the NYT story on Q
being for general information workers is whether that was Amazon exaggeration or
just the Times’ traditional struggles with understanding technology). Titan
Image Generator I suspect is Stable Diffusion under the hood. Guardrails is the
AWS version of the Copilot stack. And still no sign of a competitive LLM of
their own (and it gets harder and harder to catch up with every passing day).
But most of all, the “three steps into a 10k race” metaphor seems to have
finally been deprecated, for which we are all thankful.

Previous: re:Invent or irrelevant?, Scuttling Titan(ic), Scaling Olympus?,
Anthropic Adds Billions For More Existential AI Hand-wringing, Culture Clash
Conjecture: Amazon and Anthropic, Titan or Titanic?, The Amazon AI Inferiority
Complex

Instability AI

Previous: Stability AI: Not So Stable, Seeing Post Facto Red Flags Ex Ante

Private Equity Shot: VMware Private Equity Chaser: VMware

Broadcom isn’t a private equity firm, but they’re absolutely running the private
equity playbook. History strongly suggests price increases are next. RIP VMware.

Software Migration Alert: VMware

When private equity comes amalgamating, it is time to start migrating. Why?
Because private equity ruins software companies.

Previous: Perfidious Private Equity, Software Migration Alert: Avid Technology,
Software Migration Alert: EngageSmart, This Week’s Software Migration Alerts

Software Migration Alert: Rover

When private equity comes amalgamating, it is time to start migrating. Why?
Because software companies go to the dogs under private equity ownership.

“We Have the GPUs!!!”

I don’t know this research firm but interesting if true.

And the claim “server unit shipments for 2023 could crash by up to 20 percent”
isn’t auspicious for Team Repatriation. Nor are Dell (servers and networking
+9%) and HPE’s (“The Compute division, which contains HPE’s traditional server
business, dropped 31% to $2.6 billion, a steeper fall than analysts projected“)
latest results.

Previous: Cloud Repatriation: The Search Continues, Cloud Repatriation: Still
Not a Thing, Platformonomics Repatriation Index – Q1 2023: Surf’s Up?, The
Inexorable Attraction Between Terrible Things: Cloud Repatriation and Private
Equity

Company Killed By Its Customers

Elon wants you to know that eX-Twitter’s problems have absolutely nothing to do
with erratic 3am decisions by a part-time CEO who doesn’t understand the
business for which he drastically overpaid.

Related: Elon Tries to Cheer Up Linda Yaccarino After Her Trainwreck Code
Conference Interview, Says “Hold My Beer” (watch the whole interview with
Jonathan — there is so much more than than Elon telling advertisers to f*ck off)

Technology and the Liberal Arts

Apple’s control freak censorship proclivities make it very hard to be “the new
HBO” and perhaps signal the end of the golden age of television.

Previous: Useful Idiots, Fellow Travelers and Unregistered Foreign Agents: Apple


MORE
December 40
18.6.2024, 16:30All Things Distributed (Feed)
Tech predictions for 2024 and beyond
The coming years will be filled with innovation in areas designed to democratize
access to technology and help us keep up with the increasing pace of every-day
life -- and it starts with Generative AI.

MORE
December 10
Today, 17:55ongoing by Tim Bray (Feed)
Late Light

Late in the year, I mean. We’re hurtling toward the solstice and the photon flow
in the short hours between official sunup and sundown hardly deserves the name
“daylight”. But when the Pacific Northwest cloud and rain and fog let up, the
always-slanting sunlight is very pure.

Which I plan to salute but then disrespect via whimsical photomanipulation. Not
apologizing either.

What happened was, I ran a few neighborhood errands in the rare November clarity
and then there was this tree.

That’s about how it looked.

[Narrator: Enter Lightroom.]

But here’s the thing. That torrent of autumnal yellow is cool and I liked it,
but I thought, well, maybe I could do something with that; good raw material!
Say, turn it down a bit so the color is speaking not shouting.

And yes, I liked that color and thought it freed up the tree’s shape.

And the sky looks nice too but I think also distracting..

The observant reader will have noticed that I was walking round the tree
pointing my phone at it, which all the strangers who noticed smiled and nodded
at, because those leaves.

Anyhow, subtracting color seemed to be going in a good direction, so why not Go
All The Way?

Look at the right edge of the photo, there’s a B&W bird! If you believe shape
and contrast are the most important things about a picture, this one subtracts
everything else.

I think I’ve done enough to that poor tree. Just a little later, I pointed the
lens (by the way everything here is Pixel 7) at the sidewalk just in front of my
feet.

Those would be oak leaves.

Winter sucks and there’s plenty of it incoming. It will however not be entirely
free of light, so keep your eyes open.


MORE
November 300
Today, 14:28Phil Windley's Technometria (Feed)
SSI is the Key to Claiming Ownership in an AI-Enabled World

I've been trying to be intentional about using generative AI for more and more
tasks in my life. For example, the image above is generated by DALL-E. I think
generative AI is going to upend almost everything we do online, and I'm not
alone. One of the places it will have the greatest impact is its use in personal
agents, determining whether or not these agents enable people to lead effective
online lives.

Jamie Smith recently wrote a great article in Customer Futures about the kind of
AI-enabled personal agents we should be building. As Jamie points out: "Digital
identity [is how we] prove who we are to others." This statement is particularly
resonant as we consider not just the role of digital identities in enhancing
personal agents, but also their crucial function in asserting ownership of our
creations in an AI-dominated landscape.

Personal agents, empowered by AI, will be integral to our digital interactions,
managing tasks and providing personalized experiences. As Bill Gates says, AI is
about to completely change how you use computers. The key to the effectiveness
of these personal agents lies in the robust digital identities they leverage.
These identities are not just tools for authentication; they're pivotal in
distinguishing our human-generated creations from those produced by AI.

In creative fields, for instance, the ability to prove ownership of one's work
becomes increasingly vital as AI-generated content proliferates. A strong
digital identity enables creators to unequivocally claim their work, ensuring
that the nuances of human creativity are not lost in the tide of AI efficiency.
Moreover, in sectors like healthcare and finance, where personal agents are
entrusted with sensitive tasks, a trustworthy, robust, self-sovereign identity
ensures that these agents act in harmony with our real-world selves, maintaining
the integrity and privacy of our personal data.

In this AI-centric era, proving authorship through digital identity becomes not
just a matter of pride but a shield against the rising tide of AI-generated
fakes. As artificial intelligence becomes more adept at creating content—from
written articles to artwork—the line between human-generated and AI-generated
creations blurs. A robust, owner-controlled digital identity acts as a bastion,
enabling creators to assert their authorship and differentiate their genuine
work from AI-generated counterparts. This is crucial in combating the
proliferation of deepfakes and other AI-generated misinformation, ensuring the
authenticity of content and safeguarding the integrity of our digital
interactions. In essence, our digital identity becomes a critical tool in
maintaining the authenticity and trustworthiness of the digital ecosystem,
protecting not just our intellectual property but the very fabric of truth in
our digital world.

A good place to start is with our "humanness." On an earlier version of this
post, Timo Hotti commented that the most important use of verifiable credentials
might be to prove that you're human. There are a number of ways to do this.
Proof that I have a driver's license, at least for now, proves that I'm human.
Bank credentials could also be used to prove humanness because of
know-your-customer (KYC) regulations. I suggested something like this a year or
so ago as a way of cutting down on bots on Twitter.

As we embrace this new digital frontier, the focus must not only be on the
convenience and capabilities of AI-driven agents but also on fortifying our
digital identities so that your personal agent is controlled by you. Jamie ends
his post with five key questions that we shouldn't lose sight of:

Who does the digital assistant belong to? How will our personal agents be
funded? What will personal agents do tomorrow that we can’t already do today?
Will my personal agent do things WITH me and FOR me, or TO me? Which brands will
be trusted to offer personal agents?

Your digital identity is your anchor in the digital realm, asserting our
ownership, preserving our uniqueness, and fostering trust in an increasingly
automated world, helping you operationalize your digital relationships. The
future beckons with the promise of AI, but it's our digital identity that will
define our place in it.

Tags: ssi identity verifiable+credentials ai


MORE
November 280
24.6.2024, 6:00SEI Blog (Feed)
Don’t Wait for ROI on Model-Based Analysis for Embedded Computing Resources
This SEI Blog post examines the design and implementation of embedded computing
resources for CPS, the complexities of which drive the need for model building.

MORE
November 280
Today, 9:19mark nottingham (Feed)
How to Run an Australian Web Site in 2024

A while back, the eSafety Commissioner declined to register the proposed
Industry Codes that I’ve previously written about. Now, they’ve announced a set
of Industry Standards that, after a comment period, will likely be law.

If you run an online service that’s accessible to Australians, these Standards
will apply to you. Of course, if you don’t live here, don’t do business here,
and don’t want to come here, you can probably ignore them.

Assuming you do fall into one of those buckets, this post tries to walk through
the implications, as a list of questions you’ll need to ask yourself.

I’m going to try to focus on the practical implications, rather than “showing my
work” by deep-diving into the text of the standards and supporting legislation.
This is based only upon my reading of the documents and a miniscule dollop of
legal education; if there are things that I get wrong, corrections and
suggestions are gladly taken. Note that this is not legal advice, and the
Standards might change before they’re registered.

Does the Standard Apply to Your Service?

The first question to answer is whether your service is covered by the Online
Safety (Designated Internet Services – Class 1A and Class 1B Material) Industry
Standards 2024.

The short answer is “yes, even that one.”

A Designated Internet Service (DIS) is one that allows “end-users to access
material using an Internet carriage service.” This is a very broad definition
that explicitly applies to Web sites. For simplicity, the remainder of this
article will assume your service is a Web site, even though other information
services can be a DIS.

In a nutshell, if “none of the material on the service is accessible to, or
delivered to, one or more end-users in Australia”, your site is exempt.
Otherwise, it’s covered (unless one of the other Codes or Standards takes
precedence; see below).

So whether you’re Elon Musk or you have a personal Web site with no traffic,
this standard applies to you, so long as it’s available to one Australian person
– even if none actually visit. Don’t be fooled by “Industry” in the title. That
default page that your Web server comes up with when your new Linux box boots
for the first time? Covered. Note that it doesn’t even need to be on the public
Internet; things like corporate Intranet sites are covered, as are content-free
static sites like those used to park domains.

Given how broadly the legislation and standard are written, combined with how
prevalent HTTP and similar protocols are on today’s Internet, it’s also
reasonable to say that APIs are covered; there’s no inherent restrictions on
formats or protocols in the eSafety standards – in fact, the definition of
material in the Act includes “data”.

So, to be safe, any server available on the Internet is covered by the eSafety
scheme, so long as it can be accessed by Australians.

Do You Need a Risk Assessment?

Assuming that your site is covered by the Standard, your next step is to figure
out whether you need to perform a risk assessment.

Assuming that you’re not running a large commercial web site, a (ahem) “high
impact” service (i.e., one that specialises in porn, violent content, and
similar), or an AI-flavoured service, there are two interesting categorise that
might get you out of performing a risk assessment.

The first is a “pre-assessed general purpose DIS.” You can qualify for this if
you don’t allow users in Australia to post any material (including comments), or
if posting is “to review or provide information on products, services, or
physical points of interest or locations made available on the service.” It’s
also OK if they are “sharing […] with other end-users for a business,
informational, or government service or support purpose.”1

Does it seem like your site qualifies? Not so fast; that only covers
“pre-assessment.” A general purpose DIS is a

website or application that […] primarily provides information for business,
commerce, charitable, professional, health, reporting news, scientific,
educational, academic research, health, reporting news, scientific, educational,
academic research, government, public service, emergency, or counselling and
support service purposes.

Unless your site falls cleanly into one of those categories, you don’t have a
general purpose DIS.2

The second is an “enterprise DIS.” This is a site where “the account holder […]
is an organisation (and not an individual).” Basically, if your users are
companies or other organisations and not individual people, you don’t have to do
an assessment.

What Does Your Risk Assessment Contain?

Assuming you need a risk assessment (spoiler: you probably do, to be safe), you

 must formulate in writing a plan, and a methodology, for carrying out the
assessment that ensure that the risks mentioned in subsection 8(1) in relation
to the service are accurately assessed.

The risk referred to is that class 1A or class 1B material will be “generated or
accessed by, or distributed by or to, end-users in Australia using the service.”
Storage of such material is also included (even if it isn’t accessed).

To answer your next question, class 1A material is “child sexual exploitation
material”, “pro-terror material”, or “extreme crime and violence material.”
class 1B material is “crime and violence material” and “drug-related material.”
There are long definitions of each of these kinds of material in the standard; I
won’t repeat them here.

Your risk assessment must “undertake a forward-looking analysis” of what’s
likely to change both inside and outside of your service, along with the impact
of those changes. It’s also required to “specify the principle matters to be
taken into account”, including eleven factors such as “the ages of end-users and
likely end-users”, “safety by design guidance”, AI risks, terms of use, and so
forth.

Your risk assessment has to be written down in detail. You must also “ensure
that [it] is carried out by persons with the relevant skills, experience, and
expertise” – although it’s not yet clear what that means in practice or how it
will be enforced.3

What’s Your Risk Profile?

Once you’ve done a risk assessment, you’ll have a risk profile – one of Tier 1,
Tier 2, or Tier 3.

Let’s assume your site has no user-generated content, and you only upload very…
normal… content– like this site.4 You’re likely to be Tier 3.

If so, congratulations! Your work is just about done. Sections 34, 40, and 41 of
the Standard apply to you – basically, the eSafety Commissioner can demand that
you provide them with your risk assessment and how you arrived at it. You also
have to investigate complaints, and keep records.

If you’re not Tier 3 – for example, you blog about drugs or crime, or you allow
user uploads or comments, there are a whole slew of requirements you’ll need to
conform to, which are well out of scope for this blog entry (since I’m mostly
interested in the impact of regulation on small, non-commercial sites). Tip: get
some professional help, quickly.

What Other Standards Will Apply?

Keep in mind that we’ve gone through just one of the proposed Standards above.
The other one is about e-mail and chat services, so if you run a mail server (of
any flavour – maybe even on your infrastructure?), a chat server (e.g., Prosody,
jabberd), or Mastodon server, buckle up.

There are also another set of Industry Codes that cover things like hosting
services, app stores, social media, search engines, and operating systems, if
you happen to provide one of those.

Keep in mind that if you change anything on your site that impacts risk (e.g.,
adding a comment form), you’ll need to re-assess your risk (and likely conform
to new requirements for reporting, etc.).

What Does Enforcement Look Like?

There are a lot of small Internet services out there – there are a lot of IP
addresses and ports, after all. I suspect many people running them will ignore
these requirements – either because they don’t know about them, they think
they’re too small, that the eSafety Commissioner won’t care about their site, or
they’re willing to run the risk.

What is the risk, though?

Section 146 of the Online Safety Act 2021 sets the penalty for not complying
with an Industry Standard at 500 penalty units – currently, AU$156,500 (a bit
more than US$100,000).

In practice, the eSafety Commissioner is unlikely to come after any site if its
content isn’t problematic in their eyes. Whether you want to rely upon that is
up to you. Because the legislation and standard don’t have any exemptions for
small services – even with limited audiences – you are relying upon their
discretion if you don’t have a risk assessment ready for them.

What Do You Really Think?

Improving online safety is an important task that needs more focus from society,
and I’m proud that Australia is trying to improve things in this area. I’m
critical of the eSafety Industry Codes and now Standards not because of their
objective, but because of their unintended side effects.

Both the enabling instrument and this delegated legislation are written without
consideration for the chilling effects and regulatory burden they create on
parties that are arguably not its target. Requiring professional risk assessment
raises costs for everyone, and creates incentives to just use big tech
commercial services, rather than self host – leaning us further into things
being run by a few, big companies.

Moreover, if a small personal site is distributing child porn or inciting
terrorism, they’re not going to be caught because it doesn’t have a properly
considered risk assessment ready to produce on demand – the eSafety Commissioner
already has a range of other powers they can use in that case. They don’t have
the resources to go after the countless small services out there for compliance
issues, so all that will remain is the lingering chilling effects of these
pointless requirements.

I get that most people will ignore these requirements, and the eSafety
Commissioner is presumably relying upon that to give them the leeway to go after
the people they need to target. I just think that creating laws that can be
applied with so much discretion – where technically everyone is in violation,
and the regulator can pick who they prosecute – is a shitty way to run a
democracy.

Is it just me, or is “informational” a hole big enough to drive a truck through
here? ↩

Notably, the site you’re reading this on doesn’t clearly qualify for any of
them, and so when these codes are registered, I’ll likely be doing a risk
assessment (and posting it), even though it doesn’t allow comments any more
(because, spam). ↩

This seems to foretell the establishment of a new industry. ↩

Although it’s always tempting to write a blog entry that depicts, expresses or
otherwise deals with matters of drug misuse or addiction in such a way that the
material offends against the standards of morality, decency and propriety
generally accepted by reasonable adults to the extent that the material should
be classified RC. ↩


MORE
February 240
Today, 14:28Phil Windley's Technometria (Feed)
dApps Are About Control, Not Blockchains

Summary: Decentralized applications, where the identity and application data are
both controlled by the person using the app, provide the means of
disintermediting companies who leverage their privileged position to work for
their own interests and against ours.

I recently read Igor Shadurin's article Dive Into dApps. In it, he defines a
dApp (or decentralized application):

The commonly accepted definition of a dApp is, in short, an application that can
operate autonomously using a distributed ledger system. From Dive Into dApps
Referenced 2023-11-12T15:39:42-0500

I think that definition is too specific to blockchains. Blockchains are an
implementation choice and there are other ways to solve the problem. That said,
if you're looking to create a dApp with a smart contract, then Igor's article is
a nice place to start.

Let's start with the goal and work backwards from there. The goal of a dApp is
to give people control over their apps and the data in them. This is not how the
internet works today. As I wrote in The CompuServe of Things, the web and mobile
apps are almost exclusively built on a model of intervening administrative
authorities. As the operators of hosted apps and controllers of the identity
systems upon which they're founded, the administrators can, for any reason
whatsoever, revoke your rights to the application and any data it contains.
Worse, most use your data for their own purposes, often in ways that are not in
your best interest.

dApps, in contrast, give you control of the data and merely operate against it.
Since they don't host the data, they can run locally, at the edge. Using smart
contracts on a blockchain is one way to do this, but there are others, including
peer-to-peer networks and InterPlanetary File System (IPFS). The point is, to
achieve their goal, dApps need a way to store data that the application can
reliably and securely reference, but that a person, rather than the app
provider, controls. The core requirement for achieving control is that the data
service be run by a provider who is not an intermediary and that the data model
be substitutable. Control requires meaningful choice among a group of
interoperable providers who are substitutable and compete for the trust of their
customers.

I started writing about this idea back in 2012 and called it the Personal Cloud
Application Architecture. At the time the idea of personal clouds had a lot of
traction and a number of supporters. We built a demonstration app called Forever
and later, I based the Fuse connected car application on this idea: let people
control and use the data from their cars without an intermediary. Fuse's
technical success showed the efficacy of the idea at scale. Fuse had a mobile
app and felt like any other connected car application, but underneath the
covers, the architecture gave control of the data to the car's owner. Dave Winer
has also developed applications that use a substitutable backend storage based
on Node.

Regular readers will wonder how I made it this far without mentioning picos.
Forever and Fuse were both based on picos. Picos are designed to be self-hosted
or hosted by providers who are substitutable. I've got a couple of projects
tee'd up for two groups of students this winter that will further extend the
suitability for picos as backends for dApps:

Support for Hosting Picos—the root pico in any instance of the pico engine is
the ancestor of all picos in that engine and thus has ultimate control over
them. To date, we've used the ability to stand up a new engine and control
access to it as the means of providing control for the owner. This project will
allow a hosting provider to easily stand up new instance of the engine and its
root pico. For this to be viable, we'll use the support for peer DIDs my
students built into the engine last year to give owners a peer DID connection to
their root pico on their instance of the engine and thus give them control over
the root pico and all its decedents. Support for Solid Pods—at IIW this past
October, we had a few sessions on how picos could be linked to Solid pods. This
project will marry a pod to each pico that gets created and link their
lifecycles. This, combined with their support for peer DIDs, makes the pico and
its data movable between engines, supporting substitutability.

If I thought I had the bandwidth to support a third group, I'd have them work on
building dApps and an App Store to run on top of this. Making that work has a
few other fun technical challenges. We've done this before. As I said Forever
and Fuse were both essentially dApps. Manifold, a re-creation of SquareTag is a
large dApp for the Internet of Things that supports dApplets (is that a thing?)
for each thing you store in it. What makes it a dApp is that the data is all in
picos that could be hosted anywhere...at least in theory. Making that less
theoretical is the next big step. Bruce Conrad has some ideas around that he
calls the Pico Labs Affiliate Network.

I think the work of supporting dApps and personal control of our data is vitally
important. As I wrote in 2014:

On the Net today we face a choice between freedom and captivity, independence
and dependence. How we build the Internet of Things has far-reaching
consequences for the humans who will use—or be used by—it. Will we push forward,
connecting things using forests of silos that are reminiscent the online
services of the 1980's, or will we learn the lessons of the Internet and build a
true Internet of Things? From The CompuServe of Things
Referenced 2023-11-12T17:15:48-0500

The choice is ours. We can build the world we want to live in.

Tags: dapps picos blockchain pcaa


MORE
November 230
Today, 17:55ongoing by Tim Bray (Feed)
Green and Brown Sauce

Here’s a simple dish that hits the spot as Winter’s chilly dimness comes at us
from all directions. It’s a pasta sauce featuring crispy bacon and braised
greens.

Here are the ingredients: Bacon that I’ve cut up into little squares and sizzled
for twenty or so minutes starting with a bit of olive oil, until they’re that
nice golden-brown shade. Then some braising greens: Kale and chard in this case.
Note: The proportions in the picture are are wrong, there should be more greens,
they always cook down more than I expect. Now, the brown fluid…

You’re going to want some liquid for the braising, and it needs to add flavor.
I’ve seen recipes with wine, broth, beer, and so on. So if there’s nothing in
the fridge you can always make a bit of broth with bouillon. In this case it’s
miso soup, home-made by Lauren because we were too baffed to cook a couple
nights ago and ordered in sushi. Worked great.

Braising

It’s a two-stage process: Brown then simmer. Now, if you’ve got a frying pan
that you’ve just cooked bacon in, you can discard most of the fat then brown the
greens right there. Not for long; a minute is plenty. Then you toss in the
braising fluid, which should only half-cover the veggies.

Timing

The preparatory chopping and crisping takes a while; you want to start working
on this meal 45 minutes before dinner time. In the ingredients picture you can
see that I set the crispy bacon aside.

This evening I was making spaghettini, which only need six or so minutes on the
boil. That’s plenty of time do do the braising. So there’s co-ordination
required, to get the water boiling and the pan simultaneously hot. I toss the
bacon in almost immediately after the broth to share the flavor around, and then
when the pasta’s only got a minute or so to go I scoop some of its water into
the frying pan too, because I read that something proteins something coating
something.

Anyhow, you’re done. Plate the pasta, put the greens and bacon on top, pour some
of the fluid over it.

Bon appetit!

Not writing in this space too much in the moment because I’m workin’ for The Man
and There Is A Deadline. Also playing defense more than I’d like. Hang in there,
folks.


MORE
November 230
24.6.2024, 6:00SEI Blog (Feed)
Explainability in Cybersecurity Data Science
This post provides an overview of explainability in machine learning and
includes illustrations of model-to-human and human-to-model explainability.

MORE
November 200
22.6.2024, 19:00Platformonomics (Feed)
Platformonomics TGIF #31: November 17, 2023

Welcome to the New and Improved Platformonomics.

Subscribe to get this weekly newsletter and other posts delivered via email.

No issue next week due to an anticipated food coma.

GPUs are CAPEX; CAPEX is Increasingly GPUs

John Luttig has an excellent piece entitled “NVIVIA Envy” examining GPU supply
and demand today and into the future. In particular he tries to triangulate
hypercloud investment and says “each player spends ~20-35% of data center capex
on GPU build-outs”. That might be a little high given we’re really only seeing a
CAPEX spike at Microsoft. He suggests the hyperclouds are buying half of all
H100s and Microsoft was likely 29% of NVIDIA’s data center revenue in the most
recent quarter. Like everyone, he mulls the question of when GPU scarcity turns
into a glut.

Previous: Q3 Cloud CAPEX, Q2 Cloud CAPEX, The Google CAPEX Mystery, Follow the
CAPEX: Cloud Table Stakes 2022 Retrospective

re:Invent or irrelevant?

The annual AWS developer conference approaches at the end of this month. Last
year’s re:Invent keynote was pretty tedious, with innovation going increasingly
asymptotic. But the next day, Open AI launched ChatGPT, resetting the industry
focus and irritating AWS to no end (industry leaders hate it when something new
disrupts their comfortable way of life and leadership status — I learned that at
Microsoft in 1995).

This year AWS no doubt will loudly and repeatedly proclaim they are a player in
generative AI. (Maybe they’ll even take their long awaited “fourth step“). The
question is whether they will have code to back up all their words and
demonstrate they’re on the frontier. With Google luring Anthropic back to a
non-exclusive position, Amazon either needs to have made phenomenal progress on
internal development, or we’re going to be back to a lot of arm waving about
being Switzerland, having a three-tier strategy and tired metaphors. Given their
annual release cadence around this event, the pressure is on to deliver
something this month that substantiates their claims to be relevant in the
generative AI stack.

Previous: Scuttling Titan(ic), Scaling Olympus?, Anthropic Adds Billions For
More Existential AI Hand-wringing, Culture Clash Conjecture: Amazon and
Anthropic, Titan or Titanic?, The Amazon AI Inferiority Complex

BREAKING: Depreciation Schedules

NVIDIA announces the H200, a successor to the much coveted H100. Awaiting word
on how this impacts the GPU collaterization market with their assumption H100s
have a six year useful life.

Previous: Depreciation Schedules Meet Moore’s Law, GPUs as an Asset Class, GPUs
Are Better Than Any Financial Asset (For Now), Brother, Where Art Thy GPUs?

BREAKING: EU Considers Regulating Screen Colors

The stigma endures! Once the EU finishes its search engine, cloud computing
service and LLM, maybe they can do a messaging system?

When “Move Fast and Regulate Things” Breaks Down

Being first provided at least some plausible excuse for having no idea what you
were regulating. But clueless and last? Embarrassing.

Previous: AI Regulation: Move Fast and Regulate Things

Tumbleweeds In The Metaverse

JP Morgan shutters its metaverse thingy:

In early 2022, JPMorgan Chase proudly claimed it was the first bank to enter the
metaverse.

Inside a crypto-focused virtual world called Decentraland, users could visit the
swanky Onyx lounge, named after the bank’s blockchain operations, to mingle with
fellow crypto enthusiasts and watch videos of the bank’s experts discussing
crypto markets.

A virtual portrait of a beaming Jamie Dimon hung on the wall just below the Onyx
logo. A digital tiger roamed the floor. At the time, the bank said that the
metaverse could “lead to uniquely new services and business models” and that
“the virtual real estate market could start seeing services much like in the
physical world, including credit, mortgages and rental agreements.”

But in June, just a year and a half after it opened, JPMorgan quietly shut down
the Onyx lounge as its visitor numbers sank toward zero.

I may have missed the window to write up my dream of applying a Georgist tax
regime to the metaverse.

Previous: Things I Don’t Understand: Pricey Metaverse Real Estate, Things I
Still Don’t Understand: Pricey Metaverse Real Estate
Related: The ‘Georgists’ Are Out There, and They Want to Tax Your Land

IBM Touchy About Being Associated With Nazis

Perhaps with good cause.

Related: IBM Screws Employees, Again


MORE
December 40
The New and Improved Platformonomics

Everyone needs their own homestead on the Internet: a place to opine, ponder,
share, expound, sermonize, debate, learn, connect, gather feedback, be found,
pursue passions, do business, and whatever else keeps you out of trouble (or in
it).

But exercising dominion over your personal plot of digital turf requires some
effort, and shouldn’t be outsourced to Big Tech, no matter how easy they make
it.

As we learn over and over again, the digital landlords have their own interests
and will happily sacrifice yours. Whether it is evolving business priorities,
algorithmic changes, moderation policies or just a proclivity for making erratic
decisions at 3am, being an Internet sharecropper brings needless torment. You
must own your own domain and web site as the foundation of your digital
existence.

I’ve been amusing myself here since 2005 and the post-Twitter turmoil is
motivation to double down. I write to think, share, shitpost, and elicit both
feedback and scrutiny of my various theories about the world. I chronicle my
obsessions and compulsively pull threads (often long after the point has been
made). I surf the hype cycle and am prone to allergic convulsions when
confronting BS. That will all continue, even as some other things here change.

Welcome to the latest incarnation of Platformonomics.

New Design

You’ll notice a new design actually optimized for reading. There is less noise –
all the sidebar stuff is gone. And the fonts should no longer hurt your eyes.

The new design supports my weekly Platformonomics TGIF newsletter as well as the
more traditional (but less frequent) longer posts.

Platformonomics TGIF is a weekly roll-up of links, commentary, updates and
occasional snark. The goal is quicker posts that complement my less frequent
essays and allow more timely pulling of those pet threads. A lot of what once
would have been tweeted now shows up in the newsletter.

Get Updates By Email

Subscribe

You can read it here, or subscribe via RSS, ActivityPub or email. I’ll link to
my posts on Twitter until Elon demands money, and cross-post to Blue Sky,
LinkedIn and Threads. But the epicenter is right here.

Moving to WordPress.com

In addition to the new design, I’m also moving to the WordPress.com mothership.
I’ve run this blog using WordPress software for almost thirteen years, but at
another hoster. I’m excited about the boost WordPress.com brings to reliability,
security and scalability (bring it Hacker News!). And a full set of integrated
services means I don’t have to play Tetris with plug-ins.

My previous hoster has been deteriorating on multiple axes. Down time is way up.
Prices are increasing even as quality declines. And what was once high-quality
support has been shipped overseas. It is almost like they were bought by private
equity. Oh, wait, they were. So I’m following my mantra of “When private equity
comes amalgamating, it is time to start migrating”.

I considered Substack, but WordPress can do everything Substack can (sign up to
receive my newsletter!) and much more, which both broadens options and preserves
optionality. Substack has too narrow an aperture with its focus on email
newsletters. I also worry Substack suffers from the classic startup curse of
having been force fed too much capital at too high a valuation from VC investors
too eager for a quick return. Indigestion kills lots of startups.

Thanks to the folks at Automattic for all the help with the migration.

Future Directions

Beyond writing more, my hope is to drive more interaction with all the fabulous
readers of this site. As Twitter continues to deteriorate, I miss the debate,
insights, and people calling me on my malarkey that used to happen there. More
comments here please! I am exploring other ideas to make this site a better
place for discussion and discourse over and above traditional comments.

Any reactions to the new design?


MORE
December 40
24.6.2024, 6:00SEI Blog (Feed)
Generative AI Q&A: Applications in Software Engineering
This post explores the transformative impacts of generative AI on software
engineering as well as its practical implications and adaptability in
mission-critical environments.

MORE
November 160
18.6.2024, 16:30All Things Distributed (Feed)
Standing on the shoulders of giants: Colm on constant work
The Builders' Library gathers the expertise of Amazon's most experience builders
in one place. One article that really got me thinking was Colm MacCárthaigh's
"Reliability, constant work, and a good cup of coffee" which is about an
anti-fragility pattern that he developed for building simple, more robust, and
cost-effective systems.

MORE
November 170
24.6.2024, 6:00SEI Blog (Feed)
The OSATE Slicer: Fast Reachability Query Support for Architectural Models
Our post introduces the OSATE Slicer, a new extension to the Open Source AADL
Tool Environment that adapts a concept called slicing to architectural models of
embedded, critical systems.

MORE
November 130
Today, 17:55ongoing by Tim Bray (Feed)
Global Sensor Bandwidth

The photo-world is all agog over Sony’s just-announced (but not shipping till
next year) high-end ($6K) camera, the ɑ9 III, because it has a “global sensor”.
No, the “global” label didn’t mean anything to me either, when I first read it.
The write-ups about it have explainers and cool pictures (PetaPixel, DPReview).
I found myself wondering “What is this thing’s bandwidth?” and thus this note.
I’ll toss in another little explainer so you don’t have to click on a link like
a savage.

Non-global shutters

A digital camera sensor has millions of pixels arranged on a grid (in
non-obvious and geometrically interesting ways, but let’s not go there); they
are analog devices that measure how many photons hit them. To produce a digital
image the camera runs a bunch of voodoo across the sensors to produce a digital
integer array that can be saved in memory and eventually displayed as a colored
image on a screen.

But wait, how does the camera go about arranging for the photons to hit the
sensor? Well, there can be an actual physical shutter that opens for a very
short time and then closes again, or there can be a purely electronic-mode
“shutter” that turns on the pixels then reads the values off them after enough
time has passed.

But a physical shutter takes nonzero time to traverse the face of the sensor, so
the pixels at the top are not exposed at the same instant as the pixels at the
bottom. (Of course it’s more complicated than that, there are shutter geometries
and other design tricks but let’s not go there.) Which is normally OK, but
suppose you’re taking a picture of something that’s moving fast. Then you can
get what’s called “banding” or “rolling shutter”, usually shows up as unwanted
curvature. There are other problems with synchronizing a flash (but I don’t use
those) and in video mode.

Electronic shutters don’t make this problem go away. The pixels are arranged in
an array (On my Fujifilm X-T30 6240x4160, on my Pixel 7 4080x3072) and are
typically read off about as you’d expect, a row at a time. Which in practice is
like a shutter.

“Global shutter”

You’ve likely already figured it out. These things advertise that they read all
the pixels off the sensor at once. So, no matter how fast your subject is
moving, you’ll get an image of what it really looked like. And those flash and
video problems vanish. And because circuits are faster than shutters, you can
shoot at an eighty thousandth of a second.

All of which probably doesn’t do much for me, I take pictures of oceans and
flowers and trees mostly. But for people who shoot sports or wildlife or in
extreme lighting situations, this is probably a big deal. And there’s no need
for a physical shutter at all; any time you can get rid of a moving part, that’s
a win.

“Simultaneously?”

One result of all this is that the ɑ9 III can take 120 shots/second. At this
point I should mention that it has 24.6M pixels, small by modern high-end-camera
standards. So, first of all I was wondering how you read those data points
“simultaneously”. I’m not a microelectronics whiz but a few jobs ago I learned a
lot about memory controllers and, well, that’s a lot of integers to move all at
once. Then I wondered, what’s the bandwidth at 120 frames/second?

The first question that arises is, how many bytes is 24.6 million pixels?
Starting with, how many bits per pixel? The answer to this is less obvious. My
first assumption was that since the pixels on my screen have 24 bits of RGB
information it’d be three bytes/pixel, but no, each pixel only measures the
dynamic range of one color, then a process called demosaicing produces the RGB
pixels. so I thought maybe just 8 bits/pixel? As with everything else, it’s more
complicated than that; the answer seems to be somewhere between 10 and 16
bits/pixel.

So I scribbled some Ruby code, whose single argument is a guess at the number of
bits per pixel, and computes how many GB/second those 120 shots are. Here’s the
Ruby in case you want to check my arithmetic.

def data_rate(bits_per_pixel) pixels = 24.6 * 10**6 shots_per_second = 120
bits_per_sensor = pixels * bits_per_pixel bytes_per_sensor = bits_per_sensor /
8.0 bandwidth = bytes_per_sensor * shots_per_second end bpp = ARGV[0].to_f bw =
data_rate(bpp) bw_in_g = bw / 10**9 puts "data rate #{bw_in_g}G"

If you trust that Ruby code, at 10 bits/pixel, the camera is moving 3.69GB/sec;
5.90GB/sec at 16. Which I think is a pretty neat trick for a consumer product,
even a high-end one.

The future

It seems likely that global shutters will probably get a lot cheaper and become
a feature of almost every serious camera. Because those circuit designers and
semiconductor-wranglers are really smart people, and you just know they’re going
to find a whole lot of ways to take this v1.0 implementation and make it cheaper
and better.

What’s interesting is, it’s not obvious to me whether or not global shutters wil
be ubiquitous in mobile-phone cameras. They have way more CPU but way less room
inside. We’ll see.

But, you know what, I’d sort of thought that we were in a plateau of excellence
in camera design, wasn’t expecting any really significant new features to show
up. But what with C2PA and now this in the last couple of weeks, it’s pretty
clear I was wrong. Fun times!


MORE
November 140
22.6.2024, 19:00Platformonomics (Feed)
Platformonomics TGIF #30: November 10, 2023

Subscribe now before Elon colocates Starship test launches with Twitter.

Scuttling Titan(ic), Scaling Olympus?

The Amazon Generative AI soap opera just never ends (and, credit to them, they
have acknowledged, more than once, it is a long distance race they have barely
started). Anthropic won’t commit to a monogamous relationship with Amazon. Titan
is looking like the Titanic. So Amazon has a new code name for their
aspirational frontier LLM: Olympus. Presumably we’ll hear all about it at
re:Invent re:GPT.

Previous: Anthropic Adds Billions For More Existential AI Hand-wringing, Culture
Clash Conjecture: Amazon and Anthropic, Titan or Titanic?, The Amazon AI
Inferiority Complex

Looking for GPUs

My favorite part of OpenAI DevDay. Is an hour a day even remotely enough?

Previous: Depreciation Schedules Meet Moore’s Law, GPUs as an Asset Class, GPUs
Are Better Than Any Financial Asset (For Now), Brother, Where Art Thy GPUs?
Related: Dumpster Diving Forgiven for GPUs, Larry Ellison Makes Pilgrimage to
Redmond

Sanctions are Not Easy

Previous: “Lord Make Me Compliant, Just Not Yet!”

New York Times Very Concerned about Hallucinations

Yet they neglected to assess the hallucination rate of their own front page:

Software Migration Alert: Avid Technology

When private equity comes amalgamating, it is time to start migrating. Why?
Because private equity ruins software companies.

Previous: Perfidious Private Equity, Software Migration Alert: EngageSmart, This
Week’s Software Migration Alerts

Understanding the Global Economy

A succinct explanation by Michael Pettis of the world’s most important economic
imbalance (Chinese over-investment/under-consumption).

Previous: Paul Krugman is Wrong Even When He is Right

Doing Your Bit to Rebalance the Global Economy

Time for everyone to help the Chinese rebalance their over-dependence on exports
by buying 10% less stuff from China.

Previous: Xi Jinping Thought on Wall Street Running Dogs, The Perils of Working
for a Cult of Personality, Press Releases We’d Like to See, Best Nickname for
Emmanuel Macron, Winnie-the-Pooh: Lover of Honey, Champion of Freedom, Missing
the (Bamboo) Forest for the (Apple) Trees


MORE
November 110
Today, 17:55ongoing by Tim Bray (Feed)
Recent Griffiths

Back in 2014, in Saint Hilda, I posted a rave review of Nicola Griffith’s novel
Hild. I just finished its sequel Menewood and, in between, read her much-smaller
Spear. These are all extremely good books and if you like well-done portrayals
of violent conflict and royal-court politics and historical flavors and hot sex
in just about any combination, you’re apt to like these.

Hild, again

Yes, I read it again when Menewood was about to drop and boy, did I ever enjoy
doing so. I won’t revisit the details from that 2014 review linked above, just
want to re-emphasize the flavor. It blasted my suspension of disbelief to tiny
fragments, then swept them up and tossed them out. I absolutely believe that
this was what being close to the center of the action in post-Roman Britain
looked and felt and smelled like. Plus Hild and the people around her are, well,
fun to be around.

Menewood

This inhabits the world of Hild but does not move the clock that far forward.
It’s going to take Ms Griffith one more big thick book at the very least to turn
Hild into Saint Hilda, the Abbess of Whitby, and that’s a fine thing; I hope to
go along for the ride.

Now, Dear Reader, I have to share with you that I did not enjoy Menewood as much
as I had Hild. Her world built firm, Ms Griffith lets the plot lead the story
along. Yes, there are still smells and slaughters and sinful sex; the atmosphere
remains thick. But inhabiting Hild’s world is not as much fun as first exploring
it.

Also, at the center of the book, there is probably the most harrowing and
gut-wrenching sequence of fictional events and experiences that has ever kept me
shuddering as I raced through the pages. Ladies, gentlemen, and others, it does
not go well for Hild. A lot of the worst things that can happen to a human
happen and not just one at a time either. George R.R. Martin shook up his
readership pretty hard with the awful things he did to his heroes; “Hold my
beer” says Ms Griffith.

Then, the second half of the book, expanding out from the white-hot tangle of
anguish at the center, is maybe a little predictable. Didn’t stop me enjoying
it, even when I could see pretty clearly where it was heading. And I’ll buy the
next Hild book the moment it’s for sale.

Now let’s give a Griffith book an unqualified rave with no Dear-Reader tropes.

Spear

Hild and Menewood are both doorstop-size; Spear a much slighter offering, just
an evening’s reading or maybe two. It is an odd, idiosyncratic, side-looking
variation on the story of Sir Percival and the Holy Grail, which is my personal
favorite piece of Arthurian lore; at one point in mid-life I became obsessive
and Read All The Books. Hmm, should write about that someday. But I digress.

Few of the characters go by their canonical names but Arthur and Guinevere and
Lancelot and Kay and Merlin and Morgan Le Fay and Gawain are all there,
perfectly recognizable and generally good company. The story takes enjoyably
post-modern twists. And, this being Griffith, has spicy lesbian sex and bloody
passages of arms, losers usually receiving no quarter.

Griffith specialties

I have to say that Ms Griffith is preeminent among the authors I’ve read in
recent decades at depicting conflict conducted with swords and lances and
daggers and staffs and so on. In particular, Spear’s fight between Peretur and
the Red Knight is just unequalled; I defy anyone to offer up a superior literary
episode of single combat.

Let me broaden that a bit. Griffith is just really good at describing the
physical experience of life. Pain, lust, odor, heat/cold, weariness, touch,
touch, and more touch. And by the way, seventh-century Britain was a lot
smellier than the world we live in. Read these books, you’ll come as close as
you can, absent a time machine, to smelling those smells. And you’ll be glad you
did.


MORE
November 110
Today, 14:28Phil Windley's Technometria (Feed)
Permissionless and One-to-One

Summary: Symmetric technology democratizes speech. The web, despite all the good
it's done is fundamentally assymetric. We need a model that is permissionless
and one-to-one.

In a recent post, Clive Thompson speaks of the humble cassette tape as a medium
that had a "a weirdly Internet-like vibe". Clive is focusing on how the cassette
tape unlocked creativity, but in doing so he describes its properties in a way
that is helpful to discussions about online relationships in general.

Clive doesn't speak about cassette tapes being decentralized. In fact, I chuckle
as I write that down. Instead he's focused on some core properties. Two I found
the most interesting were that cassette tapes allowed one-to-one exchange of
music and that they were permissionless. He says:

If you wanted to record a cassette, you didn’t need anyone’s permission.

This was a quietly radical thing, back when cassette recorders first emerged.
Many other forms of audio or moving-image media required a lot of capital
infrastructure: If you wanted to broadcast a TV show, you needed a studio and
broadcasting equipment; the same goes for a radio show or film, or producing and
distributing an album. And your audience needed an entirely different set of
technologies (televisions, radios, projectors, record players) to receive your
messages.

From The Empowering Style of Cassette Tapes
Referenced 2023-11-02T08:01:46-0400

The thing that struck me on reading this was the idea that symmetric technology
democratizes speech. The web is based on assymetric technology: client-server.
In theory everyone can have a server, but they don't for a lot of reasons
including cost, difficulty, and friction. Consequently, the web is dominated by
a few large players who act as intervening administrative authorities. They
decide what happens online and who can participate. The web is not one-to-one
and it is decidedly not permissionless.

In contrast, the DIDComm protocol is symmetric and so it fosters one-to-one
interactions that provide meaningful, life-like online relationships. DIDComm
supports autonomic identity systems that provide a foundation for one-to-one,
permissionless interactons. Like the cassette tape, DIDComm is a democratizing
technology.

Photo Credit: Mix Tape from Andreanna Moya Photography (CC BY-NC-ND 2.0 DEED)

Tags: decentralization identity ssi autonomic didcomm


MORE
November 80
Cloudless: Computing at the Edge

Summary: New use cases will naturally drive more computing away from centralized
cloud platforms to the edge. The future is cloudless.

Doc Searls sent me a link to this piece from Chris Anderson on cloudless
computing. Like the term zero data that I wrote about a few weeks ago, cloudless
computing is a great name that captures an idea that is profound.

Cloudless computing uses cryptographic identifiers, verifiable data, and
location-independent compute1 to move apps to the data wherever it lives, to
perform whatever computation needs to be done, at the edge. The genius of the
name cloudless computing is that it gets us out of the trenches of dapps, web3,
blockchain, and other specific implementations and speaks to an idea or concept.
The abstractions can make it difficult get a firm hold on the ideas, but it's
important to getting past the how so we can speak to the what and why.

You'd be rightly skeptical that any of this can happen. Why will companies move
from the proven cloud model to something else? In this talk, Peter Levine speaks
specifically to that question.

One of the core arguments for why more and more computing will move to the edge
is the sheer size of modern computing problems. Consider one example: Tesla Full
Self Driving (FSD). I happen to be a Tesla owner and I bought FSD. At first it
was just because I am very curious about it and couldn't stand to not have
first-hand experience with it. But now, I like it so much I use it all the time
and can't imagine driving without an AI assist. But that's beside the point. To
understand why that drives computing to the edge, consider that the round trip
time to get an answer from the cloud is just too great. The car needs to make
decisions onboard for this to work. Essentially, to put this in the cloudless
perspective, the computation has to move to where the data from the sensors is.
You move the compute to the data, not the other way around.2

And that's just one example. Levine makes the point, as I and others have done,
that the Internet of Things leads to trillions of nodes on the Internet. This is
a difference in scale that has real impact on how we architect computer systems.
While today's CompuServe of Things still relies largely on the cloud and
centralized servers, that model can't last in a true Internet of Things.

The future world will be more decentralized than the current one. Not because of
some grand ideal (although those certainly exist) but simply because the
problems will force it to happen. We're using computers in more dynamic
environments than the more static ones (like web applications) of the past. The
data is too large to move and the required latency too low. Cloudless computing
is the future.

Notes Anderson calls this deterministic computer. He uses that name to describe
computation that is consistent and predictable regardless of how the application
gets to the data, but I'm not sure that's the core idea. Location independence
feels better to me. An interesting point is that training the AI that drives the
car is still done in the cloud somewhere. But once the model is built, it
operates close to the data. I think this will be true for a lot of AI models.

Photo Credit: Cloudless Sunset from Dorothy Finley (CC BY 2.0 DEED - cropped)

Tags: cloudless picos iot tesla


MORE
November 60
24.6.2024, 6:00SEI Blog (Feed)
The SEI SBOM Framework: Informing Third-Party Software Management in Your Supply
Chain
This post presents a framework to promote the use of SBOMs and establish
practices and processes that organizations can leverage as they build their
programs.

MORE
November 70
22.6.2024, 19:00Platformonomics (Feed)
Platformonomics TGIF #29: November 3, 2023

Subscribe now before Elon MuX does something really erratic!

Why Not Cut Out the Middleman and Just Name AI Safety Rules After Dystopian
Science Fiction Movies?

Previous: “An iron curtain has descended across the continent”, AI Regulation:
Move Fast and Regulate Things, Italy Makes OpenAI Offer They Can’t Refuse

Existential Risk Now Has a Precise Threshold

Suggestions include the Bergeron Constant, the Conga line, the Altman line and
the Cristiano Ronaldo FLOPS line. When will my laptop do 10^26 FLOPs?

Previous: You Can’t Spell Alien Invasion Without A and I, Existential AI Risk +
Twitter-level Nuance

Depreciation Schedules Meet Moore’s Law

Some financiers are furiously hitting F9 in Excel as they realize their
collateralized NVIDIA H100s may not in fact have the six to eight year useful
life they expected a week or two ago. Track H100s on eBay here.

Previous: GPUs as an Asset Class, GPUs Are Better Than Any Financial Asset (For
Now), Brother, Where Art Thy GPUs?

Anthropic Adds Billions For More Existential AI Hand-wringing

Is it vendor financing if Google Cloud has no GPUs available?

Previous: Culture Clash Conjecture: Amazon and Anthropic

OMG, People are Using Electricity

The New York Times has done a search-and-replace on their prior hysterical
stories about power consumption for cloud computing and cryptocurrency, this
time for AI. The scarcity mindset doesn’t seem to learn, even from their own
reporting. Abundant energy is the foundation of advanced civilization. And
businesses will manage costs.

Previous: Cloud Power Up, More Power Consumption Hysteria, Everyone
Hallucinates, Not Just LLMs


MORE
November 40
Today, 9:19mark nottingham (Feed)
Technical Standards Bodies are Regulators

There are lots of ways to view what Internet standards bodies like the IETF and
W3C do. They are horizontal agreements between competitors as well as
mission-driven public-good charities. One might believe they’re the homes of
innovation that brought us the Internet and the Web, or that they’re boring,
ineffective and slow committee talking shops. Some see them as vibrant, diverse
communities, while others believe they’re completely dominated by big tech.

My preferred way of thinking of them these days, however, is as regulators. Just
like the FTC in the US, the eSafety Commissioner in Australia, or the ICO in the
UK, Standards Developing Organizations (SDOs) have a fundamentally regulatory
aspect to them, and considering them in this way clarifies how they relate to
Internet governance.

In particular, it helps to understand what kind of regulator they are, what
tools they use, and the nature of the regime they operate within.

A specific kind of regulator

When most people think of a regulator, they assume it’s always state-backed;
sovereign power (and hopefully a democratic mandate) imbues the regulator with
legitimacy. As Julia Black put it back in 2002:

The core understanding that many have of ‘regulation’ is some form of ‘command
and control’ (CAC) regulation: regulation by the state through the use of legal
rules backed by (often criminal) sanctions. ‘CAC’ has also however become
shorthand to denote all that can be bad about regulation, including poorly
targeted rules, rigidity, ossification, under- or over- enforcement, and
unintended consequences.

Modern conceptions of regulation are much more expansive (or ‘decentered’),
encompassing not only public (government) regulation but also regulation by
private actors. For example, lex mercatoria – commercial law and customs
followed by merchants – goes back to at least medieval times, and is now
considered a kind of regulation. States regularly defer to such ‘soft law’, and
while it can always be overridden in a single jurisdiction by legal power,
policymakers have strong motivations to avoid over-regulating areas that are
capable of self-regulation.

Further complicating Internet regulation is its global span, which means that
more than one state is involved. Transnational Private Regulators (TPRs) are
non-government regulators who work across national boundaries.

Internet SDOs are often used as examples of TPRs. Other common examples include
organisations like the Forestry Stewardship Council, the Fairtrade Foundation,
the International Accounting Standards Board, and the ISEAL Alliance.

Caffagi identified a few factors that have “caused and helped to accelerate the
emergence of TPRs”:

Because “many goods and services today transcend national boundaries [they] can
hardly be regulated by national regulations. This is particularly the case with
global public goods […] for which international regulatory co-operation is
substantially needed to avoid a ‘race to the bottom’ between domestic
regulations.” This is very much the case for the Internet. “There are markets
that exhibit fast-changing dynamics [that are] difficult for public policy
makers to try to regulate[.] In particular, this is the case of high-tech and
knowledge-intensive markets [which] effectively leads policymakers to rely on
private parties, at least for the definition of implementing measures and
technical specifications.” Finally, “there are policy problems that inevitably
require heavy reliance on the expertise of private actors, [who] are the most
informed parties, or the players in the best position to deal with a given
failure, or simply the only parties holding control over central essential
resources.”

Importantly, the legitimacy (and therefore authority) of a TPR isn’t based on
democracy – inherently they have no demos so they cannot be democratic in the
sense that a state is. Instead, they draw on other sources of legitimacy,
including their input (who participates), their output (what impact they have),
and their throughput (what processes they use to assure fair and good outcomes).

With unique regulatory tools

The regulatory tools available to Internet SDOs are specific and limited – they
write down technical specifications that, on a good day, get reflected in code.

This is ‘architectural regulation’, according to Lessig. It sits alongside other
modalities of regulation like law, norms, and markets. Where the FTC uses law,
the IETF uses architecture – shaping behaviour by limiting what is possible in
the world, rather than imposing ex post consequences.

While much of regulatory theory and practice is taken up with issues like
monitoring and enforcement, architectural regulation doesn’t need those tasks to
be performed; the best approximation is conformance testing (which the IETF and
W3C don’t formally do anyway; they certainly don’t attempt certification).

Another interesting aspect of this form of regulation is its quasi-voluntary
nature. Internet standards are optional to adopt and implement; no one is
forcing you to do so. However, if they’re successful and widely adopted, they do
constrain your behaviour while you’re on the Internet, because everyone else is
following them. In that sense, they are mandatory.

Architectural regulation of the Internet is also constrained in how it can
introduce change. While a law can be repealed or overridden by a newer law,
Internet protocol standards have to consider the dependencies that people
already have on infrastructure; we can’t have a ‘flag day’ where we change how
the Internet works. Instead, we have to carefully extend and evolve it, working
within the constraints of what people already do, because once code is deployed,
we lose control.

These features provide interesting advantages to SDOs as regulators. While one
might see a non-state regulator without an enforcement problem as too powerful,
standards’ lack of binding force means that an SDO can’t just impose its will;
its product has to be proven by market adoption. A successful, widely adopted
standard is (qualified) proof of cooperation, and thus has gained legitimacy at
the same time it becomes binding.

In a large regulatory space

If we step back from this, we can now consider the context of this regulation -
Internet Governance overall. Plenty has been written about this that I won’t
attempt to summarise, but there are a couple of aspects that I’d like to point
out.

First of all, there are (obviously) other regulators present too – legal
regulators especially (from various governments around the world), but also
others using various combinations of the regulatory modalities.

Second, Internet Governance is polycentric (also referred to as ‘regulatory
pluralism’) - there is no hierarchy and no regulator can tell another what to
do. There are many sources of power (of various natures) that interact in
different ways – sometimes reinforcing each other, occasionally conflicting.

Lessig talks about this (with ‘constraints’ being a synonym for ‘regulators’):

The constraints are distinct, yet they are plainly interdependent. Each can
support or oppose the others. Technologies can undermine norms and laws; they
can also support them. Some constraints make others possible; others make some
impossible. Constraints work together, though they function differently and the
effect of each is distinct. Norms constrain through the stigma that a community
imposes; markets constrain through the price that they exact; architectures
constrain through the physical burdens they impose; and law constrains through
the punishment it threatens.

Third, the regulatory space is also fragmented. Information, authority,
responsibility, and capacity to regulate are dispersed unevenly across multiple
regulators. As Scott points out, ‘[r]elations can be characterized as complex,
dynamic horizontal relations of negotiated interdependence.’

This means that no regulator in the space is truly independent. Standards have
to operate in the legal contexts where they’re deployed; laws need to take the
reality of the deployed Internet into account. Each party can act unilaterally,
and might even meet their immediate goals, but the reaction to imprudent actions
might be worse than the original issue they were trying to address.

Overall, this is healthy. Power is not concentrated in any one institution.
States are able to claim sovereignty over what happens inside their borders, but
if they differ too much from the global norm, they put at risk the economic and
cultural benefits of being part of the global Internet.

What does this mean for the IETF and W3C?

Accepting the regulatory nature of SDOs leads to a few conclusions.

First, the IETF and W3C need to coordinate more closely with other regulators –
especially national regulators who have their sights set on taming particular
aspects of the Internet.

That doesn’t mean that SDOs should defer to national regulators – far from it.
I’ve heard more than a few conversations where technical people think they need
to implement the law in protocols. This is not the case, because laws are
generally limited to a specific territory; countries can’t regulate the entire
Internet by themselves. Furthermore, laws typically don’t apply to the standards
themselves; instead, they apply to their use.

It doesn’t even mean that standards work should block on getting input from
policymakers (just as policymakers don’t block lawmaking on feedback from
SDOs!); doing so would introduce problematic incentives, muddy the technical
decision-making process, and remove many of the advantages of private
regulation.

It does mean that technical discussions should be informed by ‘policy
considerations’, even if they’re ultimately dismissed. Understanding how legal
regulators see the Internet, what their goals are, and how they attempt to use
the regulatory tools in their hands helps technical regulators evaluate what
additional constraints are likely to be layered onto the Internet. That might
result in alignment between technical regulation and legal regulation, but this
is emphatically not a requirement – in some cases, they might conflict.

Those conflicts should be avoided when they’re unnecessary, so SDOs need to do
their part to inform legal regulators as well, particularly when their proposals
have impact on the architecture.

This is not a new perspective – there has been considerable discussion in both
the IETF and the W3C recently about ‘policy engagement.’ What’s different here
is the emphasis on being a peer of other regulators, rather than automatically
subject to them. That is fundamentally different than the relationship that most
corporate policy units have with regulators, for example.

Second, this view reinforces the notion that regulation by technical standards
bodies has very specific sources of legitimacy – the technical expertise that it
embodies, and the demonstrated success of its output. That legitimacy might be
enhanced by the unique global scope of these bodies – unlike national
regulators, they are responsible for the entire Web and Internet.

That suggests the positions taken by these bodies need to be focused on their
areas of expertise, rather than trying to draw on other sources of legitimacy
(for example, pseudo-democratic ones, or notions of openness, although the
latter does enhance their legitimacy). This is well-recognised in the IETF,
where arguments like Pervasive Monitoring is an Attack are couched in technical
terms, not value-driven ones.

Third, the polycentric and fragmented nature of the regulatory space suggests
that it’s entirely appropriate for architectural regulators like SDOs to focus
on areas where their tools are most effective.

For example, the HTTP Cookie specification has been working towards eradicating
third-party cookies for some time, because they’re horrible for privacy. Some
point out that this doesn’t address the privacy issues with first-party cookies
- a site you’re interacting with can still track your activity, profile you, and
so on.

That doesn’t mean that we should back away from regulating third-party cookies
with architecture; they’re extremely amenable to this form of regulation
(because of the user agency of the browser), and legal regulation of third-party
cookies has proven difficult. On the other hand, regulating first-party privacy
abuses on the Web with architecture is hard – if you interact with someone,
you’re giving them your data – but legal regulation of how entities handle
first-party data is on much firmer ground (provided there is a political will to
do so).


MORE
November 20
23.6.2024, 18:04The Architect Elevator (Feed)
Spooky: Platform Strategy Awakens
What better day to launch The Grim Wrapper than Halloween?

MORE
November 10
1.5.2024, 16:56James Governor's Monkchips (Feed)
RedMonk beers at Kubecon

RedMonk beers are going to be a blast at Kubecon this year. We’ve found just the
place, which reflects what we’re all about. The venue is called Maria’s Packaged
Goods and Community Bar – described as a “Hybrid liquor store & neighborhood
tavern supplying a large rotating menu of rare craft beers.” It has plenty of
outdoor space.

Rare craft beers, community and outdoor seating? Definitely on brand. Definitely
our vibe.

One reason our team was excited when the CNCF announced Kubecon would be in
Chicago this year is that Morgan Harris, our account and engagement manager,
lives in the city. So it felt like we could mix Cloud Native and Chicago Native.
Naturally we asked Morgan to scout a venue for us, and of course she came up
trumps.

The story behind the venue is lovely.



Maria is a South Korean, who moved to Chicago after marrying her husband, an
American serviceman called James back in the 1970s. Sadly he passed away at a
young age. Maria though found her feet in a tough neighborhood and community in
Bridgeport, Chicago. Today she is a local icon – and has a beer in her name –
The Duchess of Bridgeport – a sour red ale we can’t wait to try. Locals call her
“mom” or the Peggy Guggenheim of Bridgeport because of the way she supports
local artists. If you’d like to know more about this history of the place this
is a great article to check out.

The food is Korean Polish, which sounds amazing (think Polish Sausage, Pierogis
and Korean chicken wings), with some nods to Italian. We look forward to hosting
our own community in this community institution. Join us at Maria’s Community
Bar on Wednesday, November 8th. Don’t forget your ID, because Maria herself may
be checking them at the door.

We’ve booked a few tables and look forward to seeing old friends there and
meeting some new ones.

Date: November 8

Time: 7:30-9:30pm CT

Location: Maria’s Packaged Goods & Community Bar; 960 W 31st St
Chicago IL 60608

Bonus post material – be sure to read this great thread about Chicago before you
travel.

So you're looking to attend #KubeCon in Chicago? Welcome! Chicago is an
incredible city and I hope you have a blast while here. I've put some thoughts
together here for anyone coming to the city for the conference. I've only lived
here for just over a year but really like it. 1/

— Kai @kaipmdh@hachyderm.io (@KaiPMDH) October 28, 2023



 

 

 


MORE
October 310
24.6.2024, 6:00SEI Blog (Feed)
Dealing with Noisy Behavioral Analytics in Detection Engineering
This post discusses a process for tuning and related work you can do to make
behavioral analytics more viable in your environment, without sacrificing
analytic coverage.

MORE
October 300
Today, 17:55ongoing by Tim Bray (Feed)
On C2PA

Leica, the German maker of elegant but absurdly-expensive cameras, just released
the M11-P. The most interesting thing about it is a capability whose marketing
name is “Content Credentials”, based on a tech standard called C2PA (Coalition
for Content Provenance and Authenticity), a project of the Content Authenticity
Initiative. The camera puts a digital watermark on its pictures, which might
turn out to be extremely valuable in this era of disinformation and sketchy AI.
Herewith a few words about the camera (Leicas are interesting) but mostly I want
to describe what C2PA does and why I think it will work and how it will feel in
practice.

M11-P

To start with, this thing lists at over $9,000 in the US. There are lots of
awesome lenses for it, and you might be able to find one for under $5,000, but
not usually.

On the other hand, it’s a lovely little thing.

The obvious question: Can it possibly be worth that much money? Well… maybe. Any
camera on sale today, including your phone, can reliably take brilliant
pictures. But people who use Leicas (I never have) rave about the ergonomics, so
you might be a little quicker on the draw. And they say it’s fun to use, which
means you’re more likely to have it with you when the great photo op happens.
And there’s no denying it looks drop-dead cool.

C2PA is puzzling

I’ve been impressed by the whole C2PA idea ever since I first heard about it,
but damn is it hard to explain. Every time I post about it, I get annoying
replies like “I don’t want my camera to track me!” and “This is just NFTs
again!” and “It’ll be easy to fake by tampering with the camera!” All of which
are wrong. I conclude that I’m failing to explain clearly enough.

Whatever, let’s try again.

Signing

Inside the M11-P there is special hardware, and inside that hardware are two
closely-linked little blobs of binary data called the “public key” and the
”private key”; we call this a “keypair”. The hardware tries to be
“tamper-proof”, making it very hard for anyone to steal the private key. (But
nothing is perfect; a real security expert would assume that a serious
well-resourced hacker could crack and steal. More below.)

When you take a picture, the camera makes a little data package called a
“manifest”, which records a bunch of useful stuff like the time, the camera
serial number, the name of the person who owns the camera, and so on. Then it
runs a bunch of math over the private key and manifest data and the image pixels
to produce a little binary blob called the “signature”; the process is called
“signing”. The manifest and the signature are stored inside the metadata (called
“EXIF”) that every digital photo has.

Then, you share the public key with the world. Email it to your colleagues.
Publish it on your website. Whatever. And anyone who gets your picture can run a
bunch of math over the public key and manifest and pixels, and verify that those
pixels and that manifest were in fact signed by the private key corresponding to
the public key the photographer shared.

Geeky interlude for PKI nerds

(If the “PKI” acronym is new to you, do please skip forward to the “Chaining”
section.)

Leica has posted a Content Credentials demo page with a sample image. Big thanks
to Sam Edwards (@samedwards@mastodon.social), who dug around and found the
actual JPG, then taught me about c2patool. All this happened in a nice chatty
set of Mastodon threads starting here; the Fediverse is really the place for
substantive conversation these days.

The actual image.

(That is, once you click to enlarge it. But watch out, it’s 21M).

Applying c2pa to the JPG yields the JSON manifest, which has a selection of
useful EXIF fields. It turns out the signing relies on traditional PKI-wrapped
certs; there’s one associated uniquely with this camera, with a proper signing
chain through a Leica cert, all apparently rooted at D-Trust, part of Germany’s
Bundesdruckerei which also prints money. All very conventional, and whatever
programming language you’re using has libraries to parse and verify. Sadly,
ASN.1 will never die.

By the way, the actual C2PA spec feels way more complicated than it needed to
be, with Verifiable Credentials and Algorithm Agility and JSON-LD and CBOR and
COSE etc etc. I haven’t had the bandwidth to slog all the way through. But…
seems to work?

Chaining

I’ve described how we can attach a signature to a photo and anyone who has the
camera’s public key can check whether it was signed by that camera. That’s
clever, but not very useful, because before that picture gets in front of human
eyes, it’s probably going to be edited and resized and otherwise processed.

That’s OK because of a trick called “signature chaining”. Before I explain this,
you might want to drop by Leica’s Content Credentials page and watch the little
video demo, which isn’t bad at all.

Now, suppose you change the photo in Photoshop and save it. It turns out that
Photoshop already has (a beta version of) C2PA built in , and your copy on your
own computer has its own private/public keypair. So, first of all, it can verify
the incoming photo’s signature. Then when you save the edited version, Photoshop
keeps the old C2PA manifest but also adds another, and uses its own private key
to sign a combination of the new manifest, the old manifest (and its signature),
and the output pixels.

There’s enough information in there that if you have the public keys of my
camera and my copy of Photoshop, you can verify that this was a photo from my
camera that was processed with my Photoshop installation, and nobody else got in
there to make any changes. Remember, “signature chaining”; it’s magic.

If you run a news site, you probably have a content management system that
processes the pictures you run before they hit your website. And that system
could have its own keypair and know how to C2PA. The eventual effect is that on
the website, you could have a button labeled “Check provenance” or some such,
click it and it’d do all the verification and show you the journey the picture
took from camera to photo-editor to content-management system.

Why?

Because we are surrounded by a rising tide of disinformation and fakery and
AI-powered fantasy. It matters knowing who took the picture and who edited and
how it got from some actual real camera somewhere to your eyes.

(By the way, AI software could do C2PA too; DALL-E could sign its output if you
needed to prove you were a clever prompter who’d generated a great fantasy pic.)

But this can’t possibly work!

Like I said above, every time I’ve posted something nice about C2PA, there’ve
been howls of protest claiming that this is misguided or damaging or just can’t
work. OK, let’s run through those objections one by one.

No, because I want to protect my privacy!

A perfectly reasonable objection; some of the most important reportage comes
from people who can’t afford to reveal their identity because their work angers
powerful and dangerous people. So: It is dead easy to strip the metadata,
including the C2PA stuff, from any media file. In the movie linked from the
Leica web site above, you’ll notice that he has to explicitly turn C2PA on in
both the camera and in Photoshop.

Yes, this means that C2PA is useless against people who steal your photos and
re-use them without crediting or paying you.

I can’t see any reason why I’d attach C2PA to the flowers and fripperies I
publish on my blog. Well, except to demonstrate that it’s possible.

No, because I’m not a cryptography expert!

Fair enough, but neither am I. This demo page shows how it’ll work, in practice.
Well, early-stage, it’s kind of rough-edged and geeky. Eventually there’ll be a
nicely-styled “verify” button you click on.

No, because corporate lock-in!

Once again, reasonable to worry about, and I personally do, a lot. Fortunately,
C2PA looks like a truly open standard with no proprietary lock-ins. And the
front page of the Content Authenticity Initiative is very reassuring, with
actual working code in JavaScript and Rust. I’m particularly pleased about the
Rust SDK, because that can be wired into software built in C or C++, which is,
well, almost everything, directly or indirectly.

For example, the Leica-provided image you see above has no C2PA data, because
it’s been resized to fit into the browser page. (Click on it to get the
original, which retains the C2PA.) The resizing is done with an open-source
package called ImageMagick, which doesn’t currently do C2PA but could and I’m
pretty sure eventually will. After which, the picture above could have a link in
the signature chain saying “resized by ImageMagick installed on Tim Bray’s
computer.”

No, because of the “analog hole”, I’ll just take a picture of the picture!

This doesn’t work, because the signing computation looks at every pixel, and
you’ll never get a pixel-perfect copy that way.

No, because bad guys will sign fake images!

Absolutely they will, no question about it. C2PA tells you who took the picture,
it doesn’t tell you whether they’re trustworthy or not. Trust is earned and
easily lost. C2PA will be helpful in showing who has and hasn’t earned it.

No, because it will lead to copyright abuse!

It is definitely sane to worry about over-aggressive copyright police. But C2PA
won’t help those banditos. Sure, they can slap a C2PA manifest, including
copyright claims, on any old image, but that doesn’t change the legal landscape
in the slightest. And, like I said, anyone can always remove that metadata from
the image file.

No, because artists will be forced to buy in!

Yep, this could be a problem. I can see publishers falling overly in love with
C2PA and requiring it on all submissions. Well, if you’re a film photographer or
painter, there’s not going to be any embedded C2PA metadata.

The right solution is for publishers to be sensible. But also, if at any point
you digitize your creations, that’s an occasion to insert the provenance data.
We’ll need a tool that’s easy to use for nontechnical people.

No, because it’s blockchain! Maybe even NFTs!

It’s not, but you can see how this comes up, because blockchain also uses
signature chains, there’s nothing in principle wrong with them. But C2PA doesn’t
need any of the zero-trust collective-update crap that makes anything with a
blockchain so slow and expensive.

No, because hackers will steal the private key and sign disinformation!

Definitely possible; I mentioned this above. When it comes to computer security,
nothing is perfect. All you can ever do is make life more difficult and
expensive for the bad guys; eventually, the attack becomes uneconomic. To steal
the private key they’d have to figure out how to take the camera apart, get at
the C2PA hardware, and break through its built-in tamper-proofing. Which I’m
sure that a sufficiently well-funded national intelligence agency can do, or a
sufficiently nerdy gang of Bolivian narcos.

But, first of all, it wouldn’t be easy, and it probably wouldn’t be terribly
fast, and they’d have to steal the camera, hack it, put it back together, and
get it back to you without you noticing. Do-able, but neither easy nor cheap.
Now, if you’re a Hamas photoblogger, the Mossad might be willing and capable.
But in the real world, when it really matters, the attackers are more likely to
use the XKCD technique.

No, because websites don’t care, they’ll run any old gory clickbait pic!

Absolutely. C2PA is only for people who actually care about authenticity. I
suspect it’s not gonna be a winner at Gab or Truth Social. I hope I’m not crazy
in thinking that there are publishing operations who do care about authenticity
and provenance.

OK then. How will it be used in practice?

I remain pretty convinced that C2PA can actually provide the provenance-chain
capability that it claims to. [Note that the C2PA white papers claim it to be
useful for lots of other things that I don’t care about (thanks to vince for
pointing that out) and this piece is already too long, so I’ll ignore them.]
Which raises the question: Where and how will it be used? I can think of two
scenarios: High-quality publishing and social media.

The Quality Publishing workflow

We’re looking at The Economist or New Yorker or some such, where they already
employ fact checkers and are aggressive about truth and trust. Their photos
mostly come from indies they work with regularly, or big photo agencies.

Let’s look at the indie photographer first. Suppose Nadia has been selling pix
to the pub for years, now they want to do C2PA and Nadia has a camera that can.
So they tell Nadia to send them a picture of anything with C2PA enabled. They
have a little database (Microsoft Access would be just fine) and a little app
that does two things. First, when they get the sample photo from Nadia, there’s
a button that reads the photo, extracts and verifies the C2PA, and writes an
entry in the database containing Nadia’s camera’s public key and the way she
likes to be credited.

From then on, whenever they get a pic from Nadia, they feed it to the app and
press the other button, which extracts the C2PA and looks up the public key in
the database. If it doesn’t match anything, there’s a problem and they probably
shouldn’t run that picture without checking things out. If everything’s OK,
it’ll create a nice little chunk of HTML with the credit to Nadia and a link to
the HTML-ized provenance chain to show to anyone who clicks the “provenance”
button beside the picture.

You know, if I were building this I’d make sure the database record included the
email address, then I’d set the app up so the photog just emails the picture to
the app, then the app can use the pubkey to pull the record and see if the email
sender matches the database.

In the case of the agency photographers, the agency could run the database and
app  on its website and the publisher could just use it. Neither option sounds
terribly difficult or expensive to me.

The idea is that displaying the provenance button emphasizes the seriousness of
the publisher and makes publishers who aren’t using one look sketchy.

The social-media workflow

The thinking so far seems to have been aimed at the high-end market I just
discussed; after all, the first camera to implement C2PA is one of the world’s
most expensive. I understand that Nikon has a camera in the pipeline and I bet
it’s not going to be cheap either. [Sad footnote: I gather that Sony is building
this into its cameras too but, being Sony, it’s not using C2PA but some
Sony-proprietary alternative. Sigh.]

But on reflection I’m starting to think that C2PA is a better fit for social
media. In that domain, the photos are overwhelmingly taken on mobile-phone
cameras, and every app, bar none, has a media-upload feature.

Speaking as a former Android insider, I think it’d be pretty easy to add C2PA to
the official Camera app or, failing that, to a C2PA-savvy alternate camera app.

I also think it’d be pretty easy for the Instagrams and TikToks of this world to
add C2PA processing to their media-upload services. Obviously this would have to
be explicitly opt-in, and it’d probably work about the same way as the Quality
Publishing workflow. You have to initially upload something with a C2PA manifest
to get your public key registered and tied to your social-media identity. Then
you’d have to decide whether you wanted to attach C2PA to any particular picture
or film-clip you uploaded.

I dunno, on a snakepit of sketchy information like for example Reddit I think
there’d be real value, if I happened to get a good picture of an cop brutalizing
a protester or a legislator smooching the Wrong Person or Ukrainian troops
entering a captured town, to C2PA-equip that image. Then you could be confident
that the trustworthiness of the image is identical to the trustworthiness of the
account.

And if some particularly-red hot video capture either didn’t have the
“provenance” badge, or it did but was from Igor48295y2 whose account was created
yesterday, well then… I’m not so optimistic to think it’d be dismissed, but it’d
be less likely to leak into mainstream media. And — maybe more important — if it
were super newsworthy and CP2A-attributable to someone with a good record of
trust, it might get on national TV right away without having to wait for the
fact-checkers to track down the photog and look for confirming evidence.

Are you done yet, Tim?

Sorry, this one got a little out of control. But over the decades I’ve developed
a certain amount of trust in my technology instincts. C2PA smells to me like a
good thing that could potentially improve the quality of our society’s public
conversation with itself.

And I’m in favor of anything that helps distinguish truth from lies.


MORE
November 10
22.6.2024, 19:00Platformonomics (Feed)
Platformonomics TGIF #28: October 27, 2023

Be sure to subscribe below and to the right to receive all my output via email

Q3 2023 Cloud CAPEX

Amazon $12.76B -19%
Google $8.06B +11%
Microsoft $11.2B +70%

(This is corporate level CAPEX, not just cloud infrastructure)

Amazon’s decline seems like the continued reduction in logistics/fulfillment
spending. We’ll get a breakout of AWS CAPEX next quarter. The Google mystery
continues: the normal server refresh is getting pushed ever further out with no
sign of a ramp in GPUs (and hard to argue TPUs are cheap). Microsoft passing
Amazon for total spend would be a milestone.

Previous: Q2 Cloud CAPEX, The Google CAPEX Mystery, Follow the CAPEX: Cloud
Table Stakes 2022 Retrospective

GPUs as an Asset Class

I have suggested developing nations peg their currency to GPUs. We’re not quite
there yet (Argentina, don’t miss your chance!), but the insatiable
financialization impulses of Wall Street are turning GPUs into “a new asset
class”. This will end poorly.

Previous: GPUs Are Better Than Any Financial Asset (For Now), Brother, Where Art
Thy GPUs?
Related: GPUs’ Latest Magic Act: Securing Data Center Debt

Clown CAPEX

IBM is in the midst of a strategic transformation from a cloud to an AI . The
absence of CAPEX in either domain ensures you are a .

Previous: Bonsai AI, Follow the CAPEX: Keeping Up With The Clowns, Follow the
CAPEX: Clown Watch, Follow the CAPEX: Separating the Clowns from the Clouds
Related: DXC Technology Hires Industry Veteran Howard Boville to Senior
Leadership Team

Cloud Repatriation: The Search Continues

So the data center guys aren’t benefiting from cloud repatriation or AI…

Previous: Cloud Repatriation: Still Not a Thing, Platformonomics Repatriation
Index – Q1 2023: Surf’s Up?, The Inexorable Attraction Between Terrible Things:
Cloud Repatriation and Private Equity

The Amazon AI Inferiority Complex

Previous: Excellent Trolling by The Information, Amazon Now Forecasts They Won’t
Be Competitive In Generative AI For Over 13,000 Years, Amazon’s AI
Communications Clown Show: Cause or Effect?, Amazon Pauses Lectures on
Generative AI, Three Steps into a 12-Step Program, Culture Clash Conjecture:
Amazon and Anthropic, AWS: “It’s Day 0.1 in generative AI”, AWS: “We’re three
rankles into a 10k rankle”, “Every Single” Amazon PR Team is Working on
Generative AI (Part I’ve Lost Count), “Every Single” Amazon PR Team is Working
on Generative AI, Titan or Titanic?, Amazon Pioneers LLM-less Generative AI, A
PR Campaign About Nothing: AWS and Generative AI, Day Two of Amazon’s PR
Campaign About Nothing, Amazon’s Relentless PR Campaign About Nothing
Related: Amazon Cranks Up Its AI Euphoria

Saying the Quiet Part Out Loud: AWS Edition

Considering revising “Two pizza teams are organizationally indistinguishable
from Soviet spy cells” to something involving a million monkeys and Shakespeare.

Regulating Your Way to Prosperity

Once the European car industry is gone, what will be left? Wine, cheese,
tourism, anything else?

Previous: The EU’s Reckoning Approaches, Collapsing Economic Models: German
Edition, EU Rethinking Its Stance on Mercantilism, German Auto Industry Reaches
Hail Mary Stage, How Do You Say “Burning Platform” in German?, Volkswagen’s
Death Throes Continue, The Fate of the European Economy: Automotive Edition,
Europe: Investing in Slave Labor

Software Migration Alert: EngageSmart

When private equity comes amalgamating, it is time to start migrating. Why?
Because private equity ruins software companies.

Previous: Perfidious Private Equity


MORE
October 280
Today, 17:55ongoing by Tim Bray (Feed)
Wild World

I’ve come to loathe autumn. It echoes the current stage of my own life a little
too closely. Plus it’s plenty hard to just stay level, any season of this filthy
year. I saw someone on social media asking (may not have the words quite right)
“How you folks holding it together while the world’s burning?” I looked for
something useful and short to say, came up empty. But now maybe this.

There’s the plague and the fascists and mass shooters and people killing babies
for the greater glory of Palestine and Israel. Lies burn hot on every Internet
screen and who’s got time to refute them all? Greed lights fires everywhere.
Feaugh.

Also I have family issues that hurt and teach hard lessons. Long Covid in da
house too.

And among my friends and contemporaries and relations, there’s a steady rumble
of mortality: cancer diagnoses and dementia symptoms and heart trouble and more
like that.

Anyhow, what happened was, last Saturday morning I had stuff to pick up and drop
off at the local library. It’s across from the local farmers market. So I biked
up there and hit both. Nobody else in the family was up to the trip, for a
variety of reasons.

Cool, but I bought sprouts in a nice little bag, no stems.

I got Brussels sprouts and focaccia and apples. The market has a spot for
musicians, last week’s a guy with a strong voice singing old songs everyone
knows; didn’t pay much attention.

All this is good, right? This is the “15 minute city” model, I think. Yes, it
was. I enjoyed it.

So I lugged the books and movies and fruits and vegetables back to the bike
rack, knelt to unlock and stow. Then the minstrel started singing Wild World and
I knew I was in trouble.

I’m placid generally; among the calmest percentiles of the population. I’m
reasonable. I deal with shit. I do a good job of seeming balanced. None of which
helped.

But then a lot of nice things turn bad out there…

Oh, baby, baby, it's a wild world

It's hard to get by just upon a smile

I leaned my head against the bike and wept hard, cheeks wet. Dunno where it came
from. Everywhere I guess. Nobody else was unhooking their bike just then so
maybe nobody noticed.

It’s a wild world, no shit. Hang on for dear life. Teach your children. Then
wish ’em luck. They’re going to need it.

Brussels sprouts

A few minutes in the Instant Pot air fryer with a bit of olive oil and they come
out slightly blackened and just great.

The lesson

You don’t have to keep it together. Losing it sometimes is OK.


MORE
October 290
1.5.2024, 16:56James Governor's Monkchips (Feed)
DevSecOps and Progressive Delivery at Dunelm – modern enterprise development
practices

I saw a great presentation from Dunelm, the UK’s leading home furnishings
retailer, at GitLab’s DevSecOps world tour event in London recently. The company
has clearly successfully retooled for ecommerce – with a builder mentality, and
engineering practices that should stand it in good stead. The company is
accelerating its transformation – in 2022 it was doing 11 production deployments
a day. Now, in 2023, it’s doing 26. One aspect of the talk I appreciated was the
focus on DevSecOps – and by that I mean security enabling and helping developers
to build secure apps, rather than getting in their way.

Jan Claeyssens, DevSecOps Principal Engineer at Dunelm, explained that his role
is to enable and engage with developers and engineering teams:

The development teams are my customers. Security needs to stop saying no but
lean in, listen to what they want and try and help them. No one wants more
checks after they have finished. Security scanners should not impact the APIs
too much.

Claeyssens also pointed out how important education is. You can’t expect
developers to use security features if they don’t know they’re there. Platform
engineering and security teams need to do a better job of developer education in
order to get the results they want. The whole tone from Claeyssens was
refreshing from a security perspective.

Serve the user where they live and show them what the features are. Security at
Dunelm has to help the business become better.

Amen Jan.

Talking of amen, I was really pleased to hear the company has adopted
Progressive Delivery as an approach. Given I coined the term, it was great to
hear concepts read out from an enterprise company. To be fair GitLab has used
Progressive Delivery in its marketing so it should not surprise me that a
customer would be using the language. But it was still pleasing.

Paul Kerrison, Director of Engineering and Architecture at Dunelm said:

Our engineering work is becoming more experiment driven, we are moving towards
progressive delivery, the new kid on the block.”

You get to try more risky things but safely. We can put something in production,
test it with one one cohort, then roll it out more broadly.

Cohorts, phased rollouts, reduced risk with more experimentation, testing in
production. This is the way.

 

disclosure: GitLab is a client.


MORE
October 240
Today, 14:28Phil Windley's Technometria (Feed)
Zero Data

Summary: Like Zero Trust, Zero Data represents a paradigm that organizations can
embrace to enhance their overall security posture. But the benefits go far
beyond better security.

A few weeks ago, I came across this article from StJohn Deakin from 2020 about
zero data. I've been thinking a lot about zero trust lately, so the name leapt
out at me. I immediately knew what StJohn was speaking about because I've been
talking about it too. My new book, Learning Digital Identity, talks about the
concept. But the name—zero data—is simply brilliant. I want to dig into zero
data in this post. I'll discuss the link between zero data and zero trust in a
future post.

StJohn describes the idea like this:

Personal data should be held by humans first, and by the companies,
organisations and governments that humans choose to interact with, second. The
ultimate in ‘data minimisation’ is for the platforms in the centre to simply
facilitate the interactions and not hold any data at all. This is the exact
opposite of our Google/Facebook/Amazon dominated world where all human data is
being concentrated in a few global silos.

A zero data society doesn’t mean that data isn’t shared between us, quite the
opposite. With increased trust and participation, the data available and needed
to drive our global society will explode exponentially.

From The Future of Data is ‘Zero Data’
Referenced 2023-09-30T17:37:15-0600

If you think about this in the context of how the internet has worked for the
last three decades, the concept of zero data might seem baffling. Yet, consider
a day in your life. How often do you establish lasting relationships—and thus
share detailed information about yourself—with every individual or entity you
come across? Almost never. It would be absurd to think that every time you grab
a coffee from the local store, you'd need to form a lasting bond with the coffee
machine, the cashier, the credit card terminal, and other customers just to
facilitate your purchase. Instead, we exchange only the essential information
required, and relevant parties retain just the data that is needed long term.

To build a zero data infrastructure we need to transfer trustworthy data
just-in-time. Verifiable credentials (VCs) offer a way to represent information
so that its authenticity can be verified through cryptographic means. They can
be thought of as digital attestations or proofs that are created by an issuer
about a subject and are presented by the holder to a verifier as required.

Verifiable Credential Exchange (click to enlarge)

Here are some of the interaction patterns facilitated by verifiable credentials:

Selective Disclosure: VCs enable users to share only specific parts of a
credential. For instance, a user can prove they are of legal age without
revealing their exact date of birth. Credential Chaining: Multiple credentials
can be linked together, enabling more complex proofs and interactions. For
example, an employer might hire an employee only after receiving a VC proving
they graduated and another proving their right to work. Holder-Driven Data
Exchange: Instead of organizations pulling data about users from third parties,
VCs shift the interaction model to users pushing verifiable claims to
organizations when needed. Anonymous Credential Proofs: VCs can be designed to
be presented anonymously, allowing users to prove a claim about themselves
without revealing their identity. For example, VCs can be used to prove the
customer is a human with less friction than CAPTCHAs. Proofs without Data
Transfer: Instead of transferring actual data, users can provide cryptographic
proofs that they possess certain data or prove predicates about the data,
reducing the exposure of personal information. For example, VCs can be used to
prove that the subject is over 21 without revealing who the subject is or even
their birthdate. Adaptive Authentication: Depending on the sensitivity of an
online interaction, users can be prompted to provide VCs of varying levels of
assurance, enhancing security in adaptable ways. I plan to talk about this more
in my next post about zero data and zero trust.

These interaction patterns change traditional data management and verification
models, enabling businesses to retain considerably less data on their clients.
Verifiable credentials have numerous benefits and features of that provide a
positive impact on data management, security, and user trust:

Data Minimization: As we've seen, with VCs, users can prove facts without
revealing detailed data. By selectively sharing parts of a credential,
businesses only see necessary information, leading to overall reduced data
storage and processing requirements. Reduced Redundancy & Data Management:
Trustworthy VCs reduce the need for duplicate data, simplifying data management.
There's less need to track, backup, and maintain excess data, reducing
complexity and associated costs. Expiration, Revocation, & Freshness of Data:
VCs can be designed with expiration dates and can be revocable. This ensures
verifiers rely on up-to-date credentials rather than outdated data in long-term
databases. Trust through Standardized Protocols: VCs, built on standardized
protocols, enable a universal trust framework. Multiple businesses can thus
trust and verify the same credential, benefiting from reduced integration
burdens and ensuring less custom development. Enhanced Security & Reduced
Exposure to Threats: Data minimization reduces the size of the so-called honey
pot, reducing the attraction for cyber-attacks and, in the event of a breach,
limit the potential damage, both in terms of data exposed and reputational harm.
Compliance, Regulatory Benefits & Reduced Liability: Adhering to data
minimization aligns with many regulations, reducing potential legal
complications. Storing minimal data also decreases organizational liability and
regulatory scrutiny. Cost Efficiency: By storing less data, organizations can
achieve significant savings in storage infrastructure and IT operations, while
also benefiting from focused data analytics. Enhanced User Trust & Reputation:
By collecting only essential data, organizations can build trust with users,
gaining a competitive edge in a privacy-conscious market that is increasingly
growing tired of the abuses of surveillance capitalism.

In essence, verifiable credentials shift the paradigm from "data collection and
storage" to "data verification and trust." This is what Marie Wallace means with
her analogy between VCs and music streaming. Online interactions are provided
with the assurance they need without the business incurring the overhead (and
risk) of storing excessive customer data. Zero data strategies not only reduce
the potential attack surface for cyber threats but also offers a variety of
operational, financial, and compliance benefits.

The biggest objection to a zero data strategy is likely due to its decentralized
nature. Troves of user data make people comfortable by giving them the illusion
of ready access to the data they need, when they need it. The truth is that the
data is often unverified and stale. Nevertheless, it is the prevailing mindset.
Gettng used to just-in-time, trustworthy data requires changing attitudes about
how we work online. But the advantages are compelling.

And, if your business model depends on selling data about your customers to
others (or facilitating their use of this data in, say, an ad network) then
giving up your store of data may threaten precious business models. But this
isn't an issue for most businesses who just want to facilitate transactions with
minimal friction.

Zero data aligns our online existence more closely with our real-world
interactions, fostering new methods of communication while decreasing the
challenges and risks associated with amassing, storing, and utilizing vast
amounts of data. When your customers can prove things about themselves in real
time, you'll see several benefits beyond just better security:

Reduced Sign-Up Friction: For services that rely on verified attributes (e.g.,
age, membership status, qualifications), users can provide these attributes
quickly with VCs, eliminating lengthy sign-up or verification processes.
Cross-Platform Verification: A VC issued by one service can be verified and
trusted by another, facilitating smoother cross-platform interactions and
reducing the need for users to repetitively provide the same information. Fewer
intermediaries: VCs can allow for direct trust between parties without the need
for a centralized authority. This fosters more direct and decentralized
interactions.

Zero data, facilitated by verifiable credentials, represents a pivotal
transition in how digital identity is used in online interactions. By minimizing
centralized data storage and emphasizing cryptographic verifiability, this
approach aims to address the prevalent challenges in data management, security,
and user trust. Allowing online interactions to more faithfully follow
established patterns of transferring trust from the physical world, the model
promotes just-in-time data exchanges and reduces unnecessary data storage. As
both businesses and individual users grapple with the increasing complexities of
digital interactions, the integration of verifiable credentials and a zero data
framework stands out as a practical, friction-reducing, security-enhancing
solution for the modern digital landscape.

Tags: zero+data identity verifiable+credentials zero+trust


MORE
October 240
Internet Identity Workshop 37 Report

Summary: The latest IIW was great with many high intensity discussions of
identity by people from across the globe.

We recently completed the 37th Internet Identity Workshop. We had 315 people
from around the world who called 163 sessions. The energy was high and I enjoyed
seeing so many people who are working on identity talking with each other and
sharing their ideas. The topics were diverse. Verifiable credentials continue to
be a hot topic, but authorization is coming on strong. In closing circle someone
said (paraphrashing) that authentication is solved and the next frontier is
authorization. I tend to agree. We should have the book of proceedings completed
in about a month and you'll be able to get the details of sessions there. You
can view past Books of Proceedings here.

As I said, there were attendees from all over the world as you can see by the
pins in the map at the top of this post. Not surprisingly, most of the attendees
were from the US (212), followed by Canada (29). Japan, the UK, and Germany
rounded out the top five with 9, 8, and 8 attendees respectively. Attendees from
India (5), Thailand (3), and Korea (3) showed IIW's diversity with attendees
from APAC. And there were 4 attendees from South America this time. Sadly, there
were no attendees from Africa again. Please remember we offer scholarships for
people from underrepresented areas, so if you'd like to come to IIW38, please
let us know. If you're working on identity, we want you there.

In terms of states and provinces, California was, unsurprisingly, first with 81.
Washington (32), British Columbia (14), Utah (11), Ontario (11) and New York
(10) rounded out the top five. Seattle (22), San Jose (15), Victoria (8), New
York (8), and Mountain View (6) were the top cities.

Doc Searls has several hundred photos from Day 1, Day 2, and Day 3 of IIW on his
Flickr account.

As always the week was great. I had a dozen important, interesting, and timely
conversations. If Closing Circle and Open Gifting are any measure, I was not
alone. IIW is where you will meet people to help you solve problems and move
your ideas forward. Please come! IIW 38 will be held April 16-18, 2024 at the
Computer History Museum. We'll have tickets available soon.

Photo Credits: Doc Searls

Tags: identity iiw


MORE
October 240
24.6.2024, 6:00SEI Blog (Feed)
The 13 Key Elements of an Insider Threat Program
COVID-19 changed the nature of the workplace. In this evolving climate,
organizations need to be increasingly vigilant against insider incidents. In
this post, we present the 13 key elements of an insider threat program.

MORE
October 240
Today, 17:55ongoing by Tim Bray (Feed)
Feminized Feed

My online social life is almost all on the Fediverse (obligatory plug for our
member-owned co-op Mastodon instance) and I find it generally satisfying. But
back in early September, I noticed that my feed was extremely male-dominated. I
set out to fix that and — a rare piece of good news — it worked! Herewith
methods and numbers.

Numbers

I wish I’d been a little more careful before I started this, as in recording the
gender ratio in my feed at a more sophisticated level than “seems to be mostly
men”. And also noting how many people I was following. Oh well, rear-view is
20/20.

Late Friday afternoon, I scrolled back an hour and a bit in my feed and
categorized 205 posts as “men”, “women”, and “neutral”. Here’s what I found.

WomenMenNeutral Posts20553.7%29.8%16.6% Accounts6653.0%30.3%16.7%

So, mission somewhat accomplished. Note that the percentages of accounts and
posts are about the same; no evidence here that either gender is more prolific.

Methods

Almost every day, I pop open my Mastodon “Notifications” tab, go through it, and
gender-check any account who’s followed or replied or liked or boosted but I
don’t already follow. Intermittently and unsystematically, I do the same for
boosted posts from people I don’t follow. If they’re female, and they’ve ever
actually posted or boosted, I follow them.

What does “gender-check” mean, exactly? I try to err on the side of inclusion.
If you offer “she/her” pronouns, OK. If your avatar is a wombat or a potato but
your name is “Cynthia”, yep. If you have neither pronouns nor a feminine account
name, I take a look at your avatar through cisgendered heterosexual male eyes
and make a judgment call. So scientific.

I’ve subsequently unfollowed a small handful of women for the usual reasons.

Findings

My feed is different than it used to be. I find that I lack the words to
describe the difference, even vaguely. But, I’ve no urge to go back. Oh, my
progressive bubble is bubblier; the Fediverse’s women are a lefty lot.

Wow, are there ever a lot of trans women here. And lesbians. But we already knew
that.

Next

I have a huge advantage because I’m a loudmouth who’s been doing social media
for decades, so my Notifications tab is rarely empty. Which means that the
people I add have already chosen to interact with me and are thus likely to
intersect with my interests. So I’m not sure this would work for the many people
who are read-mostly, or who are still building their audience.

Now I’m feeling an obligation to find a good way to help feminize others’ feeds.
But I’m nervous about over-curating; why should I think that the people who
interest me would do the same for others? Would it be crazy to, every week, pick
a dozen random female accounts that I follow and post them to a #FollowFriday?
At the moment, I can’t think of anything better.

But I do recommend that other high-interaction people give this a try. It’s easy
and if you’re like me you won’t want to go back to a mostly-male feed.


MORE
October 210
22.6.2024, 19:00Platformonomics (Feed)
Platformonomics TGIF #27: October 20, 2023

Be sure to subscribe below and to the right to receive all my output via email.

Excellent Trolling by The Information

The lede from an interview with Amazon’s generative AI product manager, Andy
Jassy:

Paragraph three:

Amazon Now Forecasts They Won’t Be Competitive In Generative AI For Over 13,000
Years

Our previous analysis concluded that if you were running a 10K and had spent a
month on the third step, it would take over 800 years to finish the race. But
we’re now in the fourth month still stuck on that third step AND the race has
been extended to a marathon.

Previous: Amazon Pauses Lectures on Generative AI, Three Steps into a 12-Step
Program, Culture Clash Conjecture: Amazon and Anthropic, AWS: “It’s Day 0.1 in
generative AI”, AWS: “We’re three rankles into a 10k rankle”, “Every Single”
Amazon PR Team is Working on Generative AI (Part I’ve Lost Count), “Every
Single” Amazon PR Team is Working on Generative AI, Titan or Titanic?, Amazon
Pioneers LLM-less Generative AI, A PR Campaign About Nothing: AWS and Generative
AI, Day Two of Amazon’s PR Campaign About Nothing, Amazon’s Relentless PR
Campaign About Nothing

Amazon’s AI Communications Clown Show: Cause or Effect? You Can’t Spell
Misinformation Without MSM

The megaliths of mainstream media deeply embarrassed themselves in the last week
by eagerly acting as unquestioning press agents for Hamas (while simultaneously
bemoaning misinformation on social media that was the corrective to the media’s
misinformation), and then digging in to defend their errors. It isn’t clear if
they’re just committed to a narrative regardless of reality and/or are just bad
at journalism. But we’ve seen this before in their tech coverage.

Previous: Breaking: New York Times Agrees with Tech Industry, New York Times
Gives Up on First Amendment?, Living in a Glass House: The New York Times On
Tech, Living in a Glass House: What is the New York Times?, The World’s Worst
Technology Column (Twitter Threads)

More Power Consumption Hysteria

To quote myself, “hyperbolic power consumption predictions have a dismal track
record” (and from that researcher in particular). We should be focused on
abundant energy, not debating what is and isn’t worthy power consumption under a
scarcity mindset. But if the BBC is worried, they could opt to shut themselves
down to save power. Turning off the BBC would also reduce misinformation, riots,
and failed diplomatic efforts.

Previous: Cloud Power Up

Useful Idiots, Fellow Travelers and Unregistered Foreign Agents: Apple

It is time to start calling out the Chinese Communist Party’s allies in the West
who act on their behalf.

Previous: Xi Jinping Thought on Wall Street Running Dogs, The Perils of Working
for a Cult of Personality, Press Releases We’d Like to See, Best Nickname for
Emmanuel Macron, Winnie-the-Pooh: Lover of Honey, Champion of Freedom, Missing
the (Bamboo) Forest for the (Apple) Trees


MORE
October 210
1.5.2024, 16:56James Governor's Monkchips (Feed)
Introducing the AI Bill of Materials

What’s in the black box? As we go forward we will need a model and machine
readable bill of materials.

It’s becoming increasingly clear that we’re going to need an AI bill of
Materials (AIBOM). Just as with a Software Bill of Materials (SBOM), there are a
set of potentially important questions when we take advantage of an AI model,
which we’re just not set up to meaningfully answer at the moment. The You Only
Live Once (YOLO) approach that OpenAI took when launching ChatGPT is not going
to provide end users with the confidence they need to adopt AI models in
strategic initiatives. Sure we’ll see shadow AI adoption (like shadow IT
adoption, but for AI tools), by marketing departments and sales and so on. The
lines of business won’t allow governance to get in the way of productivity. But
as ever bottom up needs to meet top down, which is where governance comes in.
From an enterprise perspective we will need to get a much better understanding
of the models we’re using, and what went into them. Which is to say, a bill of
materials. Trust and safety are not the enemy. We need to better understand the
AI supply chain.

The AIBOM will need to consider and clarify transparency, reproducibility,
accountability and Ethical AI considerations. After introducing the idea of an
AIBOM to Jase Bell he began to work on a schema, which you can find on GitHub
here. It’s a useful starting point for discussion so please check it out.

Transparency: Providing clarity on the tools, hardware, data sources, and
methodologies used in the development of AI systems.

Reproducibility: Offering enough information for researchers and developers to
reproduce the models and results.

Accountability: Ensuring creators and users of AI systems are aware of their
origins, components, and performance metrics.

Ethical and Responsible AI: Encouraging the documentation of training data
sources, including any synthetic data used, to ensure there’s knowledge about
potential biases, limitations, or ethical considerations.

Weighting, and decisions behind it, become ever more important. Also – origins
are not the only important issue. Sometimes the intended use case is where the
higher duty of care is required. We may want to read the ingredients on a cereal
box, but it’s not a matter of life or death. Taking medicine on the other hand,
you definitely want to know exactly what the ingredients are. So what are we
predicting or generating? The EU regulatory framework for AI discussed below
establishes a hierarchy of low or high risk AI use cases. We’ll see more of
that.

One part of the industry already groping towards the need for an AIBOM as a
business critical issue is commercial open source.There was an interesting post
recently by Matt Asay about AI and open source licensing a few days ago – Making
sure open source doesn’t fail AI. The argument is that open source licensing had
failed in the age of the cloud, and we needed to avoid making the same mistakes.
What jumped out at me however was a quote from Stefano Maffulli, executive
director of the Open Source Initiative (OSI), which certifies what is, and what
is not, open source from a licensing perspective. He said you’d need

“A very detailed description of what went into creating the artifact.”

As Asay writes

“In this world, you’d need to publish all the scripts that went into assembling
the data set, the weights that govern the LLM, the biases you bring to the
model, etc. This, in my view, is a much more interesting and useful way to think
about open source AI, but it’s also far more complicated to deliver in
practice.”

Complicated? Absolutely. But increasingly important.

In a really smart tweet Jordan Hamel makes the case that OpenAI is the Napster
of the LLM revolution.

OpenAI in some ways is the Napster of AI who learned from its mistakes by
getting MSFT on board since they’ve traveled through the legal journey of the
underworld and back and have the $$$ to make it legal. A machine learning model
that can generate copyright derivative material in a variety of modalities does
conflict with existing copyright law and one way or another it’s going to come
to a breaking point. Imagine DMCA for your generated content? Napster blew
people’s minds in the 90’s for good reason and it took well over a decade for
the legal products to exceed its quality and content distribution.

This is spot on. We’ve been shown something incredible, and we want to use it,
obviously. The great softening up has begun. And we need an AIBOM from a
business perspective. Why? The answer is, as ever, found in governance, risk,
and compliance.

Large Language Models (LLMs) could take a wrecking ball to regulated industries.
Any regulation that concerns itself with user privacy, for example, is not going
to survive first contact with LLMs. The HL7 patient data interoperability
standard wasn’t designed with the cloud in mind, let alone AI. Or think about
HIPAA, or GDPR even. So enterprises are justifiably concerned about feeding LLMs
with user data. In areas such as manufacturing, engineering or polluting
industries regulations abhor a black box, but that’s exactly what LLMs are.
Copyright infringement is another potential concern – the first class action
lawsuits have been lodged by authors against OpenAI.Then of course there is the
fear that using public LLMs might lead to security breaches and leakage of trade
secrets. Data residency continues to be a thing, and public LLMs are not set up
to support that either – if German data needs to be held in a German data center
how does that chime with a model running in a US Cloud. And how about prompt
injection as an emerging vector for security threats?

So far tech vendors have been remarkably relaxed about these fears, while
enterprises have been somewhat less confident. Google and Microsoft have
promised to indemnify users if they are challenged on copyright grounds, for
example. Their highly paid corporate lawyers are evidently pretty confident that
a fair use argument will prevail in court, for outputs from AI models. As ever
this is a question about tolerance for risk.

And sometimes the promises about trust don’t stand up to scrutiny. Thus for
example Adobe said it’s Firefly image generation model was “commercially safe”
because artists had signed up for it – this was [apparently a surprise to some
content creators. Adobe, however, has continued pushing its trust credentials
with the introduction of a “made by” symbol for digital images, establishing
provenance, including, for example, if it was made with AI tools.

The EU is moving towards some far-reaching (some might argue over-reaching)
requirements around model-training with its coming EU AI Act

Some notable statements in the positioning document. Considerations of “high
risk” (All high-risk AI systems will be assessed before being put on the market
and also throughout their lifecycle.)

include:

AI systems that are used in products falling under the EU’s product safety
legislation. This includes toys, aviation, cars, medical devices and lifts
(elevators).

Meanwhile, here is the copyright kicker.

Publishing summaries of copyrighted data used for training

Perhaps we can just ignore EU law. A lot of folks consider the GDPR to be more
of an annoyance than anything else. Facebook can easily afford to pay its $1.3bn
fine – that’s just a cost of doing business, right? The US has replied with
sabre rattling that regulation will only serve to entrench the major players.

Some companies might feel confident in ignoring EU law – YOLO – but if they want
to do business in China, that’s not really an option that’s open. This thread
from Matthew Sheehan is essential reading for anyone interested in AI
regulation, or the lack of it. Also this post. The TDLR – China is literally
years ahead on AI regulation. In China at least:

The draft standard says if you’re building on top of a foundation model, that
model must be registered w/ gov. So no building public-facing genAI applications
using unregistered foundation models.

So that’s certainly a potential future. China has an AIBOM-like requirement and
policies and procedures and corporate responsibilities that go with it. We’re
all going to have to think through this stuff – Norway just announced a Minister
for Digitalisation and Governance with Responsibility for Artificial
Intelligence, Karianna Tung.

According to Norwegian Prime Minister Jonas Gahr Støre:

Artificial intelligence offers enormous opportunities, but requires knowledge,
management and regulation. Because it must still be the people who determine the
development of technology, not the technology that controls the people.

Sentiments that I agree with. And regulation – that’s going to need an AIBOM.
Major vendors are talking a lot about trust and AI, and jostling for market
positioning accordingly – again, this is where an AIBOM is going to come into
play.

 

disclosure : Adobe is a RedMonk client. OpenAI is not.


MORE
October 190
Today, 17:55ongoing by Tim Bray (Feed)
Rock Tech

Last night, my son and I took in Guns N’ Roses’ Vancouver show at “B.C. Place”,
the big football/soccer stadium, its roof closed for the event. It was a fine
show, even though I wasn’t and still am not much of a GnR fan. The show deserves
a few words, and also arena-rock concert technology is changing in interesting
ways.

Enlarge to see Slash.

The band

I impulse-bought the tickets when the show was announced. G&R have real cultural
heft and I’d never seen them, seemed like a good enough reason.

There are things to like: The songs are good. And, I learned, they put on a
long, polished, high-intensity show, really working hard to show the fans a good
time.

Here’s the set list. Is that choice of covers weird or what? I suffered
cognitive dissonance watching fresh-faced high-school girls, in
carefully-constructed trashy-rock-chick black leather, howling gleefully along
to Knockin’ on Heaven’s Door, a 50-year-old song written by an 82-year old being
performed by a 61-year-old.

But G&R has never really done it for me; I find Axl’s vocal histrionics
frequently grating. Slash is flashy and has good riffs, but all of his solos
sound like all of his other solos to me. The lyrics are often gratuitously nasty
and ugly. Axl Rose has historically not been a good person, although in 2018 I
was amused on Twitter at him going off his bad-boy brand.

Tech and staging

In my youth, I spent a few early years as the house stage manager for a few rock
venues and got pretty familiar with the way things were done. In recent years,
my musical tastes have become more niche-y, with the pleasant side effect that I
attend concerts in clubs and small theatres. So I’ve been out of touch with
Arena Rock.

Back in the day, in big venues, the show depended heavily on the use of monster
spotlights like for example the Super Trouper. You might have been way back from
the stage, but when someone was going to belt out a crucial lyric or run down
the front of the stage to shred, the spots would light ’em up, obvious where to
look.

No longer. At the show last night (see the photo above) the musicians were
well-enough lit, but lacked that insane Super-Trouper glow. Which didn’t matter,
because of the huge video screens highlighting whoever is currently the key
performer. If you look close in that picture, you can see Slash on a little
downstage-center platform, about as well as I could from where I was standing.
So most of the people at the show are watching a live TV broadcast most of the
time. Except when they’re looking at their phones which they’re using to record
the broadcast.

I shouldn’t be so negative. I mean, it works, you get to see facial expressions
and guitar moves and so on that you never would unless you were at the very
front of the hall. And playing in front of Super Troupers was notoriously hard
on the eyes, Abba even wrote a song about it: “Super Trouper beams are gonna
blind me…”

Interestingly, G&R didn’t project themselves on the huge backing backing screen;
that was a continuous video presentation, nicely correlated with the music, with
an abstract sci-fi/horror aesthetic. Worked for me.

The sound

A tip o’ the hat to whoever in the G&R team hired the audio crew. They got
sparkling sound in B.C. Place, which is not exactly a regular occurrence.

In particular, they changed the mix all the time. Every time there was a big
guitar break, the instrument was unsubtly boosted for effect, and it was a good
effect. Likewise for vocals and other instruments.

Now, those vocals. Axl Rose was getting a lot of unsubtle technology support for
his singing. He would reach down to a device on his belt and adjust himself
regularly between sections of a song, notably when switching between his normal
mid-range voice and that trademarked Axl shriek. Possibly he’s just lost the
high register and is having it added back synthetically? Also there were
occasional shudder-and-buzz effects added unsubtly.

Sometimes it worked — I mean, you just can’t do Welcome To The Jungle without
going to a glass-shattering yowl. Sometimes it sounded fakey. A couple of times
it just didn’t work, Axl’s lips were moving but no sound was coming out.

Anyhow, glad I went, I loved the crowd, the youngsters in their rock-n-roll
finery and the grizzled middle-agers in basic black. Parents with kids. Couples
mooning, melting into each other during November Rain. The band does a great job
on Live and Let Die. They worked hard to show me a good time and they did. Won’t
go again.


MORE
October 180
24.6.2024, 6:00SEI Blog (Feed)
Assuring Cyber-Physical Systems in an Age of Rising Autonomy
This post discusses research by the SEI and Georgia Tech to address a number of
challenges related to the rising volume of sensor data in autonomous
cyber-physical systems.

MORE
October 160
22.6.2024, 19:00Platformonomics (Feed)
Platformonomics TGIF #26: October 13, 2023

The snark must flow, but not this week…

A Message Not Just For Israelis…

MORE
October 140
1.5.2024, 16:56James Governor's Monkchips (Feed)
Postgres: the next generation. Investing in the next generation of committers.

PostgreSQL isn’t getting any younger. Which is fine – after all, databases
generally improve with age. The platform is going from strength, and is a
default choice for a big chunk of modern software development. But Postgres has
been around for a while – it launched in 1986 – which has an implication for the
folks actually building the database. Just how long will they want to do the
heavy lifting of maintaining a high profile codebase that so many folks rely on?
Postgres is a close knit group and project. Robert Haas, Postgres committer and
chief database scientist at EnterpriseDB writes a regular contribution post and
the latest numbers are salutary – Who Contributed to PostgreSQL Development in
2022?

I calculate that, in 2022, there were 192 people who were the principal author
of at least one PostgreSQL commit. 66% of the new lines of code were contributed
by one of 14 people, and 90% of the new lines of code were contributed by one of
40 people.

The core development community is aging somewhat – the average age is probably
around 50. Which is totally fine. 50 year olds are more than capable of doing a
shitload of work – don’t ask me how I know. Tom Lane, who works at Crunchy Data,
is 68 and he’s still the Postgres project’s fulcrum. Long may that continue.

The Postgres community is amazing. Open Postgres governance is something we can
and do rely on, which is refreshing in the current era of commercial open source
licensing rugpulls. But as an axis to consider in terms of open source
sustainability let’s assume that Postgres is still going strong in say, 20
years. Who is going to be doing the work in 2043? I had a fascinating
conversation with Nikita Shamgunov, CEO of Neon recently and one of the subjects
we discussed was aging in tech projects and its relationship to project
sustainability. Neon is a fully managed Postgres database optimised for
serverless apps, separating storage from compute – the database is just a URL.
That’s the design principle. It allows for branching, with preview deployments –
thus Neon’s partnership with Vercel. Make it easy, make it modern, make it a
zero config API. Neon has 62 employees and has raised $108m so far. It competes
with the likes of Supabase. But back to the subject at hand.

According to Shamgunov:

If you look at the Postgres committer crowd they’re in their 50s, 60s, 40s and
maybe a few in their 30s. It takes a lot of effort to become a committer but
very little to be a contributor – you just need to write good code.

I think we’re doing good to the world by hiring more junior people and training
them to become committers and hopefully maintainers. It’s very important that
the Postgres engine continues to evolve.

Neon is being intentional about investing in the next generation of
contributors, committers and maintainers. The natural move for a lot of
companies is to try and hire the existing top talent, rather than fostering new
blood.

We debated whether to just find more Postgres committers and hire them. But it’s
not clear that would be spending our money in the best way. If we train new ones
it’s better, and that’s how we can keep ramping the Postgres team.

There are some interesting questions here. For example – consider Neon’s IP,
which is currently permissively licensed, but Shamgunov is not an open source
zealot. What happens if in a few years the company decides to relicense, as
other database companies have – see for example Redis, MongoDB and Elastic. Neon
would be perfectly within its rights to relicense under more restrictive terms,
potential community blowback aside. But any code they had contributed to
Postgres? That’s not going to be affected. Having core Postgres maintainers on
staff is a pretty good example of enlightened self interest and should serve to
keep the company honest. Whatever decisions Neon makes in future, assuming they
have employees dedicated to making Postgres better, then the community and core
codebase still wins.

Cohort aging is certainly not a problem that’s unique to Postgres. Anyone
remember the year 2000 bug? Communities and ecosystems do get older, which can
be an issue when it comes to skills and staffing and rejuvenation. IBM has done
a great job of bringing younger developers into the mainframe fold, for example
with vocational education programs at universities – here is a post I wrote
about that a while back.

There are plenty of projects with literally millions of users that are run by
one or two people and don’t have the level of corporate sponsorship with see
with projects such as Postgres or even Kubernetes. Postgres isn’t in any sense
struggling to attract new users – there are plenty of 22 year olds defaulting to
it today. It’s a hugely popular platform. But yes, ensuring the ongoing
maintenance of the project will require some intentionality, funding, and
enlightened self interest.

 

Disclosure: Neon is not a RedMonk client. Crunchy Data, IBM and Vercel are all
RedMonk clients. This piece is published independently of any client
relationships.

The illustration above was created with Midjourney.


MORE
October 100
24.6.2024, 6:00SEI Blog (Feed)
Secure by Design at CERT
The national cybersecurity strategy calls on tech providers to ensure that all
their products are secure by design and secure by default. This post highlights
the SEI CERT Division's continued and longstanding efforts to ensure security by
design in fielded software.

MORE
October 90