Retired since 2012.
2337 stories
·
5 followers

The Strangers’ Case

1 Comment
I don’t normally say this, but if you watch one thing on kottke.org today, this week, this month, make it this speech written by Shakespeare and performed by Sir Ian McKellen on The Late Show with Stephen Colbert. The segment starts at ~20:00; McKellen sets it up:

It’s all happening 400 years ago. In London, there’s a riot happening. There’s a mob out in the streets and they’re complaining about the the presence of strangers in London, by which they mean the recent immigrants who’ve arrived there. And they’re shouting the odds and complaining and saying that the immigrants should be sent back home wherever they came from. And the authorities send out this young lawyer, Thomas Moore, to put down the riot, which he does in two ways. One by saying that you can’t riot like this. It’s against the law. So, shut up, be quiet. And also, being by Shakespeare, with an appeal to their humanity.

The riot took place on May 1, 1517 and is referred to as Evil May Day:

According to the chronicler Edward Hall (c. 1498–1547), a fortnight before the riot an inflammatory xenophobic speech was made on Easter Tuesday by a preacher known as “Dr Bell” at St. Paul’s Cross at the instigation of John Lincoln, a broker. Bell accused immigrants of stealing jobs from English workers and of “eat[ing] the bread from poor fatherless children”.

The same as it ever was. The text of the play, Sir Thomas More, is available at Project Gutenberg; here are the bits that McKellan performed, after the crowd calls for the removal of the strangers (some translation help, if you need it):

Grant them removed, and grant that this your noise
Hath chid down all the majesty of England;
Imagine that you see the wretched strangers,
Their babies at their backs and their poor luggage,
Plodding to th’ ports and costs for transportation,
And that you sit as kings in your desires,
Authority quite silent by your brawl,
And you in ruff of your opinions clothed;
What had you got? I’ll tell you: you had taught
How insolence and strong hand should prevail,
How order should be quelled; and by this pattern
Not one of you should live an aged man,
For other ruffians, as their fancies wrought,
With self same hand, self reasons, and self right,
Would shark on you, and men like ravenous fishes
Would feed on one another.

You’ll put down strangers,
Kill them, cut their throats, possess their houses,
And lead the majesty of law in line,
To slip him like a hound. Say now the king
(As he is clement, if th’ offender mourn)
Should so much come to short of your great trespass
As but to banish you, whether would you go?
What country, by the nature of your error,
Should give you harbor? go you to France or Flanders,
To any German province, to Spain or Portugal,
Nay, any where that not adheres to England,—
Why, you must needs be strangers: would you be pleased
To find a nation of such barbarous temper,
That, breaking out in hideous violence,
Would not afford you an abode on earth,
Whet their detested knives against your throats,
Spurn you like dogs, and like as if that God
Owed not nor made not you, nor that the claimants
Were not all appropriate to your comforts,
But chartered unto them, what would you think
To be thus used? this is the strangers case;
And this your mountainish inhumanity.

And of course, McKellen performs this wonderfully — he originated the role and has been performing it since the 1960s. Again…I urge you to watch it.

Tags: Ian McKellen · immigration · politics · Sir Thomas More · TV · video · William Shakespeare

💬 Join the discussion on kottke.org

Read the whole story
cjheinz
3 hours ago
reply
Todo.
Lexington, KY; Naples, FL
Share this story
Delete

Minimalistic City Map Posters

1 Comment
This Github project from Ankur Gupta allows you to “generate beautiful, minimalist map posters for any city in the world”. There are a variety of different themes you can choose from and the resulting images are big enough to print out actual posters (20-inch height maximum). You can install the Python scripts on your computer or use this website (which seems quite slow). Also, I wonder if the height/width minimums can be changed to output bigger posters?

Tags: Ankur Gupta · design · maps

💬 Join the discussion on kottke.org

Read the whole story
cjheinz
1 day ago
reply
This looks cool.
I luv maps!
Lexington, KY; Naples, FL
Share this story
Delete

Seu Jorge’s Lovely Tribute to David Bowie

1 Comment
For his 2004 film The Life Aquatic with Steve Zissou, Wes Anderson enlisted Brazilian musical artist Seu Jorge to perform several of David Bowie’s songs in Portuguese. Jorge released an album of the songs about a year or so later. A few weeks ago, to mark the 10th anniversary of Bowie’s death, Jorge released a hour-long set of him performing those songs: Just an acoustic guitar, a microphone, and the beautiful coastline of São Paulo.

Tags: movies · music · Seu Jorge · The Life Aquatic · video · Wes Anderson

💬 Join the discussion on kottke.org

Read the whole story
cjheinz
2 days ago
reply
Check it out n
Lexington, KY; Naples, FL
Share this story
Delete

Yvonnick Prené • Un Harmonia Pour Django

1 Comment
The only time that Django Reinhardt recorded with a harmonica player was on May 31, 1938, when Larry Adler was joined by the Quintet of the Hot Club of France for four songs. When he was ten, Yvonnick Prené heard the Django-Adler session. A virtuosic master of the harmonica from France, Prené is usually heard […]
Read the whole story
cjheinz
4 days ago
reply
Here's the 4 tracks on YouTube:

https://youtu.be/Epq8siKTg1c
https://youtu.be/wYaJIjfD-eA
https://youtu.be/ViO_DUbAZak
https://youtu.be/CPHc-k_R0_A
Lexington, KY; Naples, FL
Share this story
Delete

Kurt Vonnegut on the Simplest, Hardest Secret of Happiness

1 Share

The meaning of life, in a short verse.


Kurt Vonnegut on the Simplest, Hardest Secret of Happiness

“Don’t make stuff because you want to make money — it will never make you enough money. And don’t make stuff because you want to get famous — because you will never feel famous enough,” John Green advised aspiring writers. “If you worship money and things … then you will never have enough. Never feel you have enough. It’s the truth,” David Foster Wallace admonished in his timeless commencement address on the meaning of life. But what does it really mean to “have enough?”

There is hardly a better answer than the one implicitly given by Kurt Vonnegutman of discipline, champion of literary style, modern sage, one wise dad — in a poem he wrote for The New Yorker in May of 2005, reprinted in John C. Bogle’s Enough: True Measures of Money, Business, and Life (public library):

JOE HELLER

True story, Word of Honor:
Joseph Heller, an important and funny writer
now dead,
and I were at a party given by a billionaire
on Shelter Island.

I said, “Joe, how does it make you feel
to know that our host only yesterday
may have made more money
than your novel ‘Catch-22’
has earned in its entire history?”
And Joe said, “I’ve got something he can never have.”
And I said, “What on earth could that be, Joe?”
And Joe said, “The knowledge that I’ve got enough.”
Not bad! Rest in peace!

Complement with Vonnegut on how to write with style, the writer’s responsibility and the limitations of the brain, the shapes of stories, his daily routine, his heart-warming advice to his children, and his favorite erotic illustrations.


donating = loving

For seventeen years, I have been spending hundreds of hours and thousands of dollars each month composing The Marginalian (which bore the outgrown name Brain Pickings for its first fifteen years). It has remained free and ad-free and alive thanks to patronage from readers. I have no staff, no interns, no assistant — a thoroughly one-woman labor of love that is also my life and my livelihood. If this labor makes your own life more livable in any way, please consider lending a helping hand with a donation. Your support makes all the difference.


newsletter

The Marginalian has a free weekly newsletter. It comes out on Sundays and offers the week’s most inspiring reading. Here’s what to expect. Like? Sign up.

Read the whole story
cjheinz
4 days ago
reply
Lexington, KY; Naples, FL
Share this story
Delete

AI in government: From tools to transformation

1 Comment

Transformation depends on institutions

Artificial intelligence offers potential for governmental transformation, but like all emerging technologies, it can only catalyze meaningful change when paired with effective operating models. Without this foundation, AI risks amplifying existing government inefficiencies rather than delivering breakthroughs.

The primary barrier to AI-based breakthroughs is not an agency’s interest in adopting new tools but the structures and habits of government itself, particularly excessive risk management; rigid hierarchies; and organizational silos rather than adaptive problem solving and effective service delivery. Structural reform is critical and must accompany adoption of AI. 

Defining the tools: Generative AI and Agentic AI

Many types of AI have already permeated daily life and government operations, including predictive models, workflow automation tools, and computer vision. Two relatively new categories, generative AI and agentic AI, have been attracting the most attention lately.

Generative AI: Generative AI systems produce new content or structured outputs by learning patterns from existing data. These include large language and multimodal models that generate text, images, audio, or video in response to user prompts. Examples include tools built on models such as OpenAI’s ChatGPT, Anthropic’s Claude, and Google’s Gemini. In government, generative AI can accelerate sensemaking by summarizing documents, drafting text, generating code, and transforming unstructured information into structured formats. These capabilities are increasingly embedded in enterprise software platforms already authorized for government use, lowering barriers to experimentation and adoption.

Agentic AI: Agentic AI systems go beyond producing analysis or recommendations by autonomously taking actions to achieve specified goals. They can coordinate multistep workflows, integrate information across organizational silos, monitor conditions in real time, and execute predefined actions within established guardrails. Agentic systems often rely on generative AI as a component — for example, to synthesize information, draft communications, or generate code as part of a larger autonomous workflow. The transformative potential of agentic AI lies in augmenting an organization’s capacity to identify issues, make coordinated decisions, and implement solutions at scale, while maintaining human oversight to ensure alignment with legal, ethical, and strategic requirements.

For either to deliver value, leaders and managers must support execution with clearly defined goals and effective governance, keeping humans firmly in control of constraints, decision making, and accountability.

The risks of misguided modernization

Public debate about AI often polarizes around future hypotheticals — on one side, AI as an existential threat; on the other, AI as an inevitable, world-transforming revolution. Both frames focus on risks and capabilities that are not yet present in available systems. In reality, there is a pragmatic middle path that addresses today’s challenges: strengthening institutions, improving decision making capacity, and integrating AI only where it can reliably support public missions rather than distort them. As with all emerging technologies, the benefits of AI will depend less on the specific tools and far more on the governance frameworks, incentives, and organizational capacities that shape how those tools are actually used.

Many government leaders today face pressure to “use AI,” often without a clear understanding of what that entails or what problems it is meant to solve. This can lead to projects or procurements that are technology first rather than problem first, generating little real value for the agency or the public they serve. It is critical that leaders remain grounded in the reality of what current AI systems can and cannot deliver, as the vendor marketplace often oversells capabilities to drive adoption.

A more effective approach starts by asking, “What challenges are we trying to address?” or “Where do we encounter recurring bottlenecks or inefficiencies?” Only after defining the problem should agencies consider whether and how AI can augment human work to tackle those challenges. Anchoring adoption in clear objectives ensures that AI serves public missions rather than being implemented for its own sake.

Technology waves of the past offer lessons for the present and point to paths for the future. Each wave arrived with outsized promises, diffuse fears, and a tendency for early decisions to harden into long-term constraints. During the first major wave of cloud adoption, for instance, agencies made uneven choices about vendors, security boundaries, and data architecture. Many became locked into single-vendor ecosystems that still limit portability and flexibility today. 

The most successful reforms did not come from chasing cloud technologies themselves, but from strengthening the systems that shaped their adoption: governance, procurement, workforce skills, and iterative modernization practices. Progress accelerated only when agencies implemented clearer shared-responsibility models, modular cloud contracts, robust architecture and security training, and test-and-learn migration frameworks.

The basics of state capacity

AI is no different. The challenge now is to avoid repeating old patterns of overreaction or abdication and instead build the durable institutional capacity needed to integrate new tools responsibly, selectively, and in service of public purpose.

To navigate this moment effectively, government should focus less on the novelty of AI and more on the institutional choices that will shape its long-term impact. That shift in perspective reveals a set of practical risks that have little to do with model architectures and everything to do with how the public sector acquires, governs, and deploys technology. These risks are familiar from previous modernization waves — patterns in which structural constraints, vendor incentives, and fragmented decision making, if left unaddressed, can undermine even the most promising tools. This is why the basics of state capacity are so important.

Understanding these universal dynamics is essential before turning to the specific challenges AI introduces.

  • The vendor trap: Modernization efforts often risk simply moving outdated systems onto newer but still decades-old technology while labeling the work with the current buzzword, whether it be “mobile,” “cloud” or “AI.” Vendors actively market legacy-to-cloud migrations or modernized rules engines as “AI-ready,” enriching their businesses without delivering transformational change. 
  • Procurement challenges: Because most federal IT work is outsourced and budgeting rules often prevent agencies from retaining savings, taxpayers frequently see little benefit from cost efficiencies. Large project budgets with spending deadlines and penalties under the Impoundment Control Act, for example, incentivize agencies to spend the full amount regardless of actual costs. Previous technology waves, such as cloud adoption, demonstrated the same pattern: declining infrastructure costs rarely translated into government savings. 
  • The structural constraint: Adopting AI tools for processes such as claims or fraud detection will only yield limited, incremental efficiencies if such core issues as mandatory multioffice approvals, legacy systems, and paper documentation remain. AI accelerates part of the workflow, but the overarching, inefficient structure remains a hard limit on overall impact.
  • Strategic control: Because most federal IT work is outsourced, there is a tendency to also outsource the framing of the problem technology should solve. Such vendor framing prioritizes vendor benefit. It is imperative that the government and not vendors frame the problems that AI should solve, prioritizing public benefit. This requires hiring, retaining, and empowering civil servants with AI expertise to ensure that outcomes-based procurement aligns with public interests.
  • High cost, low impact for taxpayers: The ultimate consequence is that taxpayers often pay for costly projects that yield limited public benefit. Without structural reforms — clarity on goals, strengthened governance, and outcome-focused procurement, for example — AI adoption risks funneling value to vendors rather than serving the public interest.

Fail to scale

Even well-intentioned AI deployments can create “faster inefficiency” if agencies ignore structural, procedural, and governance constraints. Tools that accelerate individual tasks may produce measurable gains in isolated workflows, but without addressing the broader organizational bottlenecks, any gains will fail to scale. For example, the Office of Personnel Management could deploy an AI tool to extract data from scanned retirement files. But if the underlying paper records stored underground in Boyers, Pennsylvania, still require manual retrieval and physical handling, the overall processing time will barely improve. In effect, AI can make inefficient systems move more quickly, exacerbating friction rather than removing it. Recognizing this risk underscores why modernization must combine technology, government capacity, and deliberate institutional reform: Only by aligning tools, processes, and incentives can AI generate real improvements in service delivery and operational effectiveness.

Some argue that AI, and particularly artificial general intelligence (AGI), will soon become so capable that human guidance and governance will be largely unnecessary. In this view, institutions would no longer need to define problems, structure processes, or exercise judgment; governments could simply “feed everything into the system,” specify desired outcomes, and allow AI to determine how to achieve them. If this were plausible, it would raise a fundamental question: Does AI obviate the need to fix the underlying institutions of government, or does it require designing an entirely new system from scratch?

In practice, we are very far from this reality, especially at the scale and complexity of government. Even defining what “everything” means in a whole-of-government context is a formidable challenge, let alone securing access to, governing, and integrating the hundreds of thousands of data sources, systems, legal authorities, and operational constraints involved. These are not primarily AI problems; they are large-organization problems rooted in fragmentation, ownership, security, and accountability. This helps explain why some leaders today are tempted to mandate the unification of “all the data” or “all the systems” as a prerequisite for AI adoption. Such approaches are not only operationally infeasible and insecure, but they are also inconsistent with how effective large-scale systems are built in the private sector, which relies on modularity, interfaces, and clear boundaries rather than wholesale consolidation.

Nor does AI require governments to preemptively design an entirely new institutional model. AI is not going to “change everything” overnight. Its impact will be uneven, incremental, and highly dependent on existing structures, incentives, and governance. The more realistic and more effective path forward is to strengthen the fundamentals of government: clarify goals, modernize operating models, improve data governance, and build the capacity to experiment and learn. AI can meaningfully augment these efforts, but it cannot substitute for them. Institutions that are already capable of defining problems, coordinating action, and exercising accountability will be best positioned to benefit from AI; those that are not will simply automate their dysfunction more quickly.

The path forward: Structural reform and pragmatic experimentation

Artificial intelligence has real potential to improve how governments understand problems, coordinate across silos, and deliver services. Emerging examples already demonstrate how AI can augment public decision-making and situational awareness. Experiments highlighted by the AI Objectives Institute show how AI can help governments reason at scale, surface insights faster, and explore tradeoffs before acting. Examples include Talk to the City, an open-source tool that synthesizes large-scale citizen feedback in near real time; AI Supply Chain Observatory, which detects systemic risks and bottlenecks; and moral learning models that test policy options against ethical frameworks.

Agentic AI also holds promise for improving how the public interacts with government. During disaster recovery, for example, individuals must navigate FEMA, HUD, SBA, and state programs, each with distinct rules, portals, and documentation. A trusted agentic assistant, operating on the user’s behalf, could help citizens reuse information across applications, track status, flag missing documents, and explain requirements in plain language, reducing friction without requiring immediate modernization of every backend system. These kinds of user-centered applications illustrate the genuine upside of AI when applied thoughtfully.

At the same time, governments should be cautious about assuming that today’s AI systems are ready to deliver fully autonomous, self-composing systems. While the The Agentic State white paper articulates an important long-term aspiration — governments that are more proactive, adaptive, and outcomes-oriented — current institutional and technical realities impose hard limits. Agentic AI systems can only act within the trust, permissions, and constraints granted to them, which in government are shaped by security requirements, CIO oversight, IT regulations such as the Federal Information Security Modernization Act (FISMA), and legal accountability frameworks.

Similarly, while rapid AI-enabled prototyping is valuable, the idea that governments can rely on just-in-time, dynamically generated interfaces is unrealistic in the near term. Such approaches assume highly reliable, error-free, well-integrated backend systems — an assumption that’s rarely true in real production systems. In practice, proliferating interfaces also proliferate failure modes, increase QA and monitoring costs, and create operational risk. Even large private-sector companies struggle to manage this complexity. In government, these risks are compounded by stricter availability requirements, legal accountability, and public trust obligations. Governments should not take on greater operational or financial risk than the private sector, particularly for core services.

Given both the promise and the constraints, the most credible path forward is disciplined, well-governed experimentation grounded in structural reform. Transformative impact requires pairing AI adoption with effective operating models. The Product Operating Model (POM) provides a useful foundation: cross-functional teams, user-centered design, continuous iteration, and — critically — rigorous problem definition before selecting solutions. This ensures AI is applied where it genuinely improves outcomes, rather than amplifying existing inefficiencies.

Even modest agentic or AI-enabled systems require three fundamentals:

  • Clear objectives: Explicit definition of the problem, desired outcomes, and boundaries of success. Without this, systems may optimize the wrong goals or produce conflicting results.
  • Guardrails and constraints: Technical, procedural, and policy mechanisms like permissions, monitoring, escalation protocols, access controls that ensure compliance and safety.
  • Human governance: Ongoing oversight to handle legal, ethical, and strategic judgment that AI cannot replace.

Governments should begin with scoped pilots in lower-risk workflows, such as data integration and cleaning, summarization and search, drafting planning documents (e.g., RFPs), workflow orchestration and status tracking, customer service support, system diagnostics and performance monitoring, and test-and-learn policy simulations. These use cases allow agencies to build institutional muscle, evaluate governance mechanisms, and learn where AI adds value — without overcommitting or over-automating.

Throughout this process, strong data governance is essential. Inaccurate, incomplete, or poorly labeled data will yield flawed outputs, amplifying errors rather than insight. “Garbage in, garbage out” is a central risk. Clear mandates, constraints, monitoring, and escalation procedures developed by cross-functional teams spanning policy, legal, program, and engineering are what allow experimentation to remain safe, compliant, and aligned with public goals.

AI can do powerful and genuinely useful things for government — but not automatically, and not all at once. The way forward is not sweeping autonomy, but careful institutional reform paired with pragmatic, well-governed experimentation that builds trust, capability, and impact over time.

Summary

AI offers an opportunity to transform government, but its impact depends less on the technology itself and more on how institutions use it. Generative and agentic AI can accelerate analysis and automate complex workflows, yet without structural reforms, clear objectives, human oversight, and guardrails, these tools risk amplifying inefficiency rather than solving it.

The most effective approach is thoughtful, pragmatic experimentation: starting with well-scoped, low-risk pilots while continuously monitoring performance, refining governance, and ensuring alignment with public goals. Leaders should empower teams that can define problems, safely run pilots, and leverage data, tools, and governance to measure, iterate, and scale. Pilot opportunities should involve tractable, visible, high-value workflows with manageable risk and strong data, where improvements can be measured and scaled to inform future AI adoption. Example areas include data integration, workflow coordination, and customer service.

Ultimately, AI’s value to government emerges not from the tools themselves but from redesigning processes, reducing structural bottlenecks, and improving outcomes for both government and the public. Governments that combine technology with deliberate institutional reform and iterative learning will capture the transformative potential of AI while minimizing the risks of misguided modernization.

Thank you for feedback on this article: Alexander Macgillivray, Nicole Wong, Anil Dewan.

The post AI in government: From tools to transformation first appeared on Niskanen Center.

The post AI in government: From tools to transformation appeared first on Niskanen Center.

Read the whole story
cjheinz
7 days ago
reply
Generative AI & Agentic AI are both completely untrustworthy technologies at this point.
Why are we even talking about this???
Lexington, KY; Naples, FL
Share this story
Delete
Next Page of Stories