Three developments. One day. Zero federal law in sight.
April 13, 2026 was not a landmark moment for AI regulation in the way that a Supreme Court decision or a congressional vote would be. But it was a useful day for anyone trying to understand where AI governance is actually happening right now, and what the accumulating pattern means for organizations that need a compliance posture across jurisdictions.
Section 1: The Pattern
The three developments that landed on the same day reflect different stages of the same underlying dynamic: in the absence of a U.S. federal AI law, actors at every level are filling the vacuum.
Virginia Governor Abigail Spanberger signed SB 384 and HB 797 into law, directing the Joint Commission on Technology and Science to evaluate the feasibility of an Independent Verification Organization framework for AI safety. The bill creates a study mandate, not an IVO system, but it gives the concept its first statutory foothold in any U.S. state.
Louisiana’s HB 459, introduced by Representative Mandie Landry, advanced out of the House Governmental Affairs Committee. The bill would require political advertisements using AI-generated candidate images or likenesses to carry disclosures. It’s a proposed bill, not enacted law. But its committee advancement signals active legislative momentum in a state that has not previously been a leader in AI policy.
Then there’s Japan. Its AI Promotion Act, reportedly in effect since May 2025, according to legal analysis of the legislation, operates on a different premise entirely. Multiple independent legal analysts, including Bird & Bird and the International Bar Association, characterize it as a soft-law instrument: principles-based, non-prescriptive, with no enforceable duties attached to non-compliance. Japan’s government has signaled an ambition to position the country as a leading destination for AI development. That ambition shaped a law that looks nothing like what Virginia is studying or what Louisiana is proposing.
None of these three actions requires action from most organizations today. That’s exactly what makes the pattern worth studying.
Section 2: Three Models, Three Philosophies
These three developments don’t just represent different geographic jurisdictions. They represent three structurally distinct models for governing AI, and understanding the differences is the foundation of any multi-jurisdictional compliance strategy.
| Jurisdiction | Model | Current Status | What It Requires (Now) | Who Is Affected |
|---|---|---|---|---|
| Virginia (USA) | Safety-verification (IVO) | Study authorized (JCOTS evaluation) | Nothing yet, study mandate only | AI developers and deployers (future obligation if IVO enacted) |
| Louisiana (USA) | Disclosure | Proposed, committee advancement | Nothing yet, bill not passed | Political campaigns, media buyers, AI content vendors |
| Japan | Promotion-first, soft law | In effect (reportedly since May 2025) | Voluntary adherence to principles; no prescriptive duties per current analysis | All AI developers and deployers with Japan operations |
Virginia’s IVO model is architecturally ambitious. The concept, developed by Fathom, the organization advocating for the framework, would have state governments set outcome-based safety goals and authorize a marketplace of independent verifiers to assess AI products against those standards. According to Fathom’s materials, the model would create a compliance pathway that doesn’t require a prescriptive federal law, it works through market mechanisms and state-set goals. Whether that produces rigorous safety verification or a softer compliance marketplace depends on how goals get defined, a question JCOTS will now examine.
Louisiana’s disclosure model is simpler and more politically tractable. Behavioral requirements, you must label AI content, are easier to draft, easier to explain, and easier to enforce than capability-based restrictions. The tradeoff is that disclosure rules only work if “AI-generated” has a stable, enforceable definition. That definition is contested across jurisdictions and has proven difficult to pin down in bill language.
Japan’s soft-law approach represents the least restrictive end of the spectrum. Voluntary compliance and government-as-enabler framing make Japan’s current framework the closest thing to a global “innovation permissive” model. The EU’s framework sits at the other end: precautionary, rights-based, and prescriptive. The U.S. federal government remains absent from this spectrum entirely.
Section 3: The Fragmentation Problem
For an organization operating across U.S. states and internationally, the compliance exposure isn’t any single law. It’s the combinatorial complexity of multiple jurisdictions running different models on different timelines with different enforcement mechanisms.
This is not a new observation. The existing “Federal vs. State AI Authority: Two Tracks, One Compliance Problem” brief on this hub lays out the structural tension. What April 13 adds is three concrete data points from a single day. Virginia is experimenting with a market-based verification model. Louisiana is pushing a disclosure-first approach. Japan is running a promotion-first alternative to both. And the EU’s AI Act, covered in depth on the EU AI Act hub page, continues moving toward full implementation on its own timeline.
A company building and deploying AI systems across these jurisdictions doesn’t face one compliance question. It faces a matrix: What does each jurisdiction require? From whom? On what timeline? With what documentation? The answer differs by jurisdiction, by AI use case, and by how each framework defines its scope. Building a compliance architecture that can flex across that matrix, rather than treating each jurisdiction as a one-time project, is the design challenge that emerges from this pattern.
Section 4: What to Watch
The forward-looking items here are specific and trackable.
Virginia’s JCOTS will conduct an IVO feasibility study. The study’s findings and timeline will determine whether Virginia introduces actual IVO legislation in a future session. Organizations developing or deploying AI systems should monitor JCOTS outputs and track whether other states issue similar study directives. The IVO concept has also surfaced in Minnesota’s legislature, Virginia’s action may accelerate that process.
Louisiana’s HB 459 continues through the legislative process. Committee advancement is early-stage. The bill may be amended, stalled, or fail on a floor vote. Political campaigns, media buyers, and AI content tool vendors with Louisiana exposure should track the bill’s status through the current session. If passed and signed, disclosure obligations would apply to AI-generated political advertising in the state.
Japan’s Basic AI Plan, reportedly approved in late 2025 according to legal trackers, represents the operational layer of the AI Promotion Act. Companies with Japan operations should monitor guidance issued under the Plan, even where no binding obligation currently attaches, voluntary guidance has a way of hardening into expectation over time.
Section 5: Compliance Takeaway
The practical implication of this pattern is not that organizations need to act on any of these three developments today. It’s that a reactive compliance posture, one that responds to laws after they’re enacted, is increasingly inadequate in an environment where legislative activity is distributed across dozens of jurisdictions running different models.
For AI developers: The Virginia IVO study is worth tracking closely. If the JCOTS recommendations support an IVO framework, and if other states follow, independent verification against state-set safety goals could become a new category of compliance obligation, one that doesn’t map to existing EU conformity assessment structures or traditional U.S. self-certification approaches.
For political campaign operatives and AI content vendors: State-level AI political advertising disclosure requirements are multiplying. Building flexibility into content workflows now, so that disclosure labels can be added or adjusted by state, is more practical than engineering a response to each new requirement individually.
For multinationals with Japan operations: Japan’s soft-law framework is not a compliance-free environment. It’s a different compliance register. Track guidance under the Basic AI Plan and build a monitoring posture that can catch any transition from voluntary principles to prescriptive rule.
This briefing is for informational purposes. Consult qualified legal counsel for compliance guidance specific to your organization and jurisdiction.