Grigori Guitchounts

Grigori Guitchounts

@guitchounts.bsky.social

AI + biology, venture creation @FlagshipPioneer | Neuroscientist @Harvard | Writer, pianist, runner, painter, dilettante

399 Followers 181 Following 314 Posts Joined Oct 2023
1 month ago

https://arxiv.org/abs/2602.08079

0 0 0 0
1 month ago

The question I can’t stop poking: which of these can we write down in engineering terms without slipping into vibes? “Signaling” and “goal-level control” are very real in tissues. What’s the smallest synthetic analogue that could actually scale?

0 0 1 0
1 month ago

I read it as a gentle jab at the default AI recipe: train a huge model, bolt on tools, cross fingers that generalization appears. Biology leans harder on agents that can build and maintain themselves while staying tightly coupled to the world that’s pushing back.

0 0 1 0
1 month ago

Design principles they emphasize (the biology route): multiscale autonomy; growth via self-assembly of active parts; constant reconstruction (repair/rewiring); letting embodiment + physics do some of the work; dense signaling so self-organization can coexist with top-down goals.

0 0 1 0
1 month ago

Their main image is a “cognitive light cone”: how far an agent can sense, predict, and steer across space, time, and scale. Evolution, in this telling, mostly expands that cone—more distant goals, longer horizons, tighter loops of control.

0 0 1 0
1 month ago

They borrow William James’ framing: intelligence as “fixed goals, variable means.” I like it as a handhold. It points straight at generalization: reaching the same outcome in a situation you didn’t rehearse, instead of replaying a polished policy.

0 0 1 0
1 month ago

A lot of biology’s “intelligence” shows up before anything like a brain: cells negotiating, wounds closing, embryos turning a noisy soup of parts into a reliable body. This paper argues that if we want robust machine intelligence, we should copy more of that— not just neurons.

0 0 1 0
1 month ago

Two choices that caught my attention: (1) trained retrievers + rerankers—so it’s not just “vector search and hope.” (2) an iterative self-feedback loop that re-retrieves + rewrites to improve factuality, coverage, and citation accuracy. Same loop makes synthetic data to train a compact 8B model.

0 0 0 0
1 month ago

Their fix is almost boringly reasonable: first retrieve relevant passages from a big open corpus, then write an answer that stays tethered to those passages. They build OpenScholar on a 45M-paper datastore with 236M passage embeddings.

0 0 1 0
1 month ago

If you ask a general LLM to “summarize the recent literature” and tack on citations, it’ll often confidently hand you references that… don’t exist. This paper actually measures it: for recent, cross-field queries, GPT-4o’s citations were fabricated ~78–90% of the time.

0 0 1 0
1 month ago

No benchmark is going to bottle end-to-end discovery. But ones like this can keep us honest about whether “agents” are improving at the workflows researchers actually run, not just test-taking. Link: https://drive.google.com/file/d/1BV5UtmBRdpbQoz9jC1AuUF8WUTRQMqK_/view

0 0 0 0
1 month ago

They report a real difficulty jump vs the original LAB-Bench: model accuracies drop in a model-specific way, but roughly ~26–46% across subtasks. Even as models get better, more lifelike framing keeps finding the soft spots.

0 0 1 0
1 month ago

What caught my attention is how unglamorous this is—in a good way. The hard part is often: find the right document, pull the one detail that matters, and resist treating a shaky source like gospel. That’s closer to “useful” than trivia.

0 0 1 0
1 month ago

The vibe is: fewer multiple-choice guardrails, more open-response + retrieval. Instead of “here’s a figure, answer X,” it sticks figures/tables back into paper-like context. Also tasks on patents, clinical trials, and judging source quality.

0 0 1 0
1 month ago

LABBench2 is trying to grade that messier layer. ~1,900 biology-flavored tasks spanning literature retrieval/understanding, data access, protocol troubleshooting, molecular biology help, and experiment planning.

0 0 1 0
1 month ago

A lot of “AI for science” eval still sounds like a school exam: answer the question, show your work. Real research is messier—dig up the right papers, read figures in context, debug a protocol that’s failing, pull sequences, plan experiments, defend the choices.

0 0 1 0
1 month ago

Fundraising + product announcement (seed $13.5M, Biomni Lab preview): https://phylo.bio/blog/company-fundraising-announcement

0 0 0 0
1 month ago

What I keep coming back to is the interaction design: steerable agents that can execute real workflows (tools, pipelines, maybe lab robotics) while keeping provenance, citations, and reproducibility intact. That’s a brutal constraint set. Also the right one.

0 0 1 0
1 month ago

Two things I’d want nailed down (maybe in follow-ups): what “rigorous, safe, reliable” looks like in evals, and what failure modes look like in the wild. Quietly wrong analyses are the nightmare—worse than slow, because you trust them.

0 0 1 0
1 month ago

They say Biomni (the open-source agent) has been adopted by tens of thousands of scientists across 7,000+ labs/orgs. That catches my eye. Scientists are famously allergic to shiny tools—unless the tool reliably buys them time back on tasks that actually hurt.

0 0 1 0
1 month ago

Phylo is pitching Biomni Lab as that workspace: one place to orchestrate agents across data modalities and drug discovery stages, with 300+ databases/tools wired in. The vibe is less “chatbot” and more “the OS you live in during real research days.”

0 0 1 0
1 month ago

A lot of biology is still “glue work.” Ten tabs open, CSVs shuttled around like buckets in a leak, half-manual analyses, scripts you swear you’ll clean up later, tables reformatted… again. The claim here: an “integrated biology environment” + agents becomes the default interface.

1 0 2 0
1 month ago

Tactile Robotics: Past and Future (Nathan F. Lepora) https://arxiv.org/pdf/2512.01106

0 0 0 0
1 month ago

If tactile robotics really matures after 2025, the payoff isn’t just nicer grippers. It becomes a testbed for sensorimotor control (what does a policy actually need to know, moment to moment?) and for telepresence where touch feedback isn’t a gimmick.

0 0 1 0
1 month ago

A note of skepticism runs through it (and through me): tactile sensing has been “a few years away” for decades. The review frames that as recurring bottlenecks—great demos that don’t survive the messy, repetitive reality of deployment. The history makes the loop hard to unsee.

0 0 1 0
1 month ago

The post-2010 “expansion and diversification” gets broken into categories that feel very familiar if you’ve been around labs: e-skins, tactile robot hands, vision-based tactile sensors (camera + squishy elastomer), soft/biomimetic touch, and the “tactile Internet” (touch over networks).

0 0 1 0
1 month ago

I like the constraint: mostly reviews only (except pre-1980). It’s a way of sampling expert consensus without pretending you read an infinite pile of one-off papers. Also: a field’s memory often lives in surveys, not in the headlines.

1 0 1 0
1 month ago

This review tries a simple forecasting move: listen to what ~150 review papers have been saying for ~50 years, then sketch “generations” of tactile robotics: origins (1965–79), foundations (1980–94), a tactile winter (1995–2009), and the post-2010 expansion.

0 0 1 0
1 month ago

Touch feels like the missing sense in a lot of “dexterous” robotics. You can watch contact, sure—but you can’t reliably feel slip, pressure spreading across a fingertip, shear, or tiny surface texture. That’s how human hands modulate force without crushing or dropping.

0 0 1 0
1 month ago

From sequence to function: bridging single-molecule kinetics and molecular diversity (Science) https://www.science.org/doi/10.1126/science.adv4503

0 0 0 0