Updated March 2026. This document attempts to present the genuine state of the science: what is real, what is plausible, what is speculative, and what is likely wrong.
1. The Smalley-Drexler Debate: “Fat Fingers” and “Sticky Fingers”
The Argument
The most famous scientific confrontation in nanotechnology history played out in a 2003 Chemical & Engineering News cover story between K. Eric Drexler (MIT PhD, author of Nanosystems) and Richard Smalley (Nobel laureate, co-discoverer of buckminsterfullerene).
Smalley raised two objections to molecular assemblers:
- Fat Fingers Problem: The fingers of a manipulator arm are themselves made of atoms. There is not enough room in a nanometer-scale reaction region to accommodate all the manipulator fingers needed to have complete control over the chemistry.
- Sticky Fingers Problem: At the nanoscale, atoms on the manipulator will adhere to the atom being moved. You cannot simply pick up and place atoms like LEGO bricks, van der Waals forces, covalent bonding tendencies, and electrostatics make things stick.
Drexler’s Rebuttal
Drexler argued Smalley was attacking a straw man. No serious proposal for mechanosynthesis involves “fingers” reaching into a reaction site to manipulate five to fifteen atoms simultaneously. Instead, the proposals use:
- Scanning tunneling microscope (STM) tips as reactive structures for positional control
- Tooltip chemistry where a single reactive site on a tool interacts with a single reactive site on a workpiece
- Sequential, not simultaneous, operations
Drexler pointed to experimental STM work (like the famous IBM “atoms” logo) as proof-of-concept that positional control of individual atoms is possible.
Where Science Stands Now
The honest answer: Smalley’s specific objections were poorly framed, but his underlying intuition had merit.
- Smalley’s “fat fingers” argument was technically wrong as stated, nobody proposed the kind of multi-finger gripper he described. Drexler correctly identified this as a straw man.
- However, Smalley’s deeper point, that controlling chemistry at the atomic scale is far harder than Drexler’s mechanical engineering analogies suggest, has been validated by subsequent work.
- Nature proves molecular assemblers exist: ribosomes, ATP synthase, polyketide synthases all perform atomically precise assembly. But they operate in solution, use thermal motion constructively, and are products of billions of years of evolution.
- In 2019, chemists at Kiel University built the first artificial molecular assembler that combines selective binding of reactants, accurate positioning, and active release of product, using light as an energy source. This is a genuine milestone, though far from a universal assembler.
- Philip Moriarty (Nottingham) received a 1.53M GBP (~$3M) grant to experimentally test diamond mechanosynthesis predictions. His assessment: limited mechanosynthesis steps are plausible with appropriate materials (e.g., H:Si(100)), but universal assemblers handling all elements are unlikely because “the viable parameter space is too narrow.”
Verdict: Neither Smalley nor Drexler was entirely right. Positional chemistry is possible. Universal molecular assembly in the Drexler sense remains unproven and faces legitimate obstacles that go beyond the straw-man “fat fingers” argument.
2. Thermodynamic Arguments: What Physics Actually Constrains
The Fundamental Constraints
Matter compilation faces several non-negotiable physics constraints:
Free Energy Requirements: For any self-assembly or directed assembly to occur spontaneously, the Gibbs free energy change must be negative. For directed assembly (building structures that would not form spontaneously), you must input energy to overcome the free energy barrier. This is not controversial, it is basic thermodynamics.
Entropy Costs: Assembling atoms into ordered structures decreases entropy locally. The second law of thermodynamics demands this entropy be exported elsewhere, typically as heat. For a matter compiler building macroscale ordered structures from disordered feedstock, the entropy cost is substantial. You must:
- Break chemical bonds in feedstock (energy input)
- Sort atoms by type (entropy decrease, requiring energy)
- Form new bonds in desired configuration (may release energy, but positioning costs energy)
- Export waste heat
Energy per Operation: At room temperature, thermal energy is ~kT = 4.1 x 10^-21 J per degree of freedom. Any mechanosynthetic operation must use energy significantly above kT to be reliable against thermal noise. Drexler’s analysis in Nanosystems suggests operations on the order of 10-100 kT for reliable positioning, which is energetically cheap per atom but adds up at scale.
The Bull Case: The energy requirements per atom are actually quite modest. Diamond formation from carbon is exothermic. Many of the proposed mechanosynthesis reactions release energy. The total energy budget for a desktop nanofactory, in Drexler’s analysis, is comparable to a desktop computer, on the order of kilowatts. Biology already does atomically precise manufacturing (protein synthesis) at room temperature using ATP, with an energy cost of ~100 kT per amino acid placement. This is not thermodynamically prohibitive.
The Bear Case: Drexler’s energy estimates assume everything works perfectly, no defects, no side reactions, no retooling. Real chemistry is messier. The entropy cost of sorting atoms from arbitrary feedstock (as opposed to pre-prepared chemical inputs) is enormous. Real matter compilers would likely need highly purified, specific feedstock molecules, not “anything you throw in.” This significantly limits the “Star Trek replicator” vision.
Verdict: Thermodynamics does not forbid matter compilation. The energy requirements per atom are small. The real issue is not energy but control, maintaining atomically precise control while doing useful work at meaningful rates, in the presence of thermal noise.
3. Grey Goo: Why It Is Mostly Dismissed
The Original Concern
Eric Drexler himself introduced the grey goo scenario in Engines of Creation (1986): self-replicating nanobots consuming all matter on Earth to make copies of themselves.
Why It Is Now Dismissed
Even Drexler has walked it back. In a 2004 statement, he conceded there is “no need to build anything that even resembles a potential runaway replicator” and that self-replicating machines are “needlessly complex and inefficient.” The reasons:
Engineering impossibility at that scale: A self-replicating nanobot would need an internal computer to store its own blueprint, error-correction systems, energy harvesting from diverse sources, and the ability to process arbitrary feedstock. This is far more complex than any proposed assembler.
No universal feedstock: No single nanomachine could efficiently break down granite, water, cellulose, metals, and organic tissue. The chemistry for each is wildly different. A “universal dissembler” is harder than a universal assembler.
Energy constraints: A free-floating nanobot harvesting energy from its environment would be thermodynamically constrained to extremely slow replication rates. Biology’s fastest self-replicators (bacteria) take ~20 minutes per generation and require specific nutrient media. A nanobot in a desert or ocean would starve.
Modern nanofactory designs don’t require it: Drexler’s later work (Radical Abundance, 2013) and the broader APM community propose fixed nanofactories with convergent assembly, factory floors, not free-ranging replicators.
What remains concerning: Not grey goo, but directed atomically precise manufacturing of weapons, pathogens, or surveillance devices by state or non-state actors. 80,000 Hours identifies this as the real risk vector.
Verdict: Grey goo was always more science fiction than science. The actual risks from APM, if it is ever achieved, are more mundane and more serious: weaponization and destabilization.
4. What Feynman Actually Said
The Famous Talk
Richard Feynman delivered “There’s Plenty of Room at the Bottom” at the American Physical Society meeting at Caltech on December 29, 1959. It is retroactively treated as the founding document of nanotechnology. Here is what he actually said and what he did not:
What he said:
- “The principles of physics, as far as I can see, do not speak against the possibility of maneuvering things atom by atom. It is not an attempt to violate any laws; it is something, in principle, that can be done; but in practice, it has not been done because we are too big.”
- All the information in all the books in the world could be stored in a cube roughly 1/200th of an inch wide.
- You could “swallow the surgeon”, tiny machines operating inside blood vessels.
- He proposed building smaller tools to build yet smaller tools, a cascade of miniaturization.
- He offered two $1,000 prizes: one for writing text at 1/25,000 scale, another for a working electric motor fitting in a 1/64-inch cube.
What he did NOT say:
- He never used the word “nanotechnology” (Norio Taniguchi coined it in 1974; Drexler popularized it in the 1980s).
- He did not propose molecular assemblers or self-replicating machines.
- He was speculating about possibilities, not laying out an engineering program.
The Dirty Secret About Influence: The talk was cited only 7 times in the first 20 years after publication. Most pioneering nanoscientists have said “Plenty of Room” did not influence their early work, and many had not even read it until later. Historians of science generally say the field started in the 1980s-1990s and retroactively appropriated Feynman’s lecture as a founding myth, a “stamp of authenticity” from a revered Nobel laureate.
Verdict: Feynman’s talk was visionary and correct on the physics, nothing forbids atomic-scale manipulation. But its influence on actual nanotechnology development has been overstated. It is more founding mythology than founding document.
5. Current Limitations: Why We Cannot Do APM at Scale
The Actual Blocking Challenges
Throughput: This is the single biggest barrier. Current STM-based atomic manipulation is agonizingly slow. A single STM tip can place atoms one at a time, with each operation taking milliseconds to seconds. To build a 1 kg object (~10^25 atoms) at 1 atom per millisecond would take ~10^22 seconds, roughly 300 trillion years. Even at megahertz frequencies (Drexler’s optimistic estimate), you need massive parallelism.
Parallelism: Drexler’s proposed solution is millions or billions of assemblers working in parallel. But building the first generation of parallel assemblers is itself a monumental bootstrapping problem. You need atomically precise manufacturing to build the tools of atomically precise manufacturing. Current multi-tip STM arrays are limited to tens or hundreds of tips, not billions.
Environmental Control: Most demonstrated atomic manipulation works only in ultra-high vacuum at cryogenic temperatures (4-77 K). Room temperature operation introduces thermal noise that makes positional control vastly harder. Biological molecular machines solve this differently, they use thermal motion constructively rather than fighting it, but this is a fundamentally different design philosophy than Drexler’s rigid-machine approach.
Error Rates and Correction: At the atomic scale, even tiny error rates compound catastrophically. If you misplace one atom in a billion, a 1 kg product has ~10^16 defects. Real systems need error detection, correction, and quality control at every step, adding enormous complexity.
Software and Design: Designing atomically precise structures is itself a computational grand challenge. You need to simulate quantum-mechanical interactions for every proposed structure, predict stability, and design reaction pathways. This is computationally expensive (see Section 7).
Feedstock Preparation: Proposals assume pre-prepared, atomically pure feedstock molecules. Producing these feedstock molecules at scale is itself a hard chemistry problem.
Verdict: The blocking challenge is not fundamental physics but engineering at scale. We can manipulate individual atoms. We cannot do it fast enough, in parallel enough, at high enough fidelity, in permissive enough environments to build macroscale objects. The gap between “can manipulate one atom” and “can build a kilogram of product” is roughly 25 orders of magnitude in throughput.
6. Convergent Technologies: How AI, Quantum Computing, and Robotics Might Accelerate APM
AI-Driven Atomic Fabrication
A major 2024 breakthrough from NUS (National University of Singapore): AI-driven atomic robotic probes that integrate scanning probe microscopy with deep neural networks for fabricating carbon-based quantum materials at the atomic scale. This is a genuine convergence of AI and atomically precise placement.
Machine Learning Force Fields
ML force fields (MACE, ANI, and related neural network potentials) are transforming molecular simulation:
- MACE-OFF (2025): Transferable ML force fields for organic molecules achieving near-quantum-mechanical accuracy at classical MD speeds.
- GPU acceleration: NVIDIA’s cuEquivariance library dramatically accelerates geometric neural networks used in molecular dynamics.
- These allow simulating molecular systems 100-1000x faster than quantum mechanical methods, with comparable accuracy. This matters for designing atomically precise structures.
Quantum Computing
Quantum computing could eventually simulate molecular systems exactly (since molecules are quantum systems), but:
- Current quantum computers are too noisy and too small for useful molecular simulation beyond trivial systems.
- Hybrid classical-quantum algorithms show promise but are years from practical utility.
- The honest timeline: quantum computing will likely contribute to APM design in the 2030s-2040s, not sooner.
Robotics and Automation
Autonomous lab systems (like Berkeley Lab’s A-Lab) can synthesize and test materials with minimal human input. When combined with AI predictions, this creates an “AI-to-lab” pipeline that dramatically accelerates the experimental cycle.
Verdict: AI is the most impactful convergent technology right now. It accelerates design, simulation, and experimental automation. Quantum computing is promising but premature. The combination could plausibly compress the path to APM from “possibly never” to “possibly decades.”
7. Computational Requirements
The Scale of the Problem
Simulating atomic-scale systems accurately requires quantum mechanical calculations:
- Density Functional Theory (DFT): Scales as O(n^3) with the number of electrons. A 1000-atom system requires hours to days on a modern cluster. Freitas’s diamond mechanosynthesis work analyzed 1620 tooltip/workpiece structures using DFT, this alone was a massive computational effort.
- Molecular Dynamics (MD): Classical MD scales O(n^2) naively, O(n log n) with optimizations. You can simulate millions of atoms for nanoseconds, but not the billions of atoms for microseconds needed for realistic nanomachine design.
- Quantum Chemistry: Full configuration interaction (exact) scales exponentially. Practical only for systems of ~20-30 atoms.
How AI Changes This
Machine learning is creating a paradigm shift:
- Neural Network Potentials: Models like ANI-2x, MACE, and FALCON achieve near-DFT accuracy at classical MD cost, a speedup of 3-5 orders of magnitude over quantum mechanical calculation.
- Coarse-Graining via ML: AI can learn effective coarse-grained representations, enabling simulation of mesoscale structures (millions of atoms) while retaining atomic-level accuracy where it matters.
- Inverse Design: Instead of simulating every candidate, ML models can predict which structures will be stable and functional, dramatically reducing the search space.
- AlphaFold: Solved protein structure prediction (cited ~43,000 times, used by 3M+ researchers). A genuine breakthrough in structural biology, though protein structure prediction is a different problem from materials fabrication.
Honest Caveats on AI for Materials
- Google DeepMind’s GNoME claim of “millions of new materials” has been criticized: many are trivial variants of known materials, and few are strikingly novel.
- A late 2024 MIT paper claiming AI-driven materials invention was found unreliable by MIT’s own economics department.
- AI excels at interpolation within known chemical space but struggles with truly novel chemistries outside its training data.
- The “AI-to-lab” gap remains large: predicting a material is stable on a computer is very different from synthesizing it in a lab.
Verdict: AI is the single most important accelerant for APM design. It compresses what would take decades of computation into months. But AI is a tool for design, not for fabrication. The physical manufacturing bottlenecks remain.
8. Scale-Up Challenges: From Atoms to Objects
The Throughput Arithmetic
This is where the dream meets brutal arithmetic:
- A 1 kg diamond cube contains approximately 10^25 carbon atoms.
- A single STM tip operating at ~1 atom per second would take 3 x 10^17 years.
- At 1 MHz (Drexler’s optimistic frequency), one assembler takes 3 x 10^11 years.
- With 10^12 parallel assemblers at 1 MHz each: ~300 seconds. This is Drexler’s convergent assembly target.
The bootstrapping problem: You need 10^12 assemblers. Each assembler is itself an atomically precise machine containing perhaps 10^6 atoms. So you need 10^18 atoms placed precisely to build your assembler array. Which requires… assemblers.
Drexler’s answer is exponential assembly: one assembler builds two, two build four, and so on. After ~40 doublings, you have 10^12. If each doubling takes hours, the whole process takes days. This is elegant in theory but:
- Error accumulation over 40 generations of self-replication
- No experimental demonstration of even ONE generation of mechanical self-replication at the nanoscale
- Quality control across 10^12 units is an unsolved problem
Convergent Assembly Architecture
Drexler and others propose a hierarchical system:
- Nanoscale fabricators produce molecular components
- Components are transported to block assemblers
- Blocks are assembled into larger blocks
- Process repeats up to macroscale
Predicted throughput: ~1 kg/hour of product per kg of nanofactory, with the final stage taking ~100 seconds. This is impressive if achievable, but rests on assumptions about:
- Reliable operation of 10^12+ nanoscale devices simultaneously
- Transport mechanisms between assembly stages
- Defect tolerance and error correction at every level
- Feedstock delivery to 10^12 reaction sites
What Actually Works Now
Current real-world nanomanufacturing uses:
- Semiconductor lithography: Atomically precise in 2D (EUV, now at sub-nm feature sizes via Zyvex’s STM lithography), but not in 3D general shapes.
- DNA origami: Can create 100 nm-scale shapes with ~6 nm resolution, but limited materials (DNA only) and limited mechanical properties.
- Chemical vapor deposition: Grows diamond and graphene with atomic-layer control in one direction, but not arbitrary 3D shapes.
Verdict: The scale-up problem is the hardest unsolved challenge in APM. The throughput gap between current capabilities and macroscale manufacturing is approximately 20-25 orders of magnitude. No credible near-term path closes this gap. Exponential assembly is theoretically elegant but experimentally undemonstrated.
9. Alternative Paths to Matter Compilation
Directed Self-Assembly (DSA)
Instead of placing every atom mechanically, you design molecules that want to assemble into your target structure.
- Status: Actively used in semiconductor manufacturing for pattern transfer. Block copolymers can self-assemble into regular nanoscale patterns.
- Strengths: Massively parallel (every molecule is an assembler), works at room temperature, scalable.
- Limitations: Limited to periodic or symmetric structures. Cannot build arbitrary shapes. Error correction is passive (thermodynamic), not active.
DNA Nanotechnology and Origami
DNA is the most programmable self-assembling material known:
- DNA Origami: A long scaffold strand folded by hundreds of short “staple” strands into arbitrary 2D and 3D shapes at ~100 nm scale.
- DNA Bricks: Modular assembly without a scaffold strand.
- 2024-2025 advances: Self-replicating DNA wireframe nanoassemblies (JACS, 2024), isothermal self-assembly at room temperature, DNA nanorobots that can alter synthetic cell membranes.
- As scaffolding for APM: DNA structures can template the placement of other materials, metals, quantum dots, enzymes, with nanometer precision.
- Limitations: DNA is mechanically weak, thermally unstable above ~60C, and limited to aqueous environments. Not suitable for structural materials.
Bio-Inspired Approaches
Reverse-engineering biological molecular machines:
- Ribosomes as existence proof: The ribosome is a molecular assembler that reads a digital program (mRNA) and produces atomically precise products (proteins) at ~20 amino acids per second. It works at room temperature in solution.
- Synthetic biology: Engineering cells to produce custom molecules, materials, and structures. Already commercially viable (insulin, spider silk, biofuels).
- 2025 advances: Autonomous ribosome biogenesis in vitro, synthetic cells with engineered functions, DNA nanorobots controlling synthetic cell membranes.
- Limitations: Biological systems are limited to biological materials (proteins, nucleic acids, lipids, sugars). Cannot build diamond, metals, or semiconductors.
Hybrid Approaches
The most promising near-term path may combine multiple approaches:
- AI designs the target structure and reaction pathways
- DNA origami provides the nanoscale scaffold and template
- Directed self-assembly positions functional components
- Enzymatic or chemical processes form the final bonds
- Bio-inspired error correction ensures fidelity
This is not Drexler’s vision of diamond mechanosynthesis, but it may be more achievable in the near term.
Verdict: Alternative paths are more mature than Drexlerian mechanosynthesis. DNA nanotechnology and synthetic biology are producing real, useful products today. They cannot build diamond or arbitrary materials, but they are not stuck in the theoretical stage. The likely path to matter compilation, if it happens, may look more like engineered biology than mechanical engineering.
10. AI’s Role in Accelerating Materials Science
What Is Already Working
AI is genuinely transforming materials science, not just hype:
- AlphaFold (DeepMind): Solved protein structure prediction. 43,000+ citations, 3M+ users across 190+ countries. Won the 2024 Nobel Prize in Chemistry (Hassabis and Jumper). Transformative impact on structural biology.
- GNoME (DeepMind): Predicted 2.2 million new crystal structures. However, a Max Planck Institute study found over 80% showed structural disorder, and a PRX Energy analysis found most “novel” predictions are trivial compositional variants. 736 were independently synthesized, a 0.03% validation rate. Useful as a screening tool, not the revolution the press release claimed. See AI in Materials Science for the full analysis.
- Autonomous Labs: Berkeley Lab’s A-Lab is a genuinely impressive robotic synthesis platform. However, its novelty claims for synthesized materials are disputed, the platform works, the discovery claims are in question.
- ML Force Fields: MACE, ANI, and related models enable molecular simulation at 100-1000x the speed of quantum methods with comparable accuracy.
- Inverse Design: ML models can now predict which compositions and structures will have desired properties, inverting the traditional trial-and-error approach.
Honest Assessment of Limitations
- No big wins yet in novel materials: Despite millions of predictions, there has not been a convincing breakthrough material discovered primarily through AI. Most AI-predicted materials are incremental variants.
- Validation gap: Computational prediction and experimental realization remain disconnected. Most AI predictions are never tested in a lab.
- Training data bias: AI models are only as good as their training data. They interpolate well within known chemical space but cannot reliably extrapolate to genuinely novel chemistries.
- Reproducibility concerns: The 2024 MIT AI materials paper found unreliable by MIT’s own economics department highlights systemic issues with overclaiming.
How AI Could Accelerate the Path to APM
The most impactful AI contributions would be:
- Reaction pathway design: AI identifying viable mechanosynthesis reaction sequences (replacing the laborious manual DFT calculations Freitas did for 1620 structures)
- Error prediction: ML models predicting defect formation rates and failure modes in proposed nanomachine designs
- Control systems: Reinforcement learning for real-time control of STM tips and atomic manipulation
- Automated experimentation: AI-driven robotic labs testing thousands of APM-relevant reactions per day
- Multi-scale simulation: ML bridging the gap between quantum-accurate atomic simulation and mesoscale device behavior
Verdict: AI is the most powerful accelerant available for materials science and, by extension, APM. But it accelerates design and discovery, not manufacturing. The physical fabrication bottleneck, building things atom by atom fast enough to matter, is not an AI problem. AI gets you to better blueprints faster. You still need the factory.
Overall Assessment: The Honest Picture
What the Bulls Get Right
- Physics does not forbid atomically precise manufacturing. Thermodynamics is not the barrier.
- Nature proves molecular assembly works: ribosomes, ATP synthase, polyketide synthases.
- AI is dramatically accelerating materials design and simulation.
- Zyvex is making real progress on STM-based atomically precise lithography for silicon electronics.
- DNA nanotechnology is producing increasingly sophisticated and functional nanostructures.
- The theoretical framework in Nanosystems has withstood decades of scrutiny on the physics. The calculations are not wrong.
What the Bears Get Right
- The throughput/scale-up problem is 20-25 orders of magnitude away from useful. This is not a gap that clever engineering closes quickly.
- No one has demonstrated even a single cycle of mechanical self-replication at the nanoscale, and convergent assembly requires this.
- Drexler’s rigid-machine paradigm fights thermal noise rather than harnessing it, unlike biology. Richard Jones’s critique that soft, flexible approaches may be more viable has merit.
- The “universal assembler” handling arbitrary elements is likely impossible. Moriarty’s assessment of “too narrow a parameter space” is probably correct.
- 25+ years after Nanosystems, the most advanced experimental demonstrations are still individual atom placements by single STM tips in vacuum. The gap between theory and practice has barely narrowed.
- AI hype in materials science has outpaced delivery. GNoME’s “millions of new materials” is mostly incremental variations, not breakthroughs.
Timeline Estimates
The expert range is enormous, reflecting genuine uncertainty:
- Drexler / Phoenix (bull case): With substantial investment, transformative APM within a decade of serious effort. (Note: Drexler has been saying versions of this since the 1990s.)
- 80,000 Hours / Ben Snodin: 4-5% probability of advanced nanotechnology by 2040 (self-described as “unstable” estimate).
- Jones / Moriarty (bear case): Advanced APM posing substantial risks or significantly changing society is “further in the future”, implying post-2050 at earliest, if ever.
- Wildcard: If transformative AI arrives first and can automate scientific research, it could dramatically compress the timeline, but at that point, APM may be the least of humanity’s concerns.
The Most Likely Path
The path to something resembling “matter compilation” will probably not look like Drexler’s diamond mechanosynthesis nanofactory. More likely:
- Near-term (2025-2035): AI-designed materials, DNA nanotechnology for nanoscale templates, advanced lithography for 2D atomically precise semiconductor devices, synthetic biology for custom biomolecules.
- Medium-term (2035-2050): Hybrid bio-nano systems, massively parallel directed self-assembly, AI-controlled robotic chemistry, limited atomically precise 3D printing of specific materials.
- Long-term (2050+): If the throughput and parallelism problems are solved, general-purpose atomically precise manufacturing of arbitrary structures. This is the “matter compiler” vision, and it remains genuinely uncertain whether it will ever be achieved.
The honest bottom line: The physics allows it. The engineering is staggeringly hard. The timeline is measured in decades at minimum, and “never” remains a live possibility for the full Drexlerian vision. The path will likely go through biology and self-assembly rather than mechanical engineering. The most valuable near-term work is in AI-accelerated materials design, DNA nanotechnology, and synthetic biology, not in building diamond mechanosynthesis tooltips.
Sources
- Drexler-Smalley Debate - Wikipedia
- Drexler-Smalley Debate - Chemical & Engineering News
- Molecular Assemblers: Molecular Machines Performing Chemical Synthesis - Chemical Science (RSC)
- Grey Goo - Wikipedia
- Nanotechnology Pioneer Slays Grey Goo Myths - Phys.org
- Grey Goo - Britannica
- Feynman’s “Plenty of Room at the Bottom” - Wikipedia
- “Plenty of Room” Revisited - Nature Nanotechnology
- Feynman’s Fancy - Chemistry World
- A Comprehensive Analysis of the Future of APM - arXiv 2024
- Risks from Atomically Precise Manufacturing - 80,000 Hours
- AI-Driven Atomic Robotic Probe - ScienceDaily
- Synergizing Quantum Dynamics, AI, and Quantum Computing for Materials Science - PMC
- MACE-OFF: Transferable ML Force Fields - JACS
- Enabling Scalable AI-Driven Molecular Dynamics - NVIDIA
- Scalable Nanomanufacturing Review - PMC
- A Matter of Scale - Nature Nanotechnology
- Self-Replicating DNA Nanoassemblies - JACS 2024
- Programming DNA Machines to Move - Nature Reviews Chemistry 2025
- AI Materials Discovery Needs Real World - MIT Technology Review 2025
- Google DeepMind GNoME - Nature
- AI Driving Materials Discovery? - Chemistry of Materials
- Convergent Assembly - Zyvex
- Diamond Mechanosynthesis - Freitas / Nanofactory Collaboration
- A Minimal Toolset for Positional Diamond Mechanosynthesis - Freitas
- Is Mechanosynthesis Feasible? - Soft Machines (Richard Jones)
- The Mechanosynthesis Debate - Soft Machines
- Zyvex APM Publications
- Atomically Precise Manufacturing of Silicon Electronics - ACS Nano
- The Coming Wave of Confluent Biosynthetic Technologies - Nature Communications 2025
- Convergence of AI and Synthetic Biology - npj Biomedical Innovations 2025
- The Next 25 Years of Nanoscience - Nano Letters 2025
- Nanotechnology: Grey Goo is a Small Issue - CRN
- Design of a Primitive Nanofactory - Chris Phoenix
- Molecular Assembler - Wikipedia