Understanding RNA Therapeutics: Revolutionizing Medicine

RNA therapeutics are medicines that use ribonucleic acid (RNA) molecules as the active drug to treat or prevent disease. Instead of traditional small-molecule drugs or protein-based biologics, they work by directly influencing how genes are expressed inside cells.

There are several main types of RNA therapeutics:

TypeHow It WorksKey Examples (Approved or Famous)Main Uses
mRNA vaccines / therapeuticsDeliver messenger RNA (mRNA) that instructs cells to produce a specific protein (e.g., viral spike protein or a missing enzyme)Pfizer-BioNTech & Moderna COVID-19 vaccines, BioNTech’s cancer vaccines (in trials)Vaccines, cancer immunotherapy, protein replacement
ASOs (Antisense Oligonucleotides)Short, synthetic single-stranded DNA/RNA-like molecules that bind to target mRNA and block or degrade itNusinersen (Spinraza) for spinal muscular atrophy, Inotersen & Patisiran for hereditary ATTR amyloidosisRare genetic diseases, neurological disorders
siRNA (small interfering RNA)Double-stranded RNA that triggers the cell’s natural RNA interference (RNAi) machinery to silence specific genesPatisiran (Onpattro) – first ever FDA-approved siRNA, Givosiran for acute hepatic porphyriaGenetic diseases, liver diseases, some cancers
saRNA (self-amplifying RNA)mRNA that encodes not only the target protein but also a viral replicase, so it copies itself inside the cell → longer, stronger protein production with tiny dosesIn development (e.g., Gritstone, Arcturus COVID/flu programs)Vaccines (especially low-dose, fridge-stable ones)
RNA aptamersFolded RNA molecules that bind proteins like antibodiesPegaptanib (Macugen) – first RNA aptamer drug (for macular degeneration)Eye diseases, anticoagulation, cancer
Circular RNA (circRNA)RNA in a closed loop → very stable, long-lasting protein expressionEarly clinical trials (e.g., Orna Therapeutics)Protein replacement, vaccines

Why RNA Therapeutics Are a Big Deal

  1. Speed of development
    – COVID mRNA vaccines went from sequence to emergency use in <1 year (vs. 10–15 years for traditional vaccines).
  2. Precision
    – You can target almost any gene or protein. If we know the genetic cause of a disease, we can design an RNA drug against it.
  3. “Undruggable” targets become druggable
    – Many diseases are caused by proteins that small molecules can’t bind well. RNA drugs act before the protein is even made.
  4. Personalization potential
    – Easy to customize mRNA sequence for a patient’s specific mutation (already happening in cancer vaccines).

Major Challenges (Why They’re Hard)

ChallengeExplanationCurrent Solutions / Progress
DeliveryNaked RNA is destroyed quickly by enzymes and can’t easily enter cellsLipid nanoparticles (LNPs), GalNAc conjugates, new polymers
Immune activationRNA can trigger strong inflammatory responsesChemical modifications (pseudouridine, etc.)
Manufacturing scale-upVery sensitive biologic; hard to make consistently at huge scaleMassive investment post-COVID; new platforms emerging
Duration of effectMost RNA effects are transient (days to weeks)circRNA, saRNA, repeated dosing, or gene editing combos
CostStill expensive compared to small-molecule pillsEconomies of scale improving rapidly

The Future (2025–2030 Outlook)

  • Hundreds of RNA programs in clinical trials (cancer, rare diseases, infectious diseases, Alzheimer’s, heart disease, etc.).
  • Next-generation delivery: targeting lungs, brain, heart muscle, tumors directly.
  • Combination with CRISPR: using mRNA to deliver gene-editing machinery (already in trials).
  • Off-the-shelf and personalized cancer vaccines likely to get approved in the next few years.

In short: RNA therapeutics are one of the fastest-growing areas in medicine right now. They turned science fiction (programmable medicines) into reality with the COVID vaccines, and the pipeline behind them is enormous.

Elon Musk, Diffusion Models, and the Rise of Mercury

  1. A New Paradigm May Be Forming
  2. Meet Inception Labs and Mercury
  3. How Mercury Works
  4. Inside the Diffusion Revolution
  5. Training and Scale
  6. Performance: 10× Faster, Same Quality
  7. A Historical Echo
  8. What Comes Next
  9. Further Reading

A New Paradigm May Be Forming

In a recent exchange on X, Elon Musk echoed a striking prediction: diffusion models — the same architecture that powers image generators like Stable Diffusion — could soon dominate most AI workloads. Musk cited Stanford professor Stefano Ermon, whose research argues that diffusion models’ inherent parallelism gives them a decisive advantage over the sequential, autoregressive transformers that currently power GPT-4, Claude, and Gemini.

While transformers have defined the past five years of AI, Musk’s comment hints at an impending architectural shift — one reminiscent of the deep learning revolutions that came before it.


Meet Inception Labs and Mercury

That shift is being engineered by Inception Labs, a startup founded by Stanford professors including Ermon himself. Their flagship system, Mercury, is the world’s first diffusion-based large language model (dLLM) designed for commercial-scale text generation.

The company recently raised $50 million to scale this approach, claiming Mercury achieves up to 10× faster inference than comparable transformer models by eliminating sequential bottlenecks. The vision: make diffusion not just for pixels, but for language, video, and world modeling.


How Mercury Works

Traditional LLMs — whether GPT-4 or Claude — predict the next token one at a time, in sequence. Mercury instead starts with noise and refines it toward coherent text in parallel, using a denoising process adapted from image diffusion.

This process unfolds in two stages:

  1. Forward Process: Mercury gradually corrupts real text into noise across multiple steps, learning the statistical structure of language.
  2. Reverse Process: During inference, it starts from noise and iteratively denoises, producing complete sequences — multiple tokens at once.

By replacing next-token prediction with a diffusion denoising objective, Mercury gains parallelism, error correction, and remarkable speed. Despite this radical shift, it retains transformer backbones for compatibility with existing training and inference pipelines (SFT, RLHF, DPO, etc.).


Inside the Diffusion Revolution

Mercury’s text diffusion process operates on discrete token sequences x \in X. Each diffusion step samples and refines latent variables z_t that move from pure noise toward meaningful text representations. The training objective minimizes a weighted denoising loss:

L(x) = -\mathbb{E}t [\gamma(t) \cdot \mathbb{E}{z_t \sim q} \log p_\theta(x | z_t)]

In practice, this means Mercury can correct itself mid-generation — something autoregressive transformers fundamentally struggle with. The result is a coarse-to-fine decoding loop that predicts multiple tokens simultaneously, improving both efficiency and coherence.


Training and Scale

Mercury is trained on trillions of tokens spanning web, code, and curated synthetic data. The models range from compact “Mini” and “Small” versions up to large generalist systems with context windows up to 128K tokens. Inference typically completes in 10–50 denoising steps — orders of magnitude faster than sequential generation.

Training runs on NVIDIA H100 clusters using standard LLM toolchains, with alignment handled via instruction tuning and preference optimization.


Performance: 10× Faster, Same Quality

On paper, Mercury’s numbers are eye-catching:

BenchmarkMercury Coder MiniMercury Coder SmallGPT-4o MiniClaude 3.5 Haiku
HumanEval (%)88.090.0~8590+
MBPP (%)76.677.1~75~78
Tokens/sec (H100)110973759~100
Latency (ms, Copilot Arena)25N/A~100~50

Mercury rivals or surpasses transformer baselines on code and reasoning tasks, while generating 5–20× faster on equivalent hardware. Its performance on Fill-in-the-Middle (FIM) benchmarks also suggests diffusion’s potential for robust, parallel context editing — a key advantage for agents, copilots, and IDE integrations.


A Historical Echo

Machine learning has cycled through dominant architectures roughly every decade:

  • 2000s: Convolutional Neural Networks (CNNs)
  • 2010s: Recurrent Neural Networks (RNNs)
  • 2020s: Transformers

Each leap offered not just better accuracy, but better compute scaling. Diffusion may be the next inflection point — especially as GPUs, TPUs, and NPUs evolve for parallel workloads.

Skeptics, however, note that language generation’s discrete structure may resist full diffusion dominance. Transformers enjoy massive tooling, dataset, and framework support. Replacing them wholesale won’t happen overnight. But if diffusion proves cheaper, faster, and scalable, its trajectory may mirror the very transformers it now challenges.


What Comes Next

Inception Labs has begun opening Mercury APIs at platform.inceptionlabs.ai, pricing at $0.25 per million input tokens and $1.00 per million output tokens — a clear signal they’re aiming at OpenAI-level production workloads. The Mercury Coder Playground is live for testing, and a generalist chat model is now in closed beta.

If Musk and Ermon are right, diffusion could define the next chapter of AI — one where text, video, and world models share the same generative backbone. And if Mercury’s numbers hold, that chapter may arrive sooner than anyone expects.


Further Reading

  • Stefano Ermon et al., Diffusion Language Models Are Parallel Transformers (Stanford AI Lab)
  • Elon Musk on X, Diffusion Will Likely Dominate Future AI Workloads
  • Inception Labs, Mercury Technical Overview (2025)

Build and Sell n8n AI Agents — an eight-plus-hour, no-code course


AI Agents and Automation

  1. AI agents have two parts: a brain—that is, a large language model with memory—and instructions in the system prompt. Together they let the agent make decisions and take actions through connected tools.
  2. Reactive prompting beats proactive prompting: begin with no prompt, then add lines only when errors appear. This makes debugging simpler.
  3. Give each user a unique session ID so the agent’s memory stays separate, enabling personal conversations with many users at once.
  4. Use Retrieval-Augmented Generation, or R-A-G. The agent asks a question, looks up an answer in a vector database, then crafts the reply—boosting accuracy.

AI Workflows and Best Practices

  1. AI workflows—straight, deterministic pipelines—are usually cheaper and more reliable than free-roaming agents, and they’re easier to debug.
  2. Wire-frame the whole workflow first. Mapping eighty to eighty-five percent of the flow upfront clarifies what to build.
  3. Combine agents in a multi-agent system: an orchestrator assigns tasks to specialist sub-agents. That raises accuracy and control.
  4. Apply an evaluator–optimizer loop. One component scores the output; another revises it, repeating until quality is high.

AI Integration and Tools

  1. n8n is a powerful no-code platform for AI automations; you can create and even sell more than fifteen working examples.
  2. Open Router picks the best large language model for each request on the fly, balancing cost and performance.
  3. Eleven Labs adds voice input to an email agent. Pair it with Google Sheets for contacts and the Gmail API for sending mail.
  4. Tavly offers a thousand free web searches per month—handy for research inside AI content workflows.

AI Agent Development Strategies

  1. Scale vertically first: perfect one domain—its knowledge base, data sources, and monitoring—before branching out.
  2. Test rigorously, add guard-rails, and monitor performance continuously before you hit production.
  3. Use hard prompting: spell out examples of correct and incorrect behavior right in the system prompt.
  4. Allow unlimited revision loops when refining text, so the workflow can keep improving its answer until it satisfies you.

AI Business Applications

  1. Three-quarters of small businesses already use AI; eighty-six percent of adopters earn over one million dollars in annual AI-driven revenue.
  2. AI-guided marketing lifts ROI by twenty-two percent, while optimized supply chains trim transport costs five to ten percent.
  3. AI customer-service agents cut response times sixty percent and solve eighty percent of issues unaided.
  4. The median small business spends just eighteen-hundred dollars a year on AI—under one-fifty a month.

AI Development Techniques

  1. Structure prompts with five parts: overview, tools, rules, examples, and closing notes.
  2. Debug one change at a time—alter a single line to isolate the issue.
  3. Log usage and cost in Google Sheets to track tokens and efficiency.
  4. Use polling in workflows: check task status at intervals before moving on.

AI Integration with External Services

  1. In Google Cloud, enable the Drive API, set up OAuth, and link n8n for file workflows.
  2. Do the same with the Gmail API to trigger flows and send replies.
  3. Build a Pinecone vector index (for example, with text-embedding-3-small) for fast R-A-G look-ups.
  4. Generate graphics through OpenAI’s image API to save about twenty minutes per post.

Advanced AI Techniques

  1. Use a routing framework to classify inputs and dispatch them to the right specialist agent.
  2. Add parallelization so different facets of the same input are analyzed simultaneously, then merged.
  3. Store text as vectors in a vector database for semantic search—meaning matters more than keywords.
  4. Deploy an M-C-P server as a universal translator between agents and tools, exposing tool lists and schemas.

AI Development Challenges and Considerations

  1. Remember: most online agent demos are proofs of concept—not drop-in, production-ready templates.
  2. Security matters; an M-C-P server could access sensitive resources, so lock it down.
  3. Weigh agents versus workflows; use agents only when you need complex reasoning and flexible decisions.
  4. Supply high-quality context—otherwise you risk hallucinations, tool misuse, or vague answers.

AI Tools and Platforms

  1. Alstio Cloud manages open-source apps like n8n for you—install, configure, and update.
  2. Tools such as Vellum and L-M Arena let you compare language-model performance head-to-head.
  3. Supabase or Firebase cover user auth and data storage in AI-enabled web apps.
  4. In self-hosted n8n, explore community nodes—for instance, Firecrawl or Airbnb—to expand functionality.

🚨 The AI Cyber‑Warfare Threat: Insights from Geoffrey Hinton on DOAC

In his appearance on The Diary Of A CEO with Steven Bartlett, Geoffrey Hinton—the so-called “Godfather of AI”—issued a compelling warning about AI’s dual-use potential. While AI offers immense benefits, “at least half” of its development is likely directed towards offensive cyber operations. This includes crafting more potent attacks, designing new malware, and automating exploits in real time.

  1. Cyber‑Attacks Supercharged by AI
    • From reactive to proactive: AI not only defends networks but also enables automated scouting for vulnerabilities and weaponized code generation.
    • Escalating sophistication: Cyber‑criminals and state actors are already leveraging AI to build advanced phishing campaigns. They are also using it to develop malware. This forces a continuous escalation in cyber warfare.
  2. Biological Risks: AI‐Designed Viruses

Hinton raised the specter of AI-aided bioengineering. This crossover risk—where cyber AI knowledge facilitates biological threats—represents a chilling frontier.:

“There’s people using it to make nasty viruses” .

  1. Election Manipulation Beyond Digital Borders
  • AI’s ability to model and influence human behavior isn’t limited to malware. According to Hinton, AI-driven tools can:
  • Craft hyper-personalized messaging to sway individuals,
  • Potentially manipulate public opinion and democratic processes.
  1. Urgent Call for Safety‑First Governance

Hinton emphasized that the moment to act is now:

  • Governments should mandate major AI firms to allocate a portion of compute resources toward safety testing,
  • This includes rigorous safety evaluations prior to release and independent oversight .
  • Without safeguards, profits and power will continue to outweigh safety—leaving us vulnerable.

📝 What a Responsible Defense Looks Like

If you’re thinking about policy and strategic frameworks, here’s a roadmap inspired by Hinton’s analysis:

Key Focus Area Recommended Action

  • Regulation & Oversight Governments must require safety audits for AI models before deployment.
  • Safety‑first R&D Major AI labs should allocate dedicated compute to adversarial safety research.
  • Global Cooperation Collaboration across countries to counter cross-border misuse including bio‑threats.
  • Public Awareness Inform citizens and organizations about AI-driven threat evolution—phishing, malware, targeted political influence.

Final Thoughts

Hinton’s warnings aren’t speculation; they’re grounded in current tech trajectories. AI isn’t just a tool; it’s fast becoming the weapon of choice in cyber and bio conflict.

But there is hope. With proactive safety commitments, regulations tailored to dual-use risk, and global collaboration, we can choose to channel AI’s power responsibly. The question is: will society act before technology outruns us?

Rise of AI Development Environments

The rise of Cursor, Copilot + VSCode, Replit, and Qwen2.5 among others, have caused me to rethink my ways. Focus will still be key in discerning what to build.


AI development environments change the global technology conversation. They also influence the pace of hiring and team augmentation decisions.

Qwen2.5-Coder Open Source

Alibaba Group has released the Qwen2.5-Coder open-source model. Qwen2.5-Coder-32B-Instruct is currently the best-performing open-source code model (SOTA), matching the coding capabilities of GPT-4o. Qwen2.5-Coder offers six different model sizes: 0.5B, 1.5B, 3B, 7B, 14B, and 32B.

Project: https://qwenlm.github.io/blog/qwen2.5-coder-family/

Each size provides both Base and Instruct models. The Instruct model engages in direct dialogue. The Base model serves as a foundational model for developers to fine-tune.

Github: https://github.com/QwenLM/Qwen2.5-Coder

Huggingface: https://huggingface.co/collections/Qwen/qwen25-coder-66eaa22e6f99801bf65b0c2f
Additionally, it provides two scenarios, code assistants and Artifacts, for exploration.

Code Assistants: https://huggingface.co/spaces/Qwen/Qwen2.5-Coder-demo
Artifacts: https://huggingface.co/spaces/Qwen/Qwen2.5-Coder-Artifacts