Replicate Intelligence #12
Flux LoRAs, Hot Zuck, and Replicate on Lex Fridman
Welcome to Replicate’s weekly bulletin! Each week, we’ll bring you updates on the latest open-source AI models, tools, and research. People are making cool stuff and we want to share it with you. Without further ado, here’s our hacker-in-residence deepfates with an unfiltered take on the week in AI.
Editor’s note
Earlier this week I turned into Hot Mark Zuckerberg. I just did it to test this deepfake library (which turned out to not even be open source after all but oh well) but now I feel weird about the future of Reality.
You know how Zuck has been rehabilitating his PR image? He grew out his hair and started dressing cool. And there was an image that went viral where they used a filter to give him a beard.
Well, I have curly hair and a beard, so I used it as a test image. I said “It’s me, Hot Mark Zuckerberg” and posted the video on X as a gag. A lot of people engaged with it — and to my surprise, many of them thought it was real.
So I’m getting these replies from people who think it’s just a video of Mark Zuckerberg, and I start to wonder, do I look that much like him? And I look in the mirror and for a second I don’t recognize myself. Shouldn’t I look… more like Zuck?
As I wrote last week, I believe we’re headed for a multiverse of virtual realities, synthetic worlds generated in parallel and populated with digital beings. We’re going to have millions of ‘verses, each of which will feel real to the beings that inhabit it.
A lot of people are going to lose track of what is real and what is not. Arguably we already have: echo chambers, filter bubbles, video games, social media, the list goes on. We’re in the dreamtime now, a shifting blurring mess of hallucinated potentialities.
But then, maybe we already were. A preprint paper this week (see Research Radar) suggests that dreams act as a form of synthetic data, the brain generating virtual worlds and scenarios to improve learning and decision making. Dreams as a form of self-play.
The worlds we dream will inform our real world decisions. We’ll be able to simulate, say, climate forecasts in different carbon scenarios, and model them more accurately and more vividly with every extra bit of data and compute. We are the model makers, and we are the dreamers of dreams.
What is reality? As ever, it is what we make it.
— deepfates
Trending models
FLUX.1: Exploring new frontiers in image generation
FLUX.1 continues to push the boundaries of what’s possible in image generation. This week, we’re highlighting two exciting developments:
-
ReFlux: An interface that allows you to use multiple fine-tuned FLUX.1 models and generate images on an infinite canvas. You can explore various styles and concepts simultaneously, opening up new creative possibilities.
-
Multi-LoRA Explorer: This tool lets you apply multiple LoRAs (Low-Rank Adaptations) in a single image, allowing for complex style combinations. Use it with LoRAs from Replicate, CivitAI, or Hugging Face.
You can use LoRAs to train image models on your own face, then combine them with other styles. On X this week we saw a great tutorial thread by community member Justine Moore, check it out if you want to learn how.
ReFlux | Multi LoRA Explorer
Cool tools
Real-time deepfake with one photo
Deep-Live-Cam is an open-source project that makes it easy to deepfake a face from a photo onto the head of a person in a video. Supposedly it has real-time face swapping; I couldn’t run it at a satisfying frame rate on a MacBook Pro.
If I succeed with real-time Hot Mark Zuckerberg, you will hear about it.
github
Research radar
Brains do self-play while we sleep
A recent study published in bioRxiv sheds light on how our brains might be “fine-tuning” themselves during sleep. The research suggests that during REM (Rapid Eye Movement) sleep, an animal’s brain simulates actions and their consequences, creating a form of synthetic data for self-improvement.
Key findings:
-
The brain’s internal representation of direction shifts during REM sleep, similar to an awake animal moving through its environment.
-
The superior colliculus (whatever that is) issues motor commands during REM sleep, despite the body’s physical immobility.
-
These “virtual” motor commands shift the internal representation of direction as if the action had been performed.
There’s some parallels here to how AI models are trained and fine-tuned. Perhaps our brains are using a form of synthetic data generation to enhance learning and decision-making capabilities.
paper
Community spotlight
Replicate mentioned on Lex Fridman podcast
Pieter Levels, founder of PhotoAI, Nomad List, and RemoteOK, recently appeared on the Lex Fridman podcast, where he discussed various topics including AI, viral startups, and the digital nomad lifestyle.
Levels gave a shoutout to Replicate, highlighting its usefulness for AI engineers and developers.
Some of what they talked about:
-
The growing importance of AI in modern software development
-
How tools like Replicate are making AI more accessible to developers
It’s cool to hear how Replicate is empowering AI engineers and democratizing access to AI models.
podcast | youtube
Bye for now
I’ll be gone next week. Expect your next newsletter in September. That is, if you believe I was ever real at all…
--- deepfates
Sign in to highlight and annotate this article

Conversation starters
Daily AI Digest
Get the top 5 AI stories delivered to your inbox every morning.
Knowledge Map
Connected Articles — Knowledge Graph
This article is connected to other articles through shared AI topics and tags.
More in Open Source AI
v4.3.1
Changes Gemma 4 support with full tool-calling in the API and UI. 🆕 ik_llama.cpp support : Add ik_llama.cpp as a new backend through new textgen-portable-ik portable builds and a new --ik flag for full installs. ik_llama.cpp is a fork by the author of the imatrix quants, including support for new quant types, significantly more accurate KV cache quantization (via Hadamard KV cache rotation, enabled by default), and optimizations for MoE models and CPU inference. API: Add echo + logprobs for /v1/completions . The completions endpoint now supports the echo and logprobs parameters, returning token-level log probabilities for both prompt and generated tokens. Token IDs are also included in the output via a new top_logprobs_ids field. Further optimize my custom gradio fork, saving up to 50 ms

From SWE-ZERO to SWE-HERO: Execution-free to Execution-based Fine-tuning for Software Engineering Agents
arXiv:2604.01496v1 Announce Type: new Abstract: We introduce SWE-ZERO to SWE-HERO, a two-stage SFT recipe that achieves state-of-the-art results on SWE-bench by distilling open-weight frontier LLMs. Our pipeline replaces resource-heavy dependencies with an evolutionary refinement strategy: (1) SWE-ZERO utilizes large-scale, execution-free trajectories to master code semantics and repository-level reasoning, and (2) SWE-HERO applies targeted, execution-backed refinement to transition these semantic intuitions into rigorous engineering workflows. Our empirical results set a new benchmark for open-source models of comparable size. We release a dataset of 300k SWE-ZERO and 13k SWE-HERO trajectories distilled from Qwen3-Coder-480B, alongside a suite of agents based on the Qwen2.5-Coder series.

A Quick Note on Gemma 4 Image Settings in Llama.cpp
In my last post, I mentioned using --image-min-tokens to increase the quality of image responses from Qwen3.5 . I went to load Gemma 4 the same way, and hit an error: [58175] srv process_chun: processing image... [58175] encoding image slice... [58175] image slice encoded in 7490 ms [58175] decoding image batch 1/2, n_tokens_batch = 2048 [58175] /Users/socg/llama.cpp-b8639/src/llama-context.cpp:1597: GGML_ASSERT((cparams.causal_attn || cparams.n_ubatch > = n_tokens_all ) "non-causal attention requires n_ubatch >= n_tokens" ) failed [58175] WARNING: Using native backtrace. Set GGML_BACKTRACE_LLDB for more info. [58175] WARNING: GGML_BACKTRACE_LLDB may cause native MacOS Terminal.app to crash. [58175] See: https://github.com/ggml-org/llama.cpp/pull/17869 [58175] 0 libggml-base.0.9.11.dylib 0


Discussion
Sign in to join the discussion
No comments yet — be the first to share your thoughts!