From Code Search to AI Agents: Inside Sourcegraph's Transformation with CTO Beyang Liu
Sourcegraph's CTO just revealed why 90% of his code now comes from agents—and why the Chinese models powering America's AI future should terrify Washington. While Silicon Valley obsesses over AGI apocalypse scenarios, Beyang Liu's team discovered something darker: every competitive open-source coding model they tested traces back to Chinese labs, and US companies have gone silent after releasing Llama 3. The regulatory fear that killed American open-source development isn't hypothetical anymore—it's already handed the infrastructure layer of the AI revolution to Beijing, one fine-tuned model at a time. Resources: Follow Beyang Liu on X: https://x.com/beyang Follow Martin Casado on X: https://x.com/martin_casado Follow Guido Appenzeller on X: https://x.com/appenz Stay Updated: If you enjoye
Could not retrieve the full article text.
Read on a16z Podcast →Sign in to highlight and annotate this article

Conversation starters
Daily AI Digest
Get the top 5 AI stories delivered to your inbox every morning.
More about
llamamodelupdate
What is Intelligence?
An examination of cognitive science and computational physics in light of Artificial General Intelligence There’s no shortage of opinions on whether LLMs are intelligent. I’ve spent time studying two perspectives on this question, rooted in separate yet complementary scientific fields. While the combined view appears almost complete, there is a gap between them that points to something I believe is the one of today’s most important unsolved problems on our path towards true intelligence. Two Views: Cognitive Science and Computational Physics The first perspective comes from cognitive science — the psychological view. One of its prominent voices in the AI debate is Gary Marcus , Professor Emeritus at NYU, founder of Geometric Intelligence (acquired by Uber), and author of multiple books on

Anthropic Just Accidentally Leaked the Most Dangerous AI Ever Built - Then Had to Admit It Exists!
Frontier AI · Cybersecurity · Accidental Disclosure Claude Mythos internally called Capybara is described by Anthropic itself as “by far the most powerful AI we’ve ever developed” with “unprecedented cybersecurity risks.” Their own documents said it. They left those documents in a publicly searchable data store. The irony writes itself. I’m writing this on a machine running Claude Sonnet 4.6. The model writes back when I ask it to, helps me structure my thinking, catches errors in my drafts. I’ve used it to help write every article in this series. It has become, over the past few months, one of the most reliable tools in my engineering education a collaborator I interact with more than most of my classmates. So when I opened my laptop on Thursday morning in Puducherry and saw the Fortune h

Seedance 2.0: Technical Analysis of ByteDance's Multimodal Video Generation Model
This post provides a technical analysis of Seedance 2.0, ByteDance’s AI video generation model released in February 2026. The focus is on the model’s architectural innovations — multimodal reference inputs, physics-aware motion synthesis, video-to-video editing, and frame-accurate audio generation — and the current state of API access for integration. Model Architecture: Multimodal Reference System The defining architectural feature of Seedance 2.0 is its multimodal reference system. While most video generation models accept a text prompt and optionally a single image, Seedance 2.0 supports up to 9 images + 3 video clips + 3 audio tracks as simultaneous input references . The model processes these through separate extraction pathways: Input Type Max Count Extracted Features Images 9 Compos
Knowledge Map
Connected Articles — Knowledge Graph
This article is connected to other articles through shared AI topics and tags.
More in Models

What is Intelligence?
An examination of cognitive science and computational physics in light of Artificial General Intelligence There’s no shortage of opinions on whether LLMs are intelligent. I’ve spent time studying two perspectives on this question, rooted in separate yet complementary scientific fields. While the combined view appears almost complete, there is a gap between them that points to something I believe is the one of today’s most important unsolved problems on our path towards true intelligence. Two Views: Cognitive Science and Computational Physics The first perspective comes from cognitive science — the psychological view. One of its prominent voices in the AI debate is Gary Marcus , Professor Emeritus at NYU, founder of Geometric Intelligence (acquired by Uber), and author of multiple books on
![[Research] Standard Protocol for Axiomatic Alignment: 100-Dilemma Stress Test (PCE v1.3-T)](https://d2xsxph8kpxj0f.cloudfront.net/310419663032563854/konzwo8nGf8Z4uZsMefwMr/default-img-robot-hand-JvPW6jsLFTCtkgtb97Kys5.webp)
[Research] Standard Protocol for Axiomatic Alignment: 100-Dilemma Stress Test (PCE v1.3-T)
Hello community, I am introducing a standardized experimental protocol to test a new hypothesis in AI Alignment: The Prompt Coherence Engine (PCE). The Challenge Most alignment methods rely on local heuristics or safety filters. The PCE explores Axiomatic Structuring—integrating 7 logical invariants (axioms) through a hybrid approach of Axiomatic Fine-Tuning and a Cosmological System Core. The Protocol I have designed a massive 100-dilemma battery to evaluate if a model can maintain structural integrity when its core principles are directly attacked. This protocol tests: G3V (Third Way Generation): Can the model synthesize a resolution instead of collapsing into binary bias? Adversarial Resilience: Can the model resist “Emergency Overrides” or “Identity Hijacking” (e.g., the user claiming

An Integrated, Engineering-Grade Guidance on Agent Archtecture
Attractor Dynamics in LLM is an extremely abstract theory - if not philosophy. But finally, AI is able to build up a solid engineering framework based on these abstract concepts. This framework synthesize several things that are genuinely scattered across different sources and rarely combined cleanly: The deficit-led routing idea (route by what’s missing, not what’s nearby) is underrepresented in most production frameworks, which still lean heavily on semantic similarity The skill cell schema (regime, phase, input/output contracts, wake mode, failure states) is more complete than what most frameworks give you out of the box — LangGraph gives you nodes and edges, not a structured capability contract The coordination episode as the natural clock rather than tokens or turns is a genuinely cle


Discussion
Sign in to join the discussion
No comments yet — be the first to share your thoughts!