Alibaba rolls out Qwen3.6-Plus with stronger agentic AI and multimodal reasoning - Tech Critter
<a href="https://news.google.com/rss/articles/CBMidEFVX3lxTE9tdG40eHRCV2I1MTRIOHRNUzlyUWdLcEhJN1ZWSThhUHZTMkNwbGlQYlNoSkRJdVFUSTFkTGZITi10TnZXaDl0emt6bVhhYXZBcVZITDQzMmZTMF9EYWdIMjNOS0gyeGlsVW5YYnl4ZEJmQTFt?oc=5" target="_blank">Alibaba rolls out Qwen3.6-Plus with stronger agentic AI and multimodal reasoning</a> <font color="#6f6f6f">Tech Critter</font>
Could not retrieve the full article text.
Read on GNews AI multimodal →Sign in to highlight and annotate this article

Conversation starters
Daily AI Digest
Get the top 5 AI stories delivered to your inbox every morning.
More about
reasoningmultimodalagentic
Fine-tuned Gemma 4 E4B for structured JSON extraction from regulatory docs - 75% to 94% accuracy, notebook + 432 examples included
Gemma 4 dropped this week so I fine-tuned E4B for a specific task: extracting structured JSON (doc type, obligations, key fields) from technical and regulatory documents. https://preview.redd.it/v7yg80prpetg1.png?width=1026 format=png auto=webp s=517fb50868405f90a94f60b54b04608bcedd2ced Results on held-out test set: - doc_type accuracy: 75% base → 94% fine-tuned - Hallucinated obligations: 1.25/doc → 0.59/doc - JSON validity: 100% - Field coverage: 100% Setup: - QLoRA 4-bit, LoRA r=16 alpha=16, Unsloth + TRL - 432 training examples across 8 doc types - 5 epochs on a single L4, ~10 min training time - Final train loss 1.04, eval loss 1.12 The whole thing is open: notebook, dataset, serve.py for FastAPI inference. https://github.com/spriyads-vault/gemma4-docparse Some things I learned the ha

Syntaqlite Playground
Tool: Syntaqlite Playground Lalit Maganti's syntaqlite is currently being discussed on Hacker News thanks to Eight years of wanting, three months of building with AI , a deep dive into exactly how it was built. This inspired me to revisit a research project I ran when Lalit first released it a couple of weeks ago, where I tried it out and then compiled it to a WebAssembly wheel so it could run in Pyodide in a browser (the library itself uses C and Rust). This new playground loads up the Python library and provides a UI for trying out its different features: formating, parsing into an AST, validating, and tokenizing SQLite SQL queries. Tags: sql , ai-assisted-programming , sqlite , tools , agentic-engineering

Gemma 4 Uncensored (autoresearch results)
Gemma 4 Uncensored — all 4 models, MoE expert abliteration, automated research loop Released uncensored versions of all four Gemma 4 models. bf16 + GGUF for each. Collection : https://huggingface.co/collections/TrevorJS/gemma-4-uncensored-69d2885d6e4fc0581f492698 Code : https://github.com/TrevorS/gemma-4-abliteration Results Model Baseline After KL Div E2B (2.3B) 98% 0.4% 0.346 E4B (4.5B) 99% 0.7% 0.068 26B MoE 98% 0.7% 0.090 31B 100% 3.2% 0.124 Refusal rates from 686 prompts across 4 datasets (JailbreakBench, tulu-harmbench, NousResearch, mlabonne). Manually audited — most flagged refusals are actually the model complying with a disclaimer attached. 26B MoE Standard abliteration only touches dense layers, which gets you from 98% → 29% on the MoE. The remaining refusals are in the expert w
Knowledge Map
Connected Articles — Knowledge Graph
This article is connected to other articles through shared AI topics and tags.
More in Frontier Research

What I like about MATS and Research Management
Crossposted on my personal blog . This is post number 16 in my second attempt at doing Inkkaven in a day, i.e. to write 30 blogposts in a single day. MATS is an organization that pairs up-and-coming AI Safety researchers (who I call participants) with the world’s best (this is not an exaggeration) existing AI Safety researchers (called mentors), for a minimum of 3 months research experience, followed by 6 or 12 months of further time to pursue their research further if they meet a minimum standard. The most common role at MATS, called research manager but I prefer the term research coach, is all about providing 1-1 support to the participants. The participant-mentor relationship is purely based on the research: by default they meet weekly for 30 minutes and only discuss what research has h




Discussion
Sign in to join the discussion
No comments yet — be the first to share your thoughts!