Gemma 4 - 4B vs Qwen 3.5 - 9B ?
Hello! anyone tried the 4B Gemma 4 model and the Qwen 3.5 9B model and can tell us their feedback? On the benchmark Qwen seems to be doing better, but I would appreciate any personal experience on the matter Thanks! submitted by /u/No-Mud-1902 [link] [comments]
Could not retrieve the full article text.
Read on Reddit r/LocalLLaMA →Sign in to highlight and annotate this article

Conversation starters
Daily AI Digest
Get the top 5 AI stories delivered to your inbox every morning.
More about
modelbenchmark
thinking about running Gemma4 E2B as a preprocessor before every Claude Code API call. anyone see obvious problems with this?
background: I write mostly in Korean and my Claude API bill is kind of embarrassing. Korean tokenizes really inefficiently compared to English for the same meaning, so a chunk of the cost is basically just encoding overhead. the idea is a small proxy in Bun that sits in front of the Claude API. Claude Code talks to localhost, doesn't know anything changed. before each request goes out, Gemma4 E2B (llama.cpp, local) would do: - translate Korean input to English. response still comes back in Korean, just the outbound prompt is English - trim context that's probably not relevant to the current turn - for requests that look like they need reasoning, have Gemma4 do the thinking first and pass the result along — so the paid model hopefully skips some of that work and uses fewer reasoning tokens

Whats the best open source/free TTS
Hey, Im trying to see how much does synthetic data help with training ASR model. What is the best TTS? Im looking for something that sounds natural and not robotic. It would be really nice if the TTS could mimic english accents (american, british, french etc.). Thanks for the help. submitted by /u/NightMatko [link] [comments]
Knowledge Map
Connected Articles — Knowledge Graph
This article is connected to other articles through shared AI topics and tags.
More in Models

thinking about running Gemma4 E2B as a preprocessor before every Claude Code API call. anyone see obvious problems with this?
background: I write mostly in Korean and my Claude API bill is kind of embarrassing. Korean tokenizes really inefficiently compared to English for the same meaning, so a chunk of the cost is basically just encoding overhead. the idea is a small proxy in Bun that sits in front of the Claude API. Claude Code talks to localhost, doesn't know anything changed. before each request goes out, Gemma4 E2B (llama.cpp, local) would do: - translate Korean input to English. response still comes back in Korean, just the outbound prompt is English - trim context that's probably not relevant to the current turn - for requests that look like they need reasoning, have Gemma4 do the thinking first and pass the result along — so the paid model hopefully skips some of that work and uses fewer reasoning tokens

Whats the best open source/free TTS
Hey, Im trying to see how much does synthetic data help with training ASR model. What is the best TTS? Im looking for something that sounds natural and not robotic. It would be really nice if the TTS could mimic english accents (american, british, french etc.). Thanks for the help. submitted by /u/NightMatko [link] [comments]



Discussion
Sign in to join the discussion
No comments yet — be the first to share your thoughts!