Gemma 4
https://huggingface.co/collections/google/gemma-4 submitted by /u/Namra_7 [link] [comments]
Could not retrieve the full article text.
Read on Reddit r/LocalLLaMA →Sign in to highlight and annotate this article

Conversation starters
Daily AI Digest
Get the top 5 AI stories delivered to your inbox every morning.
More about
huggingface
Fastest QWEN Coder 80B Next
I just used the new Apex Quantization on QWEN Coder 80B Created an Important Matrix using Code examples This should be the fastest best at coding 80B Next Coder around It's what I'm using for STACKS! so I thought I would share with the community It's insanely fast and the size has been shrunk down to 54.1GB https://huggingface.co/stacksnathan/Qwen3-Coder-Next-80B-APEX-I-Quality-GGUF https://preview.redd.it/wu924fls1dtg1.png?width=890 format=png auto=webp s=0a060e6868a5b88eabc5baa7b1ef266e096d480e submitted by /u/StacksHosting [link] [comments]

Error While using langchain with huggingface models
from langchain_core.prompts import PromptTemplate from langchain_community.llms import HuggingFaceEndpoint import os os.environ[“HUGGINGFACEHUB_API_TOKEN”] = “hf_your_new_token_here” prompt = PromptTemplate( input_variables=[“product”], template=“What is a good name for a company that makes {product}?” ) llm = HuggingFaceEndpoint( repo_id=“mistralai/Mistral-7B-Instruct-v0.3”, temperature=0.7, timeout=300 ) chains = prompt | llm print(“LLM Initialized with Token!”) try: response = chains.invoke({“product”: “camera”}) print(“AI Suggestion:”, response) except Exception as e: print(f"Error details: {e}") when i run this i get Value error can anyone help me out? Its a basic prompt template and text gen code but still it doesnt work i used various models from Huggingface and its not working well
Knowledge Map
Connected Articles — Knowledge Graph
This article is connected to other articles through shared AI topics and tags.
More in Open Source AI


What happened to MLX-LM? What are the alternatives?
Support seems non-existent and the last proper release was over a month ago. Comparing with llama.cpp, they are just miles different in activity and support. Is there an alternative or should just use llama.cpp for my macbook? submitted by /u/Solus23451 [link] [comments]

Fine-tuned Gemma 4 E4B for structured JSON extraction from regulatory docs - 75% to 94% accuracy, notebook + 432 examples included
Gemma 4 dropped this week so I fine-tuned E4B for a specific task: extracting structured JSON (doc type, obligations, key fields) from technical and regulatory documents. https://preview.redd.it/v7yg80prpetg1.png?width=1026 format=png auto=webp s=517fb50868405f90a94f60b54b04608bcedd2ced Results on held-out test set: - doc_type accuracy: 75% base → 94% fine-tuned - Hallucinated obligations: 1.25/doc → 0.59/doc - JSON validity: 100% - Field coverage: 100% Setup: - QLoRA 4-bit, LoRA r=16 alpha=16, Unsloth + TRL - 432 training examples across 8 doc types - 5 epochs on a single L4, ~10 min training time - Final train loss 1.04, eval loss 1.12 The whole thing is open: notebook, dataset, serve.py for FastAPI inference. https://github.com/spriyads-vault/gemma4-docparse Some things I learned the ha

Gemma 4 Uncensored (autoresearch results)
Gemma 4 Uncensored — all 4 models, MoE expert abliteration, automated research loop Released uncensored versions of all four Gemma 4 models. bf16 + GGUF for each. Collection : https://huggingface.co/collections/TrevorJS/gemma-4-uncensored-69d2885d6e4fc0581f492698 Code : https://github.com/TrevorS/gemma-4-abliteration Results Model Baseline After KL Div E2B (2.3B) 98% 0.4% 0.346 E4B (4.5B) 99% 0.7% 0.068 26B MoE 98% 0.7% 0.090 31B 100% 3.2% 0.124 Refusal rates from 686 prompts across 4 datasets (JailbreakBench, tulu-harmbench, NousResearch, mlabonne). Manually audited — most flagged refusals are actually the model complying with a disclaimer attached. 26B MoE Standard abliteration only touches dense layers, which gets you from 98% → 29% on the MoE. The remaining refusals are in the expert w


Discussion
Sign in to join the discussion
No comments yet — be the first to share your thoughts!