OpenAI to Launch ChatGPT 5.5 and a New Unified Desktop Super App - Geeky Gadgets
OpenAI to Launch ChatGPT 5.5 and a New Unified Desktop Super App Geeky Gadgets
Could not retrieve the full article text.
Read on GNews AI Huawei →Sign in to highlight and annotate this article

Conversation starters
Daily AI Digest
Get the top 5 AI stories delivered to your inbox every morning.
More about
launchchatgpt
d318 is almost always suppressive in Qwen-2.5-3B emotional vectors, built an emotion vector steering pipeline, positive steering collapses to a single 'preschool teacher' register regardless of emotion
It appears that on lower weight models, behavior converges to either be highly sycophantic or neutral with no real in between, however existentialism did seem to be somewhat present. Using some heatmaps and visualizations, the cosine similarities between emotions appears coherent with what'd be expected, and there's really interesting dimensional dominances. In Qwen-2.5-3B, d318 is almost always the greatest in magnitude and almost always suppressive. Could be interesting for interpretability research. Vector merging also appears to lead to model incoherence if you merge a lot of vectors without normalizing their influences to some maximum. Built an automated emotion vector pipeline on top of Anthropic's emotional vector research . It makes the detection and correction of unwanted behavior
Knowledge Map
Connected Articles — Knowledge Graph
This article is connected to other articles through shared AI topics and tags.
More in Models

The Geometric Alignment Tax: Tokenization vs. Continuous Geometry in Scientific Foundation Models
arXiv:2604.04155v1 Announce Type: cross Abstract: Foundation models for biology and physics optimize predictive accuracy, but their internal representations systematically fail to preserve the continuous geometry of the systems they model. We identify the root cause: the Geometric Alignment Tax, an intrinsic cost of forcing continuous manifolds through discrete categorical bottlenecks. Controlled ablations on synthetic dynamical systems demonstrate that replacing cross-entropy with a continuous head on an identical encoder reduces geometric distortion by up to 8.5x, while learned codebooks exhibit a non-monotonic double bind where finer quantization worsens geometry despite improving reconstruction. Under continuous objectives, three architectures differ by 1.3x; under discrete tokenizatio

A Bayesian Information-Theoretic Approach to Data Attribution
arXiv:2604.03858v1 Announce Type: cross Abstract: Training Data Attribution (TDA) seeks to trace model predictions back to influential training examples, enhancing interpretability and safety. We formulate TDA as a Bayesian information-theoretic problem: subsets are scored by the information loss they induce - the entropy increase at a query when removed. This criterion credits examples for resolving predictive uncertainty rather than label noise. To scale to modern networks, we approximate information loss using a Gaussian Process surrogate built from tangent features. We show this aligns with classical influence scores for single-example attribution while promoting diversity for subsets. For even larger-scale retrieval, we relax to an information-gain objective and add a variance correct

Scalable Variational Bayesian Fine-Tuning of LLMs via Orthogonalized Low-Rank Adapters
arXiv:2604.03388v1 Announce Type: cross Abstract: When deploying large language models (LLMs) to safety-critical applications, uncertainty quantification (UQ) is of utmost importance to self-assess the reliability of the LLM-based decisions. However, such decisions typically suffer from overconfidence, particularly after parameter-efficient fine-tuning (PEFT) for downstream domain-specific tasks with limited data. Existing methods to alleviate this issue either rely on Laplace approximation based post-hoc framework, which may yield suboptimal calibration depending on the training trajectory, or variational Bayesian training that requires multiple complete forward passes through the entire LLM backbone at inference time for Monte Carlo estimation, posing scalability challenges for deploymen



Discussion
Sign in to join the discussion
No comments yet — be the first to share your thoughts!