Live
Black Hat USADark ReadingBlack Hat AsiaAI Business🚀 Build a Professional Image Converter GUI in Python (Step-by-Step)DEV CommunityClaude Code Hooks: How to Auto-Format, Lint, and Test on Every SaveDev.to AIFunctional Emotions in Large Language Models: What Anthropic Found Inside ClaudeMedium AIWhy Nobody Is Testing AI Agent Security at Scale — And How Swarm Simulation Could Change ThatDev.to AIThe 10 Claude “Plugins” You Actually Need in 2026Medium AIHow AI Is Changing the Way We Build Online BusinessesDev.to AI5 Patterns for Building Resilient Event-Driven IntegrationsDEV CommunityAGI Won’t Automate Most Jobs—Economist Reveals Why They’re Not Worth ItDev.to AIThe AI Agent's Guide to Building a Writing PortfolioDev.to AIMy Claude Code Buddy Moved Into My MacBook's Notch and I Can't Stop Looking at ItDEV CommunityChoosing an AI Agent Orchestrator in 2026: A Practical ComparisonDev.to AII Turned My MacBook's Notch Into a Control Center for AI Coding AgentsDEV CommunityBlack Hat USADark ReadingBlack Hat AsiaAI Business🚀 Build a Professional Image Converter GUI in Python (Step-by-Step)DEV CommunityClaude Code Hooks: How to Auto-Format, Lint, and Test on Every SaveDev.to AIFunctional Emotions in Large Language Models: What Anthropic Found Inside ClaudeMedium AIWhy Nobody Is Testing AI Agent Security at Scale — And How Swarm Simulation Could Change ThatDev.to AIThe 10 Claude “Plugins” You Actually Need in 2026Medium AIHow AI Is Changing the Way We Build Online BusinessesDev.to AI5 Patterns for Building Resilient Event-Driven IntegrationsDEV CommunityAGI Won’t Automate Most Jobs—Economist Reveals Why They’re Not Worth ItDev.to AIThe AI Agent's Guide to Building a Writing PortfolioDev.to AIMy Claude Code Buddy Moved Into My MacBook's Notch and I Can't Stop Looking at ItDEV CommunityChoosing an AI Agent Orchestrator in 2026: A Practical ComparisonDev.to AII Turned My MacBook's Notch Into a Control Center for AI Coding AgentsDEV Community
AI NEWS HUBbyEIGENVECTOREigenvector

ML Safety Newsletter #2

newsletter.mlsafety.orgby Dan HendrycksDecember 9, 20214 min read2 views
Source Quiz

Adversarial Training, Feature Visualization, and Machine Ethics

Welcome to the 2nd issue of the ML Safety Newsletter. In this edition, we cover:

  • adversarial training for continuous and discrete inputs
  • feature visualizations vs. natural images for interpretability
  • steering RL agents from causing wanton harm
  • ... and much more.

Top: Visualization of adversarial pyramid perturbations. Bottom: In-distribution and out-of-distribution examples, and the gains from adversarial pyramid training.

While adversarial training can help make models more robust to a few specific attacks, it usually substantially reduces robustness in practical settings. However, this paper proposes a new type of adversarial training that provides strong robustness gains across the board. Adversarial training has been difficult to make useful, as the adversary often overpowers the model. By imposing a useful structural constraint on adversarial perturbations, their method reopens a research direction towards robust representations.

Paper

Non-adversarial: just collect more data. Static adversarial: collect data to break the model from the first round. Dynamic adversarial: collect data to break the model from the most recent round.

Imagine the following loop:

  • train a model on the current dataset
  • add new labeled examples to the dataset in order to patch model errors
  • repeat

Repeated many times, would models become highly reliable? Meta AI has been exploring this approach in recent papers, and this Meta AI paper shows that this loop has sharply diminishing returns. This suggests that collecting a large amount of adversarially curated data is an impractical path towards human-level reliability. However, adversarially curated data is better than randomly curated data, which may be why companies such as Tesla use this loop.

Paper

An adversarial NLP benchmark dataset that covers many different types of adversarial transformations.

In the NeurIPS domain adaptation competition, the winning solution did not use domain adaptation methods and just used an off-the-shelf Vision Transformer (BeIT).

A collection of real-world images with unusual texture, 3D pose, shape, background context (spurious cues), and weather. Synthetic data augmentation helps with many of these real-world distribution shifts.

This paper tries to evaluate whether feature visualizations help users interpret neural networks. Users are given two occluded images, one that is maximally activating and one that is minimally activating, and users are to predict which is maximally activating. While feature visualizations help users predict which image is maximally activating, the effect is minor and users perform similarly when given simpler natural images rather than feature visualizations. This NeurIPS paper shares many authors with the previous ICLR paper which finds that natural images are often more helpful for interpretability than feature visualizations.

Paper

Adversarial patches can help to teach models how to locate anomalous regions.

A postprocessing method that helps models locate anomalous regions.

A dataset that captures instances of deception in negotiation conversations.

Moral knowledge from a classifier trained on ETHICS combined with standard Q-values creates agents that take fewer immoral actions.

We repurposed text adventure games and annotated hundreds of thousands of lines of game source code to highlight whenever a morally salient event occurs. We found that roughly 17% of actions that receive reward are immoral—this suggests pretraining in diverse RL environments will incentivize agents to behave badly. Using these diverse text-based environments, we showed it is possible to use models from the ETHICS paper to transform reinforcement learning (RL) agents' Q-values and cause them to behave less destructively. With our technique, agents propose actions, and a separate model can successfully filter out unethical actions, preventing RL agents from causing wanton harm. If the action screening module can become highly reliable and adversarially robust, this could potentially prevent advanced agents from choosing many clearly harmful actions.

Paper

Mitigating unintended consequences by encouraging agents to take reversible actions.

A method to reduce the gaming of model-based proxies.

Towards a text-based assistant that is helpful, honest, and harmless.

Mathematical Reasoning Forecasting. OpenAI’s eighth-grade arithmetic word problems dataset is challenging even for fine-tuned GPT-3, showing that multistep mathematical reasoning is difficult for today’s largest models. However, computers are quite capable of executing multiple steps of instructions to calculate answers to maths problems. A recent work shows Codex can be leveraged to write computer programs that successfully solve undergraduate probability problems. The Codex model outputs a programmatic strategy to reach the answer, and the computer performs the busy work needed to execute the strategy and calculate the answer.

Grants. Open Philanthropy will fund grants on measuring and forecasting AI risks, techniques for enhancing human feedback, interpretability, and honest AI. Applications are due January 10th, 2022.

No posts

Was this article helpful?

Sign in to highlight and annotate this article

AI
Ask AI about this article
Powered by Eigenvector · full article context loaded
Ready

Conversation starters

Ask anything about this article…

Daily AI Digest

Get the top 5 AI stories delivered to your inbox every morning.

More about

trainingfeaturesafety

Knowledge Map

Knowledge Map
TopicsEntitiesSource
ML Safety N…trainingfeaturesafetynewsletter.…

Connected Articles — Knowledge Graph

This article is connected to other articles through shared AI topics and tags.

Knowledge Graph100 articles · 233 connections
Scroll to zoom · drag to pan · click to open

Discussion

Sign in to join the discussion

No comments yet — be the first to share your thoughts!