The Adversarial Robustness Toolbox (ART) = framework that uses generative adversarial neural networks (GANs) to protect deep learning models from security attacks
Thread⬇️
GANs = the most popular form of generative models.
GAN-based attacks:
+White Box Attacks: The adversary has access to the training environment, knowledge of the training algorithm
+Black Box Attacks: The adversary has no additional knowledge
2/⬇️
The goal of ART = to provide a framework to evaluate the robustness of a neural network.
The current version of ART focuses on four types of adversarial attacks:
+evasion
+inference
+extraction
+poisoning
3/⬇️
ART is a generic Python library. It provides native integration with several deep learning frameworks such as @TensorFlow, @PyTorch, #Keras, @ApacheMXNet
If you'd like to find a concentrated coverage of ART, click the link below. You'll move to TheSequence Edge#7, our educational newsletter. thesequence.substack.com/p/edge7 5/5
• • •
Missing some Tweet in this thread? You can try to
force a refresh
1. TRM is built on the idea of the Hierarchical Reasoning Model (HRM).
HRM uses 2 small neural networks working together, each at its own rhythm, to successfully solve hard problems like Sudoku, mazes, and ARC-AGI puzzles, though it’s tiny (27 million parameters).
TRM is a simpler, smaller alternative to HRM.
2. No more complex math:
HRM depends on a mathematical “fixed-point” assumption to simplify gradients, assuming that its recursive loops converge to a stable state.
On the contrary, TRM just runs the full recursion several times and backpropagates through all steps.
This removes the need for theoretical constraints and gives a huge boost in generalization: 56.5% → 87.4% on Sudoku-Extreme.
Retrieval-of-Thought (RoT) makes reasoning models faster by reusing earlier reasoning steps as templates.
These steps are stored in a “thought graph” that shows both their order and meaning.
As a result, RoT:
- reduces output tokens by up to 40%
- speeds up inference by 82%
- lowers cost by 59%
All without losing accuracy.
Here is how it works:
RoT works by:
- Storing reasoning steps as nodes in a “thought graph.”
- Retrieving relevant steps when a new problem comes in.
- Assembling a dynamic template from those steps to guide the model.
Let’s take it step by step
1. Building the "thought graph"
Researchers collected a large set of reasoning templates (3.34k). Each step in these templates became a node in the graph, with metadata like topic tags: algebra, geometry, etc.
- Sequential edges connect steps in the natural order within a template.
- Semantic edges connect steps that mean similar things across different templates.
So this graph acts like a memory bank of reasoning fragments.
1. Intern-s1: A scientific multimodal foundation model by Shanghai AI Lab (open-source)
This is a 241B-parameter multimodal Mixture-of-Experts model with 28B active parameters, optimized for scientific reasoning:
- Trained on 5T tokens (2.5T scientific)
- Supports text, images, molecular structures, and time-series data.
- Has a dynamic tokenizer and Mixture-of-Rewards RL framework
- Outperforms both open- and closed-source models on MatBench, ChemBench, etc.
It's a 9B hybrid Mamba-Transformer LLM optimized for reasoning:
- 3–6× higher throughput than Qwen3-8B
- Matches or exceeds its accuracy across benchmarks like MATH (80.5), BFCLv3, RULER-128k, AIME24
- FP8 pretraining on 20T tokens with 128k context
- Runs on a single 22GB A10G GPU
Our top 9
▪️ Sotopia-RL: Reward Design for Social Intelligence
▪️ Agent Lightning: Train ANY AI Agents with RL
▪️ Exploitation Is All You Need... for Exploration
▪️ Learning to Reason for Factuality
▪️ VeOmni
▪️ Is Chain-of-Thought Reasoning of LLMs a Mirage?
▪️ Cognitive Loop via In-Situ Optimization
▪️ Sculptor
▪️ CoAct-1
▪️ Tool-integrated Reinforcement Learning for Repo Deep Search
▪️ RL-PLUS
▪️ SEAgent
▪️ CRINN
▪️ Training Long-Context, Multi-Turn Software Engineering Agents with RL
▪️ Beyond the Trade-off: Self-Supervised RL for Reasoning Models' Instruction Following
▪️ CompassVerifier
▪️ Are We on the Right Way for Assessing Document Retrieval-Augmented Generation?
▪️ Are Today's LLMs Ready to Explain Well-Being Concepts?
▪️ VeriGUI
▪️ Trainable Dynamic Mask Sparse Attention
▪️ LeanK
▪️ Don't Overthink It: A Survey of Efficient R1-style Large Reasoning Models
▪️ On the Generalization of SFT
▪️ SitEmb-v1.5
▪️ AttnTrace
▪️ LaTCoder
▪️ ChartCap
🧵
1. Sotopia-RL: Reward Design for Social Intelligence
Trains socially intelligent agents with utterance-level, multi-dimensional rewards to capture nuanced social behaviors
SingLoRA is a new simple version of LoRA (Low Rank Adaptation) by Technion that uses only one small matrix instead of usual two.
It multiplies it by its own transpose (like A × Aᵀ).
What does it buy you?
- No scale mismatch between different matrices
- Uses ~half the parameters of LoRA
- Stability and better learning
Here's how it works:
1. Workflow of SingLoRA:
• The original weights of the model (W₀) are frozen.
• The system adds a small adapter - a learnable piece that updates the model for your specific task.
In SigLoRA, it's A × Aᵀ, where:
- A is a small trainable matrix with n × r size, where r ≪ n
- Aᵀ is its transpose
• The original model and the adapter are combined like this:
2. SingLoRA is extended for all layer shapes, whether they are:
- Square (same input/output size), like many attention layers
- Rectangular (input ≠ output size), like MLP layers
- Non-square (here “truncated” version of A is used so the shapes line up correctly).