The Adversarial Robustness Toolbox (ART) = framework that uses generative adversarial neural networks (GANs) to protect deep learning models from security attacks
Thread⬇️
GANs = the most popular form of generative models.
GAN-based attacks:
+White Box Attacks: The adversary has access to the training environment, knowledge of the training algorithm
+Black Box Attacks: The adversary has no additional knowledge
2/⬇️
The goal of ART = to provide a framework to evaluate the robustness of a neural network.
The current version of ART focuses on four types of adversarial attacks:
+evasion
+inference
+extraction
+poisoning
3/⬇️
ART is a generic Python library. It provides native integration with several deep learning frameworks such as @TensorFlow, @PyTorch, #Keras, @ApacheMXNet
If you'd like to find a concentrated coverage of ART, click the link below. You'll move to TheSequence Edge#7, our educational newsletter. thesequence.substack.com/p/edge7 5/5
• • •
Missing some Tweet in this thread? You can try to
force a refresh
Our top 2
▪️ Xattention
▪️ Inside-Out: Hidden Factual Knowledge in LLMs
▪️ Rwkv-7 "Goose"
▪️ ϕ-Decoding
▪️ Frac-connections
▪️ DAPO
▪️ Reinforcement learning for reasoning in small LLMs
▪️ MetaLadder
▪️ Measuring AI ability to complete long tasks
▪️ Why do multi-agent LLM systems fail?
▪️ Agents play thousands of 3D video games
▪️ GKG-LLM
▪️ Privacy, Synthetic Data, and Security
▪️ Scale-wise distillation of diffusion models
▪️ Multimodal chain-of-thought reasoning
▪️ Survey on evaluation of LLM-based agents
▪️ Stop overthinking: A survey on efficient reasoning
▪️ Aligning multimodal LLM with human preference
🧵
1. Xattention by @MIT, @Tsinghua_Uni, @sjtu1896 and @nvidia
Speeds up inference with block-sparse attention and antidiagonal scoring
There’s no single “right” answer for AI models in creative writing (like creating a story tale), and their open-ended thinking is a key part of creative intelligence.
Still, models often lack output diversity, so @midjourney dropped an interesting study on this 👇
▪️ Their idea is to add diversity directly into the training process:
They measured response deviation for the same prompt and used it to train with DPO and ORPO, leading to more diversified DDPO and DORPO methods.
Here's how DDPO and DORPO work:
1. Diversified DPO (DDPO):
In the regular DPO method, the model learns by comparing a better response to a worse one.
In diversified version, researchers add more weight to rare or unique winning responses—those with higher deviation.
This helps the model pay more attention to uncommon but high-quality examples during training.
2. Diversified ORPO (DORPO):
Here too, researchers use deviation to give more importance to standout winning responses.
But ORPO has a slightly different math formula, so they apply the deviation weight to both parts of the learning signal.
DiLoCo (Distributed Low-Communication) method by @GoogleAI and @GoogleDeepMind changes how training of models happens:
Instead of constant syncing, multiple copies of the model are trained in parallel and sync only occasionally.
Scaling laws show how DiLoCo works as models' size grows🧵
At its core, DiLoCo follows a 2-level optimization process:
• Inner optimization: Each model replica (M) trains independently, making local updates.
• Outer optimization: Every H steps, replicas sync their updates to adjust a global model, which is then shared with all replicas, repeating the cycle.
Here are scaling laws for DiLoCo:
1. DiLoCo scales predictably and efficiently across model sizes, often performing better than Data-Parallel training.
DiLoCo supports larger batch sizes than Data-Parallel.
▪️ @perplexity_ai expands beyond the web
▪️ Manus: a Chinese high-performing AI agent
▪️ @Apple delayed Siri AI enhancements and new M3 Ultra chip
▪️ @CorticalLabs' CL1 computer fuses human brain cells with silicon
▪️ @MistralAI OCR
▪️ Andrew Barto and @RichardSSutton take home the 2024 Turing Award!
Find the details below 🧵
1. @perplexity_ai expands beyond the web
It partners with hardware firms to integrate its AI into everyday devices. This year, Deutsche Telekom’s AI Phone launches with Perplexity’s assistant, hinting at future moves. Phones for now, then TVs? Where next?
2. Manus: a Chinese high-performing AI agent from Monica.ai
Built on Claude Sonnet, it outperforms OpenAI and Anthropic on key benchmarks. Founded by Xiao Hong, it grew from a plugin to a $100M startup, targeting global markets and sidestepping China's AI rules. Focusing on business over AGI purists, it monetizes user data. Manus, exclusive and invite-only, might reshape China's global AI strategy.