TuringPost Profile picture
Jun 26, 2021 5 tweets 3 min read Read on X
The Adversarial Robustness Toolbox (ART) = framework that uses generative adversarial neural networks (GANs) to protect deep learning models from security attacks

Thread⬇️
GANs = the most popular form of generative models.

GAN-based attacks:
+White Box Attacks: The adversary has access to the training environment, knowledge of the training algorithm
+Black Box Attacks: The adversary has no additional knowledge
2/⬇️
The goal of ART = to provide a framework to evaluate the robustness of a neural network.

The current version of ART focuses on four types of adversarial attacks:
+evasion
+inference
+extraction
+poisoning
3/⬇️
ART is a generic Python library. It provides native integration with several deep learning frameworks such as @TensorFlow, @PyTorch, #Keras, @ApacheMXNet

@IBM open-sourced ART at github.com/IBM/adversaria….
4/⬇️
If you'd like to find a concentrated coverage of ART, click the link below. You'll move to TheSequence Edge#7, our educational newsletter.
thesequence.substack.com/p/edge7
5/5

• • •

Missing some Tweet in this thread? You can try to force a refresh
 

Keep Current with TuringPost

TuringPost Profile picture

Stay in touch and get notified when new unrolls are available from this author!

Read all threads

This Thread may be Removed Anytime!

PDF

Twitter may remove this content at anytime! Save it as PDF for later use!

Try unrolling a thread yourself!

how to unroll video
  1. Follow @ThreadReaderApp to mention us!

  2. From a Twitter thread mention us with a keyword "unroll"
@threadreaderapp unroll

Practice here first or read more on our help page!

More from @TheTuringPost

Mar 25
The freshest AI/ML research of the week:

Our top 2
▪️ Xattention
▪️ Inside-Out: Hidden Factual Knowledge in LLMs

▪️ Rwkv-7 "Goose"
▪️ ϕ-Decoding
▪️ Frac-connections
▪️ DAPO
▪️ Reinforcement learning for reasoning in small LLMs
▪️ MetaLadder
▪️ Measuring AI ability to complete long tasks
▪️ Why do multi-agent LLM systems fail?
▪️ Agents play thousands of 3D video games
▪️ GKG-LLM
▪️ Privacy, Synthetic Data, and Security
▪️ Scale-wise distillation of diffusion models
▪️ Multimodal chain-of-thought reasoning
▪️ Survey on evaluation of LLM-based agents
▪️ Stop overthinking: A survey on efficient reasoning
▪️ Aligning multimodal LLM with human preference

🧵Image
Image
Image
1. Xattention by @MIT, @Tsinghua_Uni, @sjtu1896 and @nvidia

Speeds up inference with block-sparse attention and antidiagonal scoring

huggingface.co/papers/2503.16…
Code: github.com/mit-han-lab/x-… Image
2. Inside-Out by @TechnionLive and @GoogleAI

Reveals that internal LLM knowledge often surpasses what is actually generated

huggingface.co/papers/2503.15…
Read 20 tweets
Mar 24
7 open-source AI models of the week:

• @Microsoft’s KBLaM
• Fin-R1
• @nvidia’s Cosmos-Reason1
• @nvidia’s Cosmos-Transfer1
• M3 by @nvidia
• Tencent’s T1
• Roblox’ Cube

🧵 Image
Image
Image
1. @Microsoft’s KBLaM integrates structured knowledge into LLMs with rectangular attention for low-latency, hallucination-resistant answers.

microsoft.com/en-us/research…

Code and database: github.com/microsoft/KBLa… Image
2. Fin-R1

Trains a finance-specific LLM using CoT and RL to outperform larger models on reasoning benchmarks.

huggingface.co/papers/2503.16…
Code: github.com/SUFE-AIFLM-Lab… Image
Read 9 tweets
Mar 24
There’s no single “right” answer for AI models in creative writing (like creating a story tale), and their open-ended thinking is a key part of creative intelligence.

Still, models often lack output diversity, so @midjourney dropped an interesting study on this 👇

▪️ Their idea is to add diversity directly into the training process:

They measured response deviation for the same prompt and used it to train with DPO and ORPO, leading to more diversified DDPO and DORPO methods.

Here's how DDPO and DORPO work:Image
1. Diversified DPO (DDPO):

In the regular DPO method, the model learns by comparing a better response to a worse one.

In diversified version, researchers add more weight to rare or unique winning responses—those with higher deviation.

This helps the model pay more attention to uncommon but high-quality examples during training.
2. Diversified ORPO (DORPO):

Here too, researchers use deviation to give more importance to standout winning responses.

But ORPO has a slightly different math formula, so they apply the deviation weight to both parts of the learning signal.
Read 5 tweets
Mar 18
DiLoCo (Distributed Low-Communication) method by @GoogleAI and @GoogleDeepMind changes how training of models happens:

Instead of constant syncing, multiple copies of the model are trained in parallel and sync only occasionally.

Scaling laws show how DiLoCo works as models' size grows🧵Image
At its core, DiLoCo follows a 2-level optimization process:

• Inner optimization: Each model replica (M) trains independently, making local updates.

• Outer optimization: Every H steps, replicas sync their updates to adjust a global model, which is then shared with all replicas, repeating the cycle.

Here are scaling laws for DiLoCo:
1. DiLoCo scales predictably and efficiently across model sizes, often performing better than Data-Parallel training.

DiLoCo supports larger batch sizes than Data-Parallel.
Read 10 tweets
Mar 11
The latest AI/ML news of the week:

▪️ @perplexity_ai expands beyond the web
▪️ Manus: a Chinese high-performing AI agent
▪️ @Apple delayed Siri AI enhancements and new M3 Ultra chip
▪️ @CorticalLabs' CL1 computer fuses human brain cells with silicon
▪️ @MistralAI OCR
▪️ Andrew Barto and @RichardSSutton take home the 2024 Turing Award!

Find the details below 🧵Image
Image
Image
1. @perplexity_ai expands beyond the web

It partners with hardware firms to integrate its AI into everyday devices. This year, Deutsche Telekom’s AI Phone launches with Perplexity’s assistant, hinting at future moves. Phones for now, then TVs? Where next?

telekom.com/en/media/media…Image
2. Manus: a Chinese high-performing AI agent from Monica.ai

Built on Claude Sonnet, it outperforms OpenAI and Anthropic on key benchmarks. Founded by Xiao Hong, it grew from a plugin to a $100M startup, targeting global markets and sidestepping China's AI rules. Focusing on business over AGI purists, it monetizes user data. Manus, exclusive and invite-only, might reshape China's global AI strategy.

chinatalk.media/p/manus-chinas…Image
Read 8 tweets
Mar 10
6 notable AI models of the week:

▪️ Differentiable Logic Cellular Automata @GoogleAI
▪️ Phi-4-Mini @Microsoft
▪️ Babel, Open Multilingual LLMs @AlibabaGroup
▪️ Aya Vision @CohereForAI
▪️ LLMVoX
▪️ LanDiff by Moonshot AI

🧵 Image
Image
Image
1. Differentiable Logic Cellular Automata @GoogleAI

Integrates Neural Cellular Automata with Differentiable Logic Gate Networks to enable self-healing, pattern generation, and robust computational architectures.

google-research.github.io/self-organisin…Image
2. Phi-4-Mini Technical Report @Microsoft

Introduces a 3.8B parameter multimodal model using Mixture-of-LoRAs, excelling in math, coding, and reasoning while maintaining efficiency.

huggingface.co/papers/2503.01…
Model: huggingface.co/microsoft/Phi-… Image
Read 9 tweets

Did Thread Reader help you today?

Support us! We are indie developers!


This site is made by just two indie developers on a laptop doing marketing, support and development! Read more about the story.

Become a Premium Member ($3/month or $30/year) and get exclusive features!

Become Premium

Don't want to be a Premium member but still want to support us?

Make a small donation by buying us coffee ($5) or help with server cost ($10)

Donate via Paypal

Or Donate anonymously using crypto!

Ethereum

0xfe58350B80634f60Fa6Dc149a72b4DFbc17D341E copy

Bitcoin

3ATGMxNzCUFzxpMCHL5sWSt4DVtS8UqXpi copy

Thank you for your support!

Follow Us!

:(