TuringPost Profile picture
Jun 26, 2021 5 tweets 3 min read Read on X
The Adversarial Robustness Toolbox (ART) = framework that uses generative adversarial neural networks (GANs) to protect deep learning models from security attacks

Thread⬇️
GANs = the most popular form of generative models.

GAN-based attacks:
+White Box Attacks: The adversary has access to the training environment, knowledge of the training algorithm
+Black Box Attacks: The adversary has no additional knowledge
2/⬇️
The goal of ART = to provide a framework to evaluate the robustness of a neural network.

The current version of ART focuses on four types of adversarial attacks:
+evasion
+inference
+extraction
+poisoning
3/⬇️
ART is a generic Python library. It provides native integration with several deep learning frameworks such as @TensorFlow, @PyTorch, #Keras, @ApacheMXNet

@IBM open-sourced ART at github.com/IBM/adversaria….
4/⬇️
If you'd like to find a concentrated coverage of ART, click the link below. You'll move to TheSequence Edge#7, our educational newsletter.
thesequence.substack.com/p/edge7
5/5

• • •

Missing some Tweet in this thread? You can try to force a refresh
 

Keep Current with TuringPost

TuringPost Profile picture

Stay in touch and get notified when new unrolls are available from this author!

Read all threads

This Thread may be Removed Anytime!

PDF

Twitter may remove this content at anytime! Save it as PDF for later use!

Try unrolling a thread yourself!

how to unroll video
  1. Follow @ThreadReaderApp to mention us!

  2. From a Twitter thread mention us with a keyword "unroll"
@threadreaderapp unroll

Practice here first or read more on our help page!

More from @TheTuringPost

Aug 12
The freshest AI/ML research of the week

Our top 9
▪️ Sotopia-RL: Reward Design for Social Intelligence
▪️ Agent Lightning: Train ANY AI Agents with RL
▪️ Exploitation Is All You Need... for Exploration
▪️ Learning to Reason for Factuality
▪️ VeOmni
▪️ Is Chain-of-Thought Reasoning of LLMs a Mirage?
▪️ Cognitive Loop via In-Situ Optimization
▪️ Sculptor
▪️ CoAct-1

▪️ Tool-integrated Reinforcement Learning for Repo Deep Search
▪️ RL-PLUS
▪️ SEAgent
▪️ CRINN
▪️ Training Long-Context, Multi-Turn Software Engineering Agents with RL
▪️ Beyond the Trade-off: Self-Supervised RL for Reasoning Models' Instruction Following
▪️ CompassVerifier
▪️ Are We on the Right Way for Assessing Document Retrieval-Augmented Generation?
▪️ Are Today's LLMs Ready to Explain Well-Being Concepts?
▪️ VeriGUI
▪️ Trainable Dynamic Mask Sparse Attention
▪️ LeanK
▪️ Don't Overthink It: A Survey of Efficient R1-style Large Reasoning Models
▪️ On the Generalization of SFT
▪️ SitEmb-v1.5
▪️ AttnTrace
▪️ LaTCoder
▪️ ChartCap

🧵Image
1. Sotopia-RL: Reward Design for Social Intelligence

Trains socially intelligent agents with utterance-level, multi-dimensional rewards to capture nuanced social behaviors

arxiv.org/abs/2508.03905
Project page: rl.sotopia.world Image
2. Agent Lightning: Train ANY AI Agents with RL by @MSFTResearch

Provides a general framework for applying RL to any AI agent architecture with minimal integration overhead

arxiv.org/abs/2508.03680
GitHub: github.com/microsoft/agen… Image
Read 12 tweets
Jul 11
SingLoRA is a new simple version of LoRA (Low Rank Adaptation) by Technion that uses only one small matrix instead of usual two.

It multiplies it by its own transpose (like A × Aᵀ).

What does it buy you?

- No scale mismatch between different matrices
- Uses ~half the parameters of LoRA
- Stability and better learning

Here's how it works:Image
Image
1. Workflow of SingLoRA:

• The original weights of the model (W₀) are frozen.
• The system adds a small adapter - a learnable piece that updates the model for your specific task.
In SigLoRA, it's A × Aᵀ, where:
- A is a small trainable matrix with n × r size, where r ≪ n
- Aᵀ is its transpose
• The original model and the adapter are combined like this:Image
2. SingLoRA is extended for all layer shapes, whether they are:

- Square (same input/output size), like many attention layers
- Rectangular (input ≠ output size), like MLP layers
- Non-square (here “truncated” version of A is used so the shapes line up correctly).

The A × Aᵀ adapter still forms the update.
Read 5 tweets
Jul 1
The freshest AI/ML research papers of the week

Our top 7:

▪️ OctoThinker
▪️ Performance Prediction for Large Systems via Text-to-Text Regression
▪️ Radial Attention
▪️ MADrive
▪️ Mind2Web 2
▪️ Chain-of-Experts
▪️ Ark

▪️ Where to find Grokking
▪️ Skywork-SWE
▪️ BlenderFusion
▪️ OmniGen2
▪️ LLaVA-Scissor
▪️ MMSearch-R1
▪️ LongWriter-Zero
▪️ Steering Conceptual Bias
▪️ WorldVLA

🧵Image
Image
Image
1. OctoThinker

Improves reinforcement learning alignment via mid-training strategies and math-intensive corpora

arxiv.org/abs/2506.20512
GitHub: github.com/GAIR-NLP/OctoT… Image
2. Performance Prediction for Large Systems via Text-to-Text Regression, by @Google

Models system behavior from logs and configs using text-to-text LLMs that outperform tabular regressors with few-shot adaptation

arxiv.org/abs/2506.21718
Code: github.com/google-deepmin… Image
Read 19 tweets
Jun 28
30 days, 15 AI Coding Agents, one prompt — and the results will surprise you!

Will Schenk, TheFocusAI, specially for Turing Post tested which coding tool could best build a Dockerized idea app with voting, notes, and file attachments.

You would not believe what he discovered about Cursor, v0, Copilot, and 12 others 🧵Image
1. Aider @aider_chat
This free, open-source CLI cranks out solid code faster than GitHub’s $20/month Copilot.

Grab the full June 2025 Coding Agent Report for code quality, testing, and more surprising and useful details to know what agent to hire -> github.com/The-Focus-AI/j…
2. Claude Code @AnthropicAI
Its brainy context compaction and 4/5 testing score make it a pro-grade choice for production-ready apps - hiring recommended!
Read 17 tweets
Jun 27
Chain-of-Experts (CoE) - a new kind of model architecture.

It builds on Mixture-of-Experts (MoE) idea that a model can choose a different expert each round.

➡️ As a new addition, experts work in a sequence, one after the other
within a layer.

CoE keeps the number of active experts the same as before, but:

- Uses up to 42% less memory
- Unlocks over 800× more effective expert combinations
- Improves performance

Here's how it works:Image
1. In CoE:

- The model picks a small group of experts.
- Each expert transforms the current hidden state of a token.
- The outputs are combined using gating weights.
- A residual connection helps keep the information stable.

So, the final result is the token after it's been processed by C rounds of experts, with each round learning from the last.Image
2. Adaptive routing:

Each iteration has its own router, so the model can "change its mind" about which experts to use as it learns more. For example:

- In the first step, it might send the token to general experts.
- In later steps, it can route to more specialized ones, depending on how the token has evolved.
Read 6 tweets
Jun 26
Models, datasets and benchmarks to pay attention to:

▪️ Gemini 2.5 Flash and Pro, plus Gemini 2.5 Flash-Lite
▪️ MiniMax-M1
▪️ Kimi-Dev-72B

▪️ SHADE-Arena benchmark
▪️ ESSENTIAL-WEB V1.0 dataset

🧵 Image
1. @Google introduced Gemini 2.5 Flash and Pro as stable and production-ready, and launched Gemini 2.5 Flash-Lite in preview – the fastest and most cost-efficient.

Flash-Lite outperforms 2.0 Flash-Lite in coding, math, science, reasoning, and multimodal benchmarks. It features lower latency, supports 1 million-token context, multimodal input, and connects to tools like Google Search and code execution

storage.googleapis.com/deepmind-media…Image
Read 7 tweets

Did Thread Reader help you today?

Support us! We are indie developers!


This site is made by just two indie developers on a laptop doing marketing, support and development! Read more about the story.

Become a Premium Member ($3/month or $30/year) and get exclusive features!

Become Premium

Don't want to be a Premium member but still want to support us?

Make a small donation by buying us coffee ($5) or help with server cost ($10)

Donate via Paypal

Or Donate anonymously using crypto!

Ethereum

0xfe58350B80634f60Fa6Dc149a72b4DFbc17D341E copy

Bitcoin

3ATGMxNzCUFzxpMCHL5sWSt4DVtS8UqXpi copy

Thank you for your support!

Follow Us!

:(