TuringPost Profile picture
Jun 26, 2021 5 tweets 3 min read Read on X
The Adversarial Robustness Toolbox (ART) = framework that uses generative adversarial neural networks (GANs) to protect deep learning models from security attacks

Thread⬇️
GANs = the most popular form of generative models.

GAN-based attacks:
+White Box Attacks: The adversary has access to the training environment, knowledge of the training algorithm
+Black Box Attacks: The adversary has no additional knowledge
2/⬇️
The goal of ART = to provide a framework to evaluate the robustness of a neural network.

The current version of ART focuses on four types of adversarial attacks:
+evasion
+inference
+extraction
+poisoning
3/⬇️
ART is a generic Python library. It provides native integration with several deep learning frameworks such as @TensorFlow, @PyTorch, #Keras, @ApacheMXNet

@IBM open-sourced ART at github.com/IBM/adversaria….
4/⬇️
If you'd like to find a concentrated coverage of ART, click the link below. You'll move to TheSequence Edge#7, our educational newsletter.
thesequence.substack.com/p/edge7
5/5

• • •

Missing some Tweet in this thread? You can try to force a refresh
 

Keep Current with TuringPost

TuringPost Profile picture

Stay in touch and get notified when new unrolls are available from this author!

Read all threads

This Thread may be Removed Anytime!

PDF

Twitter may remove this content at anytime! Save it as PDF for later use!

Try unrolling a thread yourself!

how to unroll video
  1. Follow @ThreadReaderApp to mention us!

  2. From a Twitter thread mention us with a keyword "unroll"
@threadreaderapp unroll

Practice here first or read more on our help page!

More from @TheTuringPost

Jul 10
Let's dive into one of the newest concept of synthetic data generation -
active inheritance.

Proposed by @CohereForAI, it's a strategy used in ML to intentionally design synthetic data to achieve specific goals.

Here's how active inheritance works: Image
1. What's in the base?

The base is a knowledge distillation technique, where a smaller LLM (student) learns from a larger, more powerful model (teacher).

The student tries to mimic the teacher outputs for the same input prompts by learning from the data the teacher generates.
2. Measuring data characteristics:

Cohere researchers use specific functions to track length, diversity, and toxicity in outputs from both teacher and student LLMs. This helps them understand what properties the student inherits from the teacher. Image
Read 7 tweets
Jun 3
A new model, Meteor, leverages multifaceted information and a Mamba architecture to enhance comprehension and response capabilities in vision-language tasks.

Let's explore its architecture and training strategy👇 Image
1. Meteor's architecture includes:

- a vision encoder (CLIP-L/14)
- vision and tor projectors (MLP modules with GELU activation)
- the Mamba-130M architecture for computational efficiency
- the InternLM2-7B as the backbone LLM. Image
2. Visual instruction tuning dataset

Covers a wide range of capabilities, including image understanding, common-sense knowledge, cognitive reasoning, and math. It contains 2.1 million question-answer pairs from various sources and diverse real-world images.
Read 11 tweets
Apr 11
TimeGPT is the first foundation model specifically designed for time series analysis.

It excels at generating precise forecasts across a diverse range of datasets and domains.

Here's what you need to know about it:

1/8 Image
The model leverages a Transformer-based architecture, optimized for time series data, with self-attention mechanisms that facilitate the handling of temporal dependencies and patterns across varied frequencies and characteristics.

2/8
It incorporates an encoder-decoder structure, local positional encoding, and a linear output layer designed to map decoder outputs to forecast dimensions.

3/8
Read 10 tweets
Mar 3
8 Free Courses to Master Large Language Models:

1. @cohere LLM University
2. @huggingface NLP course
3. @databricks courses
and more!

🧵


Image
Image
Image
Image
1. @cohere LLM University

The course offers insights into how LLMs work, and their practical applications, and guides participants on using LLMs to build and deploy applications.docs.cohere.com/docs/llmu
2. @huggingface NLP course

This course provides comprehensive knowledge about Hugging Face Transformers, Datasets, Tokenizers, and the Accelerate tool in the field of Natural Language Processing (NLP).huggingface.co/learn/nlp-cour…
Read 9 tweets
Feb 19
DoRA (Weight-Decomposed Low-Rank Adaptation) sets a new standard for optimizing AI models.

It combines the benefits of full model fine-tuning and LoRA.

How does it do that? Let's see 👇🏼

1/7 Image
The genius of DoRA lies in its unique handling of pre-trained weights.

It separates these weights into two parts:

1. one that determines the size (magnitude)
2. one that determines the orientation (direction) of the weight vectors

2/7 Image
This allows for more nuanced adjustments during fine-tuning, closely mimicking the comprehensive learning achieved through full model fine-tuning but with far fewer parameters needing adjustment.

3/7
Read 7 tweets
Dec 27, 2023
Want to understand foundation models, generative AI models, and transformers?

Here is your FREE list of 15+ resources to do that:
1. Efficient Transformers: A Survey explores the evolution of Transformer models in various domains. It provides a comprehensive overview of different Transformer variants (X-formers) to guide researchers.
arxiv.org/abs/2009.06732
2. A Survey of Transformers

The survey provides a comprehensive review of Transformer variantsin AI fields. It introduces the vanilla Transformer and a new taxonomy of X-formers, covering architectural modifications, pre-training, and applications.
arxiv.org/abs/2106.04554
Read 18 tweets

Did Thread Reader help you today?

Support us! We are indie developers!


This site is made by just two indie developers on a laptop doing marketing, support and development! Read more about the story.

Become a Premium Member ($3/month or $30/year) and get exclusive features!

Become Premium

Don't want to be a Premium member but still want to support us?

Make a small donation by buying us coffee ($5) or help with server cost ($10)

Donate via Paypal

Or Donate anonymously using crypto!

Ethereum

0xfe58350B80634f60Fa6Dc149a72b4DFbc17D341E copy

Bitcoin

3ATGMxNzCUFzxpMCHL5sWSt4DVtS8UqXpi copy

Thank you for your support!

Follow Us!

:(