Yam Peleg Profile picture
๐Ÿ‡ฎ๐Ÿ‡ฑ | AI & War it is
Felipe Meres Profile picture Jerome Ku Profile picture Santiago Cavanna Profile picture Ken Nickerson Profile picture Jason Profile picture 13 subscribed
Mar 10 โ€ข 10 tweets โ€ข 2 min read
I think Claude 3 crossed (or closely approached) an interesting threshold:

The "power users" threshold.

For the first time being able to help power users with heavy complex tasks faster than themselves.

This is a controversial topic in AI,
Let me try to explain: I never use GPT-4 for code.

I use it for:
1. Brainstorming ideas.
2. Learning new topics that I don't know about.
3. Reading long texts instead of me (ask-your-pdf)
4. Just easy tasks.

But never for code.
It never helped me coding.
Jan 24 โ€ข 11 tweets โ€ข 3 min read
Gazans going crazy when realizing there are Muslims in the IDF

A Thread ๐Ÿงต๐Ÿ‘‡ Image "Are these Muslims like us or what is their religion exactly?"Image
Oct 22, 2023 โ€ข 7 tweets โ€ข 2 min read
First week in the army.

Some thoughts:

The IDF is unbelievably different now than what is was during my (5yr) service.

Unrecognisable.

No clutter.

No bloat.

No BS procedures.

Everything (and everyone) is insanely efficient and highly motivated. Young soldiers are not complaining and oftentimes also overdo their roles voluntarily.

More reservists keeps showing up.

(Many of which volunteered and were not supposed to attend otherwise)

Civilians show up every couple of hours with food, clothing and supplies.
Jul 16, 2023 โ€ข 8 tweets โ€ข 4 min read
The strongest model you didn't know about:

- The best model.
- Trained on the best dataset.

OpenChat V2 x OpenOrca

OpenChat V2, A leading open-source model was fine-tuned last night on Open-Orca's data.

Link: https://t.co/4hiMTqXUOJ https://t.co/jG9Abqeos0huggingface.co/openchat/opencโ€ฆ
twitter.com/i/web/status/1โ€ฆ
Image OpenChat?

A series of powerful open source models leading multiple leader boards.

The teams' latest model, OpenChat V2 (weighted), Is currently leading 4 different benchmarks.

Github:
Model: https://t.co/Fv0793EEQSgithub.com/imoneoi/openchโ€ฆ
huggingface.co/openchat/opencโ€ฆ
Jul 9, 2023 โ€ข 8 tweets โ€ข 2 min read
Reverse engineering Code-Interpreter

Last night I participated in an interesting twitter space: Reverse Engineering the code-interpreter

Here are some of the most interesting bits: https://t.co/pUudeV5uuXtwitter.com/i/web/status/1โ€ฆ
Since the new model is able to run code, many people have already managed to ask it nicely for all the information it has inside the virtual machine it runs on.
Jun 21, 2023 โ€ข 8 tweets โ€ข 2 min read
I think I get it now. let's try something out:
.
Comment to this thread everything you "just guess" about GPT-4 ๐Ÿ‘‡๐Ÿ‘‡ Guess: In the dataset: Went through every undergrad major and included exams, tests & textbooks of the subject. [to create a "wow" effect for every educated person no matter the field]
Jun 21, 2023 โ€ข 4 tweets โ€ข 1 min read
Ladies and Gentlemen. GPT-4 ๐Ÿ‘‡ Image *Ported from HF's code
Jun 20, 2023 โ€ข 4 tweets โ€ข 2 min read
B R E A K I N G! @realGeorgeHotz Just said:
- GPT-4 Has 220 parameters.
- GPT-4 Is a mixture of experts with 8 experts.
- GPT-4 Is doing 16 times inference (did u mean beams? or just 2 beams per model?)
-
@realGeorgeHotz It's HUGE!! Can you confirm??
src: latent.space/p/geohot twitter.com/i/web/status/1โ€ฆ @abacaj look what you have done! I had work to do! Image
Jun 20, 2023 โ€ข 7 tweets โ€ข 3 min read
ื”ื—ืœืงื™ื ื”ื—ืกืจื™ื ืฉืœ GPT-4 (ื‘ืขื‘ืจื™ืช) ๐Ÿ‘‡
.
The missing pieces of GPT-4 (Hebrew)๐Ÿ‘‡
.
โ€‹ื˜ืจื™ืงื™ื ืฉืืฃ ืื—ื“ ืœื ืžื“ื‘ืจ ืขืœื™ื”ื ื•ืขืชื™ื“ ืžื•ื“ืœื™ ื”ืฉืคื” ื”ืคืชื•ื—ื™ื.
***
> ืืž;ืœืง: ืœืžื™ื“ื” ืžืžืฉื•ื‘ ืœื-ืื ื•ืฉื™ (!!)
.
ื”ืขื‘ืจืชื™ ื”ืจืฆืื” ืžืขื ื™ื™ื ืช ืœืคื ื™ ื›ืžื” ื™ืžื™ื ื‘ื›ื ืก #GenML ืขื ืื•ืชื• ื”ืฉื ("ื”ื—ืœืงื™ื ื”ื—ืกืจื™ื ืฉืœ GPT-4").
ื‘ื–ืžืŸ ืฉื›ื•ืœื ื• ืžื—ื›ื™ืโ€ฆ twitter.com/i/web/status/1โ€ฆ ื—ืœืง 2: ื”ื”ืฉืคืขื” ืฉืœ LLaMA
----------
ืืชื ื›ื‘ืจ ื™ื•ื“ืขื™ื ืขืœ LLaMA.
ืžื•ื“ืœ ืฉืคื” ืขื•ืฆืžืชื™ ืžื‘ื™ืช ืžื˜ื, ืฉื•ื—ืจืจ ื‘ืงื•ื“ ืคืชื•ื— ื‘ืคื‘ืจื•ืืจ ื”ืื—ืจื•ืŸ.
ื”ืžื•ื“ืœ ืขื•ืžื“ ื‘ืžืจื›ื–ื• ืฉืœ ืžืืžืฅ ืขื•ืœืžื™ ืœืฉื—ื–ืจ ืืช ื™ื›ื•ืœื•ืช ืžื•ื“ืœื™ ื”ืฉืคื” ื”ืžืกื—ืจื™ื™ื ื‘ืื•ืคืŸ ืคืชื•ื—.
ื‘ื›ืš ืœืกืคืง ืœืžื™ืœื™ื•ื ื™ ืื ืฉื™ื ื‘ื›ืœ ืจื—ื‘ื™ ื”ืขื•ืœื ื’ื™ืฉื” ืœื˜ื›ื ื•ืœื•ื’ื™ื” ื–ื•.
ื™ืฉื ื ืขืฉืจื•ืช ืืœืคื™ ืื ืฉื™ืโ€ฆ twitter.com/i/web/status/1โ€ฆ
May 2, 2022 โ€ข 24 tweets โ€ข 5 min read
How to trick AUC into scoring your stupid models higher ๐Ÿ‘Œ

๐Ÿ‘‡A - Thread - about gaming the AUC metric!๐Ÿ‘‡

There is a built-in "flaw" with the AUC metric. Today we are going to take it for a spin! One of the questions coming up as we get stuck on improving individual models further is how to best combine multiple models.

Apart from simple the baseline Avg, Median or Avg-Median-Rocket-Science-Blending[1].