I think Claude 3 crossed (or closely approached) an interesting threshold:
The "power users" threshold.
For the first time being able to help power users with heavy complex tasks faster than themselves.
This is a controversial topic in AI,
Let me try to explain:
I never use GPT-4 for code.
I use it for: 1. Brainstorming ideas. 2. Learning new topics that I don't know about. 3. Reading long texts instead of me (ask-your-pdf) 4. Just easy tasks.
But never for code.
It never helped me coding.
Jan 24 โข 11 tweets โข 3 min read
Gazans going crazy when realizing there are Muslims in the IDF
A Thread ๐งต๐
"Are these Muslims like us or what is their religion exactly?"
Oct 22, 2023 โข 7 tweets โข 2 min read
First week in the army.
Some thoughts:
The IDF is unbelievably different now than what is was during my (5yr) service.
Unrecognisable.
No clutter.
No bloat.
No BS procedures.
Everything (and everyone) is insanely efficient and highly motivated.
Young soldiers are not complaining and oftentimes also overdo their roles voluntarily.
More reservists keeps showing up.
(Many of which volunteered and were not supposed to attend otherwise)
Civilians show up every couple of hours with food, clothing and supplies.
Jul 16, 2023 โข 8 tweets โข 4 min read
The strongest model you didn't know about:
- The best model.
- Trained on the best dataset.
OpenChat V2 x OpenOrca
OpenChat V2, A leading open-source model was fine-tuned last night on Open-Orca's data.
Last night I participated in an interesting twitter space: Reverse Engineering the code-interpreter
Here are some of the most interesting bits: https://t.co/pUudeV5uuXtwitter.com/i/web/status/1โฆ
Since the new model is able to run code, many people have already managed to ask it nicely for all the information it has inside the virtual machine it runs on.
Jun 21, 2023 โข 8 tweets โข 2 min read
I think I get it now. let's try something out:
.
Comment to this thread everything you "just guess" about GPT-4 ๐๐
Guess: In the dataset: Went through every undergrad major and included exams, tests & textbooks of the subject. [to create a "wow" effect for every educated person no matter the field]
Jun 21, 2023 โข 4 tweets โข 1 min read
Ladies and Gentlemen. GPT-4 ๐
*Ported from HF's code
Jun 20, 2023 โข 4 tweets โข 2 min read
B R E A K I N G! @realGeorgeHotz Just said:
- GPT-4 Has 220 parameters.
- GPT-4 Is a mixture of experts with 8 experts.
- GPT-4 Is doing 16 times inference (did u mean beams? or just 2 beams per model?)
- @realGeorgeHotz It's HUGE!! Can you confirm??
src: latent.space/p/geohottwitter.com/i/web/status/1โฆ@abacaj look what you have done! I had work to do!
How to trick AUC into scoring your stupid models higher ๐
๐A - Thread - about gaming the AUC metric!๐
There is a built-in "flaw" with the AUC metric. Today we are going to take it for a spin!
One of the questions coming up as we get stuck on improving individual models further is how to best combine multiple models.
Apart from simple the baseline Avg, Median or Avg-Median-Rocket-Science-Blending[1].