How to get URL link on X (Twitter) App
https://twitter.com/keyonV/status/1943730486280331460See:
https://x.com/ChombaBupe/status/1759226186075390033?s=19
https://twitter.com/ashleevance/status/1906052687084531897There is a simple explanation.
https://twitter.com/DaveShapi/status/1886375604531974479The base line should be using a pure lookup function over the actual LLMs training set & comparing the performance to the LLM, then seeing how much the LLM beats a search over the training set.
https://twitter.com/VictorTaelin/status/1881858899306745925For example, a forward pass through P(), usually a decoder-only transformer model, the output is an R, an array of numbers the size of the token vocabulary.
https://twitter.com/denny_zhou/status/1835761801453306089The problem of learnability - ability to learn, is the major bottleneck in machine intelligence.
https://twitter.com/ChombaBupe/status/1857125974015439183
https://twitter.com/BlueBir75555922/status/1857156409323860250
https://twitter.com/hardmaru/status/1856912013210832918
https://twitter.com/seanonolennon/status/1787712232782274628In deep learning the function f() is a composition of many smaller functions g1(),g2(),...,gL(), arranged in layers such that one function feeds from the ones before & feeds into the others after it.
https://twitter.com/mgubrud/status/1782155170647806296This is an example of a bag-of-tokens approach where only individual tokens in a set directly votes for the tokens they have high affinity for.
https://twitter.com/spiantado/status/1775217414613131714That is, given infinite data & compute all algorithms spiral down to merely looking up the corresponding response.
https://twitter.com/chrmanning/status/1772642891761955139Without task specific finetuning from human feedback the model in raw form underperform even though trained on extremely large scale data.
https://twitter.com/ChinpoKoumori/status/1763798803860406541
https://twitter.com/OpenAI/status/1758192957386342435The important question is what did they train this thing on?
https://twitter.com/emollick/status/1743023738428256540If you asked around - some folks working in computer vision think the machines good performance on toy benchmarks like ImageNet imply that these systems beat human visual capabilities when that's very far from the truth.