Saved tweets
Just published: Do large language models understand us? https://t.co/Ks6zztTzxv Itโs sometimes claimed that ML is "just stats" and AI can't "understand". I'm arguing that LLMs have a great deal to teach us about language, understanding, intelligence, sociality, even personhood.
โ Blaise Aguera (@blaiseaguera) December 16, 2021
You use GPUs everyday, but do you (actually) know how they work?
โ Sasha Rush (@srush_nlp) July 12, 2022
GPU-Puzzles (v0.1) - 14 short puzzles in Python with a visual debugger. No background required. Do puzzles, learn CUDA.
Link: https://t.co/Yk1lWRqilN pic.twitter.com/eFs7u5lxES
When I started lifting I thought I was a hard gainer.
โ Warren English (@TheWarEnglish) August 23, 2022
Now people tell me I have good genetics.
If you struggle to gain muscle, read this: pic.twitter.com/gDSHgRdflm
There are more than 3,000 TED Talks.
โ Unleash Your Mind (@MentalUnleash) August 23, 2022
Here are 10 TED Talks that will change the way you think forever:
3. Public APIs
โ Pratham (@Prathkum) November 7, 2022
A collective list of free APIs for use in software and web development
๐ https://t.co/vDRQKBf15V pic.twitter.com/cGI11nbfeV
Here is a great explanation of the Lagrangian multiplier (the intuition of which is typically not given). ๐งต
โ Lionel Page (@page_eco) November 10, 2022
The problem: maximising a function f(x,y) under a constraint g(x,y)=k (the point has to be on the red curve). pic.twitter.com/xgLQ9UAKdp
These @alfcnz's lecture slides are *colourful*!https://t.co/E4qkoqzDj7 pic.twitter.com/HPdAjesSgz
โ milton (@tensor_fusion) November 12, 2022
I wrote a blog post on why I decided to join OpenAI instead of academia.
โ Rowan Zellers (@rown) February 12, 2023
(after I went on the academic & industry job markets, and got offers from both.)
This post (pt2 in a series) took a while ๐ - hoping my experience helps others make life decisions!https://t.co/B5z4DGP9yI
Asociados, necesito una forma de ver la f1 sin pagar un cรฉntimo a DAZN, confรญo en vosotros
โ - (@Nanoestafa) February 27, 2023
I finally find an explanation for why RL is needed for RLHF that satisfied me. It's actually like playing board games.
โ Zhengyao Jiang (@zhengyaojiang) February 28, 2023
The reward model can only judge a full answer and a "critic" is needed to efficiently improve the intermediate moves (earlier tokens in the answer) 1/4 pic.twitter.com/GMjSVC6Tee
If you sit for more than 6 hours a day, read this: pic.twitter.com/CYUr79mYF9
โ Dan Go (@FitFounder) February 28, 2023
I love Pandas! I've been using it ever since I started doing machine learning more than a decade ago.
โ Sebastian Raschka (@rasbt) March 4, 2023
Now that the Pandas 2.0 release candidate is out, I was just taking the new PyArrow backend for a test drive. It's a significant boost over the original https://t.co/vk8mtViPtJโฆ pic.twitter.com/wThAm26cdB
I packed-up a full-text paper scraper, vector database, and LLM into a CLI to answer questions from only highly-cited peer-reviewed papers. Feels unreal to be able instantly get answers by an LLM "reading" dozens of papers. 1/2 pic.twitter.com/a6PWxWyuc1
โ Andrew White ๐ฆโโฌ (@andrewwhite01) February 25, 2023
In my mid 20's to early 30's I dealt with chronic low back pain.
โ Dan Go (@FitFounder) March 11, 2023
I went to a Chiropractor & he told me I'd have to live with the pain & get adjustments for the rest of my life.
I said screw that, searched for a different solution & fixed my back.
Here's how I did it: pic.twitter.com/sy5DkpZZsp
Harsh truths I know at 32 I wish knew at 22:
โ Sahil Bloom (@SahilBloom) March 11, 2023
ML runs everywhere now because of 2 recent trends:
โ Mishig Davaadorj (@mishig25) March 17, 2023
* ๐๐๐๐๐๐๐๐๐๐-๐๐๐๐๐๐๐๐ of ML (transformers.js, huggingface.js) for running on browsers
* ๐๐๐-๐๐๐๐๐๐๐๐ of ML (llama.cpp, bloomz.cpp, whisper.cpp) for running on embedded devices
RIP IOS Developers ๐
โ peter! ๐ฅท (@pwang_szn) March 26, 2023
I just built a 95% functional IOS app in less than 2 hours with GPT-4. (With Payments and OpenAI integration)
I have zero Swift experience (I hired a freelancer to pair program thou.)
Gonna write a thread about this entire experience in 1-2 days. ๐คฏ pic.twitter.com/H1pffLfx1Y
The normalization scheme that DeepMind researchers came up with for their "linear recurrent unit" (LRU) is a nice example of how it is possible to predictably engineer circuits in artificial neural networks, when you know what you're doing. A thread: pic.twitter.com/AxzFE58OQk
โ Charles Foster (@CFGeek) March 27, 2023
If you are a PhD student, you should check out the book called "How to take smart notes".
โ Mahesh Sathiamoorthy (@madiator) April 1, 2023
I wrote a bit about this book and what I have learned about note-taking in https://t.co/ROFiqmoYC6. pic.twitter.com/vTwlC32xdL
โฌโฌ Evoluciรณn de mi patrimonio a 01/02/2023 โฌ โฌ
โ La Carrera Del Dinero (@carrera_dinero) April 1, 2023
Seguramente este sea el รบltimo mes por debajo de los 40k, y creo que para la edad que tengo estรก muy bastante bien :P
Hoy querรญa preguntaros una cosa...
ยฟEsta publicaciรณn mensual os aporta algo de valor? ยฟO no? pic.twitter.com/aXtw0lEX49
1/๐งต๐ Making sense of Principal Component Analysis (PCA), Eigenvectors & Eigenvalues: A simple guide to understanding PCA and its implementation in R! Follow this thread to learn more! #RStats #DataScience #PCA pic.twitter.com/An6qxZGLDP
โ Selรงuk Korkmaz (@selcukorkmaz) April 23, 2023
I was puzzled for a while as to why we need RL for LM training, rather than just using supervised instruct tuning. I now have a convincing argument, which is also reflected in a recent talk by @johnschulman2 . I summarize it in this post:https://t.co/DQD1wgyjg3
โ (((ู()(ู() 'yoav))))๐พ (@yoavgo) April 22, 2023
โTransformers from scratchโ by Brandon Rohrer ๐ค
โ Sanyam Bhutani (@bhutanisanyam1) April 24, 2023
This is one of the best write ups, that starts from 0 and explains every single detail of the model architecture.
Even if you need a refresher or donโt, I would still highly recommend reading it:https://t.co/D25bs6TP5X pic.twitter.com/IWOWP1P9QW
[1/9] ๐ฒ Let's talk about the difference between probability and likelihood in #statistics. These two terms are often confused, but understanding their distinction is key for making sense of data analysis! #Rstats #DataScience pic.twitter.com/Xuo19nyTvU
โ Selรงuk Korkmaz (@selcukorkmaz) April 23, 2023
This is *exactly* what I had in mind when disliking the term "emergent" recently.
โ Lucas Beyer (bl16) (@giffmana) May 1, 2023
It seems due to the metrics (like binary correct/incorrect), in reality the model does smoothly approach the right answer.
But I was too lazy to verify this intuition myself, glad this paper did! https://t.co/xLQBGPkwxs pic.twitter.com/ZPNaEtiYGM
๐๐ผ๐ป๐๐ผ๐น๐๐๐ถ๐ผ๐ป๐ ๐ณ๐ฟ๐ผ๐บ ๐ณ๐ถ๐ฟ๐๐ ๐ฝ๐ฟ๐ถ๐ป๐ฐ๐ถ๐ฝ๐น๐ฒ๐
โ Marc Lelarge ๐ป (@marc_lelarge) May 8, 2023
Linear+Shift Invariant = Convolution. A simple proof with circulant matrices ๐ปhttps://t.co/P5ifI6nPJh
๐learn how stacking convolutions with a kernel of size 3 get you a network with a receptive field of size 9๐ pic.twitter.com/swSCqdIusY
The team at @CohereAI just released an awesome API endpoint (called Rerank) that can easily improve search and recommendation offerings by using LLMs. Here's what you need to know...
โ Cameron R. Wolfe, Ph.D. (@cwolferesearch) May 9, 2023
Some background: Most search engines follow a two-step process.
1. Filtering: a rough/efficientโฆ pic.twitter.com/C15AVexFPq
Interesting paper from my ex-colleagues at @GoogleAI led by @vqctran. Generative retrieval (i.e., DSI) is one of the most fun works I've worked on (and pioneered) during my Google career.
โ Yi Tay (@YiTayML) May 22, 2023
Also, @vqctran is driving a lot of the agenda that we worked on together back then. He hasโฆ https://t.co/FYyDng4n2e
Is Adam the best optimizer to train neural networks?๐ค
โ Frank Schneider (@frankstefansch1) June 13, 2023
We don't know. And we won't know until we test training algorithms properly.
๐That's why we spent ~2.5 years building AlgoPerf, a competitive, time-to-result training algorithms benchmark using realistic workloads! pic.twitter.com/Tzf4R4UXVO
shower thought : drop the position embeddings, rewrite the transformer using complex numbers, encode the position information in the complex phase
โ Georgi Gerganov (@ggerganov) June 23, 2023
ref : see how MRI phase encoding works
this paper's nuts. for sentence classification on out-of-domain datasets, all neural (Transformer or not) approaches lose to good old kNN on representations generated by.... gzip https://t.co/6eZiXlJxOX pic.twitter.com/sF9kd1FzI4
โ Luke Gessler (@LukeGessler) July 12, 2023
github repo w/ model: https://t.co/cRh08vsHFN
โ MF FOOM (@MF_FOOM) August 3, 2023
just a modded @karpathy nanoGPT with ada embeddings projected in as input tokens
heads up, the checkpoint on HF is only ~117M parameters, and it's finetuned on a tiny subset of Wikipedia sentences so it's very easy to go OOD
every now and then thereโs an insane alpha bomb on reddit pic.twitter.com/2Gw9Yuu83Y
โ varepsilon (@var_epsilon) August 18, 2023
what's your favorite codebase for training larger scale neural networks?
โ Aleksa Gordiฤ ๐ฟ๐ค (@gordic_aleksa) August 27, 2023
fairseq?
GPT-NeoX?
MosaicML's composer/llm-foundry?
something else?
worth taking a look at https://t.co/mFpvQ8MXfA for some intuition on these new knobs pic.twitter.com/m9jEqj9Dej
โ Susan Zhang (@suchenzang) August 31, 2023
Whatever inductive bias you can bake into your model... bake it in. https://t.co/Upxj521Bh6 pic.twitter.com/PRMraEJOQN
โ Susan Zhang (@suchenzang) August 31, 2023
ยฟQuieres empezar a invertir y no sabes por dรณnde empezar?
โ Pobre Millenial (@pobremillenial) September 9, 2023
En 4 vรญdeos te doy lo mรกs bรกsico para que puedas empezar a entender este mundillo. pic.twitter.com/Z6XDigvanK
I started my career in Data Science back in 2016 โณ
โ Akshay ๐ (@akshay_pachaar) September 9, 2023
Being self-taught, I came across several courses and books, but these are some of my favourites! ๐
1๏ธโฃ ISL
Tested by time and read by millions, the bible for statistical & classical machine learning.
Mathematical conceptsโฆ pic.twitter.com/83DnSnqgkx
What is Causal Inference?
โ ๐ฅKareem Carr ๐ฅ (@kareem_carr) September 11, 2023
Causal Inference is a new science of causation. This field is nothing less than a revolution in how scientists understand data. Read on to learn more.
This is the first post in a series based on the Book of Why by Judea Pearl. I will be reading theโฆ pic.twitter.com/IOI7vHthBc