Hacker News
new
|
past
|
comments
|
ask
|
show
|
jobs
|
submit
login
dakolli
8 days ago
|
parent
|
context
|
favorite
| on:
The path to ubiquitous AI (17k tokens/sec)
try here, I hate llms but this is crazy fast.
https://chatjimmy.ai/
help
bmacho
8 days ago
|
next
[–]
"447 / 6144 tokens" "Generated in 0.026s • 15,718 tok/s"
This is crazy fast. I always predicted this speed in ~2 years in the future, but it's here, now.
reply
Lalabadie
8 days ago
|
prev
|
next
[–]
The full answer pops in milliseconds, it's impressive and feels like a completely different technology just by foregoing the need to stream the output.
reply
FergusArgyll
8 days ago
|
prev
|
next
[–]
Because most models today generate slowish, they give the impression of someone typing on the other end. This is just <enter> -> wall of text. Wild
reply
machiaweliczny
8 days ago
|
prev
|
next
[–]
We need that for this chinese 3B model that think 45s for hello world but also solves math.
reply
Bolwin
7 days ago
|
parent
|
next
[–]
Nanbeige. Yeah this seems ideal for models that scale test time compute
reply
Serenacula
7 days ago
|
prev
[–]
Do we know anything about the method?
reply
Guidelines
|
FAQ
|
Lists
|
API
|
Security
|
Legal
|
Apply to YC
|
Contact
Search: