Javascript is not enabled. This site can still works but it'll be more interactive when javascript is enabled.
loading...
Hacker News
new
|
past
|
comments
|
ask
|
show
|
jobs
|
submit
bigyabai
18 hours ago
|
on: LLM Architecture Gallery
If your definition of "competitive" is loose enough, you can write your own Markov chain in an evening. Transformer models rely on a lot of prior art that has to be learned incrementally.
reply
jasonjmcghee
18 hours ago
|
parent
Not that loose lol.
I’m thinking it’s still llama / dense decoder only transformer.
reply