I like the idea here, but the final product is just so far from what good interactive articles/explanations actually look like. E.g., this style of article:
One can smell Claude's touch with these reactive teaching material. Not quite unexpected, every sane teacher uses Claude's artefacts to teach, but not all it spits is useful for convening knowledge.
jbdamask
Thanks to everyone who tried this today and those who provided feedback. I really appreciate your time. Here are some stats:
100 papers processed.
Cost breakdown:
LLM cost $64
AWS cost $0.0003
Claude's editorial comment about this breakdown, "For context, the Anthropic API cost ($63.32) is roughly 200,000x the AWS infrastructure cost. The AWS bill is a rounding error compared to the LLM spend."
Category breakdown:
Computer and Information Sciences 41%
Biological and Biomedical Sciences 15%
Health Sciences 7%
Mathematics and Statistics 5%
Geosciences, Atmospheric, and Ocean Sciences 5%
Physical Sciences 5%
Other 22%
There were a handful of errors due to papers >100 pages. If there were others, I didn't see them (but please let me know).
I'd be interested in hearing from people, what's one thing you would change/add/remove from this app?
japoneris
Well, i do not understand the concept.
Maybe i am too used to read paper: read the abstract to get a digest of the results, read the intro to understand the problem, skip all the rest as it is too technical or only for benchmark.
In the app, i selected a few paper, as i did not know anything about the selecter paper, comparing frog A doing magic stuff is helpless. Yet, the interface is great, i think this can be improve for true understanding.
show comments
jbdamask
Someone processed a paper on designing kindergartens. Mad props for trying such a cool paper. Really interesting how the LLM designed a soothing color scheme and even included a quiz at the end.
Man. I know you just made this for your own convenience, and all the big LLMs can one-shot this, but if you found a way to improve on the bog-standard LLM "webpage" design (inject some real human taste, experience and design sensibility), you'd get a few bucks from me- per paper.
show comments
vunderba
Nice job. I have no point of comparison (having never actually used it) - but wasn't this one of the use-cases for Google's NotebookLM as well?
Feedback:
Many times when I'm reading a paper on arxiv - I find myself needing to download the sourced papers cited in the original. Factoring in the cost/time needed to do this kind of deep dive, it might be worth having a "Deep Research" button that tries to pull in the related sources and integrate them into the webpage as well.
show comments
hackernewds
"daily limit reached" on first attempt :/
show comments
throwaway140126
A light mode would be great. I know that many people ask for a dark mode for the reason that they think that a light mode is more tiring than a dark mode but for me it is the opposite.
The actual explanation (using code blocks) is almost impossible to read and comprehend.
show comments
lamename
I tried to upload a 239 KB pdf and it said "Daily processing limit reached".
show comments
swaminarayan
How do you evaluate whether users actually understand better, rather than just feel like they do?
show comments
ajkjk
cool idea
probably need to have better pre-loaded examples, and divided up more granularly into subfields. e.g. "Physical sciences" vs "physics", "mathematics and statistics" vs "mathematics". I couldn't find anything remotely related to my own interests to test it on. maybe it's just being populated by people using it, though? in which case, I'll check back later.
show comments
filldorns
Great solution!
but...
Error
Daily processing limit reached. Please try again tomorrow.
show comments
armedgorilla
Thanks John. Neat to see you on the HN front page.
One LLM feature I've been trying to teach Alltrna is scraping out data from supplemental tables (or the figures themselves) and regraphing them to see if we come to the same conclusions as the authors.
LLMs can be overly credulous with the authors' claims, but finding the real data and analysis methods is too time consuming. Perhaps Claude with the right connectors can shorten that.
show comments
fsflyer
Some ideas for seeing more examples:
1. Add a donate button. Some folks probably just want to see more examples (or an example in their field, but don't have a specific paper in mind.)
2. Have a way to nominate papers to be examples. You could do this in the HN thread without any product changes. This could give good coverage of different fields and uncover weaknesses in the product.
show comments
jbdamask
Lots of great responses. Thank you!
I increased today's limit to 100 papers so more people can try it out
jbdamask
I see a few people trying to process big papers. Not sure if you're seeing a meaningful error in the UI but the response from the LLM is,
"A maximum of 100 PDF pages may be provided"
cdiamand
Great work OP.
This is super helpful for visual learners and for starting to onboard one's mind into a new domain.
Excited to see where you take this.
Might be interesting to have options for converting Wikipedia pages or topic searches down the line.
show comments
leke
Do you happen to know if LLMs have issues reading PDFs? Would they prefer EPUB format for example?
show comments
BDGC
This is neat! As an academic, this is definitely something I can see using to share my work with friends and family, or showing on my lab website for each paper. Can’t wait to try it out.
show comments
DrammBA
> I could just as well use a saved prompt in Claude
On that note, do you mind sharing the prompt? I want to see how good something like GLM or Kimi does just by pure prompting on OpenCode.
show comments
TheBog
Looks super cool, adding to the sentiment that I would happily pay a bit for it.
show comments
toddmorey
I’m worried that opportunities like this to build fun/interesting software over models are evaporating.
A service just like this maybe 3 years ago would have been the coolest and most helpful thing I discovered.
But when the same 2 foundation models do the heavy lifting, I struggle to figure out what value the rest of us in the wider ecosystem can add.
I’m doing exactly this by feeding the papers to the LLMs directly. And you’re right the results are amazing.
But more and more what I see on HN feels like “let me google that for you”. I’m sorry to be so negative!
I actually expected a world where a lot of specialized and fine-tuned models would bloom. Where someone with a passion for a certain domain could make a living in AI development, but it seems like the logical endd game in tech is just absurd concentration.
show comments
onion2k
I want this for my company's documentation.
show comments
alwinaugustin
There is a limit for 100 pages. Tried to upload the Architectural Styles and the Design of Network-based Software Architectures (REST - Roy T. Fielding) but it is 180 pages.
show comments
sean_pedersen
very cool! would be useful if headings where linkable using anchor
show comments
Vaslo
I’d love if this can be self-hosted, but i understand you may want to monetize it. I’ll keep checking back.
show comments
croes
Are documents hashed and the results cached?
show comments
jbdamask
The app may be getting throttled. If you're waiting on a job, check back in a bit.
enos_feedler
can i spin this up myself? is the code anywhere? thanks!
show comments
relaxing
I picked the “Attention is All You Need” example at the top, and wow it is not great!
Didn’t take long to find hallucination/general lack of intelligence:
> For each word, we compute three vectors: a Query (what am I looking for?), a Key (what do I contain?), and a Value (what do I give out?).
What? That’s the worst description of a key-value relationship I’ve ever read, unhelpful for understanding what the equation is doing, and just wrong.
> Attention(Q, K, V) = softmax( Q·Kᵀ / √dk ) · V
> 3 Mask (Optional) Block future positions in decoder
Not present in this equation, also not a great description of masking in a RNN.
I like the idea here, but the final product is just so far from what good interactive articles/explanations actually look like. E.g., this style of article:
- https://mlu-explain.github.io/decision-tree/
- any article from distill.pub
- any piece from NYT
One can smell Claude's touch with these reactive teaching material. Not quite unexpected, every sane teacher uses Claude's artefacts to teach, but not all it spits is useful for convening knowledge.
Thanks to everyone who tried this today and those who provided feedback. I really appreciate your time. Here are some stats:
100 papers processed.
Cost breakdown:
LLM cost $64
AWS cost $0.0003
Claude's editorial comment about this breakdown, "For context, the Anthropic API cost ($63.32) is roughly 200,000x the AWS infrastructure cost. The AWS bill is a rounding error compared to the LLM spend."
Category breakdown:
Computer and Information Sciences 41%
Biological and Biomedical Sciences 15%
Health Sciences 7%
Mathematics and Statistics 5%
Geosciences, Atmospheric, and Ocean Sciences 5%
Physical Sciences 5%
Other 22%
There were a handful of errors due to papers >100 pages. If there were others, I didn't see them (but please let me know).
I'd be interested in hearing from people, what's one thing you would change/add/remove from this app?
Well, i do not understand the concept. Maybe i am too used to read paper: read the abstract to get a digest of the results, read the intro to understand the problem, skip all the rest as it is too technical or only for benchmark. In the app, i selected a few paper, as i did not know anything about the selecter paper, comparing frog A doing magic stuff is helpless. Yet, the interface is great, i think this can be improve for true understanding.
Someone processed a paper on designing kindergartens. Mad props for trying such a cool paper. Really interesting how the LLM designed a soothing color scheme and even included a quiz at the end.
https://nowigetit.us/pages/9c19549e-9983-47ae-891f-dd63abd51...
Man. I know you just made this for your own convenience, and all the big LLMs can one-shot this, but if you found a way to improve on the bog-standard LLM "webpage" design (inject some real human taste, experience and design sensibility), you'd get a few bucks from me- per paper.
Nice job. I have no point of comparison (having never actually used it) - but wasn't this one of the use-cases for Google's NotebookLM as well?
Feedback:
Many times when I'm reading a paper on arxiv - I find myself needing to download the sourced papers cited in the original. Factoring in the cost/time needed to do this kind of deep dive, it might be worth having a "Deep Research" button that tries to pull in the related sources and integrate them into the webpage as well.
"daily limit reached" on first attempt :/
A light mode would be great. I know that many people ask for a dark mode for the reason that they think that a light mode is more tiring than a dark mode but for me it is the opposite.
Neat!
Social previews would be great to add
https://socialsharepreview.com/?url=https://nowigetit.us/pag...
Neat! I've previously used something similar: https://www.emergentmind.com/
https://nowigetit.us/pages/8cf08b76-c5bc-4a7b-bdb4-a0c15089e...
The actual explanation (using code blocks) is almost impossible to read and comprehend.
I tried to upload a 239 KB pdf and it said "Daily processing limit reached".
How do you evaluate whether users actually understand better, rather than just feel like they do?
cool idea
probably need to have better pre-loaded examples, and divided up more granularly into subfields. e.g. "Physical sciences" vs "physics", "mathematics and statistics" vs "mathematics". I couldn't find anything remotely related to my own interests to test it on. maybe it's just being populated by people using it, though? in which case, I'll check back later.
Great solution!
but...
Error Daily processing limit reached. Please try again tomorrow.
Thanks John. Neat to see you on the HN front page.
One LLM feature I've been trying to teach Alltrna is scraping out data from supplemental tables (or the figures themselves) and regraphing them to see if we come to the same conclusions as the authors.
LLMs can be overly credulous with the authors' claims, but finding the real data and analysis methods is too time consuming. Perhaps Claude with the right connectors can shorten that.
Some ideas for seeing more examples:
1. Add a donate button. Some folks probably just want to see more examples (or an example in their field, but don't have a specific paper in mind.)
2. Have a way to nominate papers to be examples. You could do this in the HN thread without any product changes. This could give good coverage of different fields and uncover weaknesses in the product.
Lots of great responses. Thank you!
I increased today's limit to 100 papers so more people can try it out
I see a few people trying to process big papers. Not sure if you're seeing a meaningful error in the UI but the response from the LLM is, "A maximum of 100 PDF pages may be provided"
Great work OP.
This is super helpful for visual learners and for starting to onboard one's mind into a new domain.
Excited to see where you take this.
Might be interesting to have options for converting Wikipedia pages or topic searches down the line.
Do you happen to know if LLMs have issues reading PDFs? Would they prefer EPUB format for example?
This is neat! As an academic, this is definitely something I can see using to share my work with friends and family, or showing on my lab website for each paper. Can’t wait to try it out.
> I could just as well use a saved prompt in Claude
On that note, do you mind sharing the prompt? I want to see how good something like GLM or Kimi does just by pure prompting on OpenCode.
Looks super cool, adding to the sentiment that I would happily pay a bit for it.
I’m worried that opportunities like this to build fun/interesting software over models are evaporating.
A service just like this maybe 3 years ago would have been the coolest and most helpful thing I discovered.
But when the same 2 foundation models do the heavy lifting, I struggle to figure out what value the rest of us in the wider ecosystem can add.
I’m doing exactly this by feeding the papers to the LLMs directly. And you’re right the results are amazing.
But more and more what I see on HN feels like “let me google that for you”. I’m sorry to be so negative!
I actually expected a world where a lot of specialized and fine-tuned models would bloom. Where someone with a passion for a certain domain could make a living in AI development, but it seems like the logical endd game in tech is just absurd concentration.
I want this for my company's documentation.
There is a limit for 100 pages. Tried to upload the Architectural Styles and the Design of Network-based Software Architectures (REST - Roy T. Fielding) but it is 180 pages.
very cool! would be useful if headings where linkable using anchor
I’d love if this can be self-hosted, but i understand you may want to monetize it. I’ll keep checking back.
Are documents hashed and the results cached?
The app may be getting throttled. If you're waiting on a job, check back in a bit.
can i spin this up myself? is the code anywhere? thanks!
I picked the “Attention is All You Need” example at the top, and wow it is not great!
Didn’t take long to find hallucination/general lack of intelligence:
> For each word, we compute three vectors: a Query (what am I looking for?), a Key (what do I contain?), and a Value (what do I give out?).
What? That’s the worst description of a key-value relationship I’ve ever read, unhelpful for understanding what the equation is doing, and just wrong.
> Attention(Q, K, V) = softmax( Q·Kᵀ / √dk ) · V
> 3 Mask (Optional) Block future positions in decoder
Not present in this equation, also not a great description of masking in a RNN.
> 5 × V Weighted sum of values = output
Nope!
https://nowigetit.us/pages/f4795875-61bf-4c79-9fbe-164b32344...
People will do anything except work