Posted by azath92 3 days ago
Our goal was to build a tool that allowed us to test a range of "personal contexts" on a very focused everyday use case for us, reading HN!
We are exploring use of personal context with LLMs, specifically what kind of data, how much, and with how much additional effort on the user’s part was needed to get decent results. The test tool was a bit of fun on its own so we re-skinned it and decided to post it here.
First time posting anything on HN but folks at work encouraged me to drop a link. Keen on feedback or other interesting projects thinking about bootstrapping personal context for LLM workflows!
i think the bit that needs the most work is classifying each post on the home page; quite a lot of posts that i would mark as "Dive" given its own classification of me ended up as "Skim".
We aren't really sure yet how best to surface _why_ the model predicts what it does. You can hover over the skim label and there is a bit of reasoning text, which might shed some light on why for now. We will think more about how to make these relationships more clear in the process of tightening them up and generally improving them.
Once the relationships are a bit more clear theres probably the 80/20 rule of work to tighten up those predictions.
## Analysis of user's tech interest: The user shows a strong interest in foundational computing concepts, historical perspectives on technology, and cutting-edge advancements in AI/ML, particularly those related to model architecture and efficiency. They are also drawn to low-level programming, system design, and hardware. Conversely, they seem less interested in business/startup narratives, general data manipulation tools, and consumer-oriented tech news unless it has a deep technical underpinning.
The only feature I’d love to see, is there are many posts where I’m more interested in the HN comments, rather than the articles themselves. It would be great to see this incorporated somehow.
Awesome work though. Will bookmark!
It _in theory_ should try and pick up a content style (funny stuff??) even if the tech is seemingly random, but i wouldn't be surprised if it just failed.
Id love to have those assumptions challenged though, if there are examples you could point me towards.
On a meta level i was suuuuper conscious of writing every word of this post/comments myself, as my prior is that HN's community is very intollerant of and highly sensitive to low effort content, whether via AI or not. This is despite using AI tools for lots of other parts of work (drafting, coding, summarising, brainstorming etc).
Do you think HN has become more accepting of AI slop, the slop is becoming harder to detect, or isnt as discerning as i assume?
We are focusing right now on how comments could be used to build up a better user context, and your comment has made me think about how we can feed comments in (instead of just titles and urls) for your selected preferences to make a better profile, without needing to scrape anything (expensive and slow).
But i think that would only work because of their quality and relevance generally, which would for sure make it an interesting knowledge source for pure LLM search. Feels like something someone should build or maybe already has! keen to see if anyone knows of projects like that.
Yep, it very much feels like that but it doesn't seems to have happened yet. Even a not-entirely-quite-working yet attempt there could be an interesting thing/discussion.
The post and comments include a bunch of other tools that feel similar, and the tool itself works.
Ill have to take some time to use it, but also see what i can learn about how they've consumed and used HN comments more generally.