Hacker Newsnew | past | comments | ask | show | jobs | submit | quinncom's commentslogin

Although I rarely hit my limit in my $20 a month Codex plan, I can imagine this would be very useful.

The issue I have more often is that I will start a conversation in ChatGPT, and realize an hour later that I needed all that context to be in Codex, so I’ll generally ask ChatGPT to give me a summary of all facts and a copy‑paste prompt for Codex. But maybe there is a way to extract the more useful content from a chat UI to an agent UI.


imo an agent learns way more by watching the raw agentic flow than by reading some sanitized context dump. you get to see exactly where the last bot derailed and then patched itself. give that a shot—handing over a spotless doc feels fake anyway.

Screen sharing to any remote API is a nonstarter for me. I don’t care if the API claims ZDR; Snowden’s revelations are still echoing. So, I appreciate that the app supports a custom endpoint for local models.

Which local models did you try? GLM-OCR seems like it would excel at this: https://huggingface.co/zai-org/GLM-OCR


I've got it installed with Qwen3-VL-4B running in LM Studio on my MBP M1 Pro. (Yes, the fans are running.) GLM-OCR didn't work because it returns all text on the screen, despite the instructions asking only for a summary.

Screenshots are summarized in ~28 seconds. Here's the last one:

> "The user switched to the Hacker News tab, displaying item 47049307 with a “Gave Claude photographic memory for $0.0002/screenshot” headline. The chat now shows “Sonnet 4.6” and a message asking “What have I been doing in the past 10 minutes?” profile, replacing prior Signal content. The satellite map background remains unchanged."

The satellite map background remains unchanged message appears in every summary (my desktop background is a random Google Maps satellite image that rotates every hour).

I would like to experiment with custom model instructions – for example, to ignore desktop background images.

Earlier in my testing it was sending screenshots for both of my displays at the same time, which was much slower, but now it's only sending screenshots of my main screen. Does MemoryLane only send screenshots for displays that have active windows?

Here's the first test of the MCP server in Claude – https://ss.strco.de/SCR-20260217-onbp.png – it works!


Update: I switched to Qwen3 VL 2B (`qwen3-vl-2b-instruct-mlx@bf16`) which is 2.5× faster than 4B (11s vs 18s per screenshot) and my meager M1 Pro is able to keep up without the fans spinning 100% of the time.

Small nudges to steer company culture regarding AI use:

- signal disclosure as a norm: whenever you use AI, say “BTW I used AI to write this”, when you don’t use AI, say “No AI used in this document”

- add an email footer to your messages that states you do not use AI because [shameful reasons]

- normalize anti-AI language (slop, clanker, hallucination, boiling oceans)

- celebrate human craftsmanship (highlight/compliment well written documentation, reports, memos)

- share AI-fail memes

- gift anti-AI/pro-human stickers

- share news/analysis articles about the AI productivity myth [0], AI-user burnout [1], reverse centaur [2], AI capitalism [3]

[0] https://hbr.org/2025/09/ai-generated-workslop-is-destroying-... [1] https://hbr.org/2026/02/ai-doesnt-reduce-work-it-intensifies... [2] https://pluralistic.net/2025/12/05/pop-that-bubble/ [3] https://80000hours.org/problem-profiles/extreme-power-concen...


OpenClaw is actually built on top of pi-mono (for its agent runtime, models, and tools):

https://docs.openclaw.ai/concepts/agent#pi-mono-integration

https://github.com/openclaw/openclaw/blob/main/docs/pi.md


Did you scroll through the pricing options? The largest Kimi plan is $199/month. “Much better” depends on how much usage is included vs. Anthropic plans/API costs.

From what I can see, this is agentic tooling that provides similar features to OpenClaw. It’s been on GitHub since June 2024 but never seemed to catch the hype train. Some stats comparing the popularity of the two:

  Agent Zero: 14k GH stars, 3k X followers
  OpenClaw: 197k GH stars, 314k X followers

Much ado about nothing, this link has been going around the Fediverse: https://www.resistandunsubscribe.com/

Login.gov is still in use by the states, though. I recently registered for CA’s DROP, and it used Login.gov.

When this happens to me, it's either because I'm connected to a VPN, or I'm using Cloudflare's public DNS server.

Exponential growth may look like a very slow increase at first, but it's still exponential growth.

Sigmoids may look like exponential growth at first, until they saturate. Early growth alone cannot distinguish between them.

Intelligence must be sigmoid of course, but it may not saturate until well past human intelligence.

Intelligence might be more like an optimization problem, fitting inputs to optimal outputs. Sometimes reality is simply too chaotic to model precisely so there is a limit to how good that optimization can be.

It would be like distance to the top of a mountain. Even if someone is 10x closer, they could still only be within arms reach.


On the other hand: Perception of change might not be linear but logarithmic.

(= it might take an order of magnitude of improvements to be perceived as a substantial upgrade)

So the perceived rate of change might be linear.

It's definitely true for some things such as wealth:

- $2000 is a lot of you have $1000.

- It's a substantial improvement of you have $10000.

- It's not a lot you have $1m

- It does not matter if you have $1b


$2000 is not substantial over $1b on the linear scale

2k is the same on the linear scale no matter where you are. that's what the linear scale is about.

you're already interpreting this on the log scale


If it's exponential growth. It may just as well be some slow growth and continue to be so.

Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: