| ▲ | eth0up 8 hours ago | |||||||||||||
I see a need for something similar for Perplexity. Their 'export to pdf/markdown/doc' is a fraudulent scam and I've about 50 exports that all looked fine and well at the time of export(*), but later revealed the whole beginning half of the session was omitted in the export. Or worse. I've lost many days of work because of this. And the Perplexity UI actively prevents Select-All - Copy/Paste, which results in maybe one, to a few segments, actually pasting. There is no direct method of exporting or saving a long session. Test it. And trying F12 Network etc, etc, reveals even XHR is a dead-end. Effectively. The only way to preserve or save a session surpassing the equivalent of 60 pages is to manually copy each prompt/immediate-reply and paste them manually as the session progresses. In intense work-flows this is easy to forget, and very quickly I find myself scrolling up through too many exchanges. Severely distracting. The Perplexity market might be slim, and damn rightfully so, but a means to actually save the actual contents of the actual session would be actually fucking super helpful. Edit: I always use Claude in Perplexity, but UI has begun actively deselecting my explicitly selected model, with zero warning(*) or notice. I have video of this. I select Claude, query the model for ID, and it says GPT5.1 as I watch my selected model disappear in real time. | ||||||||||||||
| ▲ | meatmanek 6 hours ago | parent [-] | |||||||||||||
For the truncated session logs issue, it sounds like it's auto-compacting the context (or just truncating it), because LLMs can only handle a finite amount of context. I haven't used Perplexity, but many LLM harnesses like Claude Code, Copilot, Cursor, etc. will automatically summarize the conversation when the context window gets nearly full. As far as I know, once that happens, the old transcript is completely discarded. (I could be wrong though.) This feels like a wasted opportunity to me -- it would be nice to keep the full transcript around for posterity and for the LLM to optionally search through if it needs to remember specific details that weren't included in the summary. | ||||||||||||||
| ||||||||||||||