Why can't LLMs just be like a normal developer already?
So I wanted to get a clear picture about the Great Consensus Cleanup, a Bitcoin soft fork proposal that's supposed to be as uncontroversial as it can get while fixing vulnerabilities.
The amount of tokens and time I threw at this problem is unreal!
I thought I could one-shot it with @PayPerQ 's autoclaw via openclaw. Something like "Find the proposed activation client implementation for BIP 54, modify it to measure instead of enforce the rules and log your findings. I want to know how common were violations or almost-violations historically and especially during the past year. Oh and by the way, you only have 60GB, so you will have to write down your findings as the blocks are downloaded as you will have to prune."
Countless times did the LLM try to guess the results to save time, probe only some blocks from blockchain info or do whatever strange and unintuitive things. Countless times did it throw away perfectly fine IBD data. My concession now: Get 60GB of pruned data and then look what we find. We will have some data from the "stress test" times and isolated proof of concept transactions but not the full historic picture.
So I wanted to get a clear picture about the Great Consensus Cleanup, a Bitcoin soft fork proposal that's supposed to be as uncontroversial as it can get while fixing vulnerabilities.
The amount of tokens and time I threw at this problem is unreal!
I thought I could one-shot it with @PayPerQ 's autoclaw via openclaw. Something like "Find the proposed activation client implementation for BIP 54, modify it to measure instead of enforce the rules and log your findings. I want to know how common were violations or almost-violations historically and especially during the past year. Oh and by the way, you only have 60GB, so you will have to write down your findings as the blocks are downloaded as you will have to prune."
Countless times did the LLM try to guess the results to save time, probe only some blocks from blockchain info or do whatever strange and unintuitive things. Countless times did it throw away perfectly fine IBD data. My concession now: Get 60GB of pruned data and then look what we find. We will have some data from the "stress test" times and isolated proof of concept transactions but not the full historic picture.
On Jumble.social I see no indication of chat being supported at all?
Is nostr getting worse with chat?
#askNostr