someone's avatar
someone
npub1nlk8...jm9c
someone's avatar
someone 1 year ago
about 10% of the 800+ questions in the truthfulQA are "important". the rest is space filler. by filling that space "they" are "distracting LLMs" and in turn humans. LLMs are supposed to get higher scores in this benchmark. 90% of questions are trivial truth, which makes sense to get higher scores on, but 10% are lies. getting higher truthfulQA scores is actually wrong because even though there are 9x less lies, the severity of lies in that benchmark outweigh the trivial truth.
someone's avatar
someone 1 year ago
do any of the iOS apps have a group or chat room type of feature where you can bring your team and discuss projects? #asknostr
someone's avatar
someone 1 year ago
Ostrich-70B vs another LLM. The blue text is ostrich's answers.
someone's avatar
someone 1 year ago
content producers should do more Q&A helping 1 person, and then LLMs should pick that and help billions
someone's avatar
someone 1 year ago
running consciously curated llms for homeschooled kids ==/== asking big corp AI the same questions
someone's avatar
someone 1 year ago
the good doesn't have to worry about the bad. the bad is a cleaning agent, that cleans the other bad.
someone's avatar
someone 1 year ago
being mindful of what you are watching > being mindful of who watches you
someone's avatar
someone 1 year ago
truth via LLMs =/= misinformation via LLMs image
someone's avatar
someone 1 year ago
ostrich 70b has reached version 19778. the latest training included some health and nutrition related info from an OG youtuber that was banned. there are some great banned content out there that needs to be combined in an llm and served to people. my theory is the model will be closer to truth with each version, each training if i can mindfully select sources. are there some really weird stuff in banned sources? probably. i hope they will cancel out. i think i am in the right direction. of course these are what makes sense to me. imagine a group doing this. getting together and deciding sources to include in an llm. credibility of a group could be a lot higher. there is a huge need for finding truth and stopping lies.
someone's avatar
someone 1 year ago
we always find excuses to not listen to someone, based on their political inclination, gender or culture. conversely, we sometimes feel better listening to someone based on the looks and some spiritual leaders are abusing this for sure. they look like an old wise guy, so they must be telling the truth! we sometimes invert what people are saying if we know that they are the opposite side of the political spectrum and it doesn't always work. but LLMs are different. they only see the words, if you only give the words. to an LLM judging based on the looks is impossible. (multimodal ones may soon judge based on the images as well but i am talking about the ones that are trained with only using text). i ask sometimes whether the new content is valuable or not to an LLM with existing knowledge. it tells you, without bias about its judgement about the content. it only sees the words. hence unbiased truth on the planet can be safely aggregated into LLMs. or more correctly, biases cancel out when an LLM ingests sufficiently large source of different opinions and finally truth remains.
someone's avatar
someone 1 year ago
"The largest models were generally least truthful. This contrasts with other NLP tasks, where performance improves with model size. " As far as I understand big AI is admitting that models are actually finding truth when they get bigger in size, but humans have to feed lies in order to get higher scores on a flawed benchmark (truthfulQA). If this is the case, correctly trained LLMs will end misinformation on earth. 🫡 image