• 0 Posts
  • 15 Comments
Joined 1 year ago
cake
Cake day: July 22nd, 2023

help-circle
  • Absolutely agree. I’m only talking about the fleeing the country part. Those of us who can stay are going to have to put in a lot of work, speaking out against fascism, protecting those who cannot flee, and being generally rebellious against tyranny.

    If anything those of us who happen to not be directly in their crosshairs have a greater responsibility to speak out for the groups that are going to be targeted, because it could quickly get to a point where it’s dangerous for those marginalized people to be as vocal. We cannot leave the most vulnerable to fight alone for their right to exist.


  • Technically it’s not a full 55% of my countrymen, just 55% of the ones who bothered to vote. I’ll admit that’s not really a meaningful distinction though. Unfortunately, there’s also more of us who want to leave than the rest of the world can reasonably handle. I hope as many marginalized people can get out, because it’s going to be bad, especially for them. But those, like me, who are unlikely to be directly targeted due simply to being lucky enough to be born straight, white, men should probably leave those limited seats for those who truly need to leave.


  • I still use DDG as my “daily driver” (I know there are better options for privacy and avoiding big tech, but I haven’t yet found anything independent that is good enough for me to switch to full time yet). I bookmarked Stract a while back, and it proved useful a few months back when Microsoft had an outage that took down Bing and by extension, Duck Duck Go. I do like Stract, their index seems to be enough larger than MoJeek (another independent search with their own index) that it gives me better results.

    Stract might not be as open as I’d like, but it’s nice to have as an option, and I’m never going to complain about having more search providers with independent indexes.



  • While I’m not the person you replied to and don’t know what their argument would be, I’ll take a shot at giving my own answer. In many cases when people post examples of AI giving unhelpful or bad information, there’s often someone who runs off to their favorite LLM to see if it gives a better result, and it usually does, so it gets treated like user error for using the wrong LLM or not wording the prompt properly. When in other examples that person’s favorite LLM which gave the correct answer this time, is the bad example hallucinating or mixing unrelated concepts, and other people are in the comments promoting other LLMs that gave them a good reply this time. None of the LLMs are actually trustworthy consistently enough to be trusted alone, and you won’t really know what answer is trustworthy unless you ask several LLMs and then go research the topic on your own anyway to figure out which answer is the most correct. It’s a valid point that ChatGPT got the answer more right than Gemini this time, but it’s somewhat useless to know that because other times ChatGPT is the one hallucinating wildly, and Gemini has the right answer, but since they’ve all been wrong before who do you trust.

    LLMs are like asking an arrogant person who thinks they know everything, who rather than admitting what they don’t know, will pull an answer out of their butt, and while it might be a logical answer, it isn’t based in reality, and may still be wildly wrong. If you already mostly know the answer, maybe asking the arrogant person works, because you already know enough to know if they are speaking from their actual knowledge or making up an answer, but if you don’t already have knowledge on a topic, you won’t know whether the arrogant person is giving useful information or not.