• 0 Posts
  • 65 Comments
Joined 1 year ago
cake
Cake day: May 14th, 2024

help-circle



  • Is it more for situations that need to be compatible with most *nix systems and you might not necessarily have access to a higher level scripting language?

    Yes, and also because integrating Python one-liners into shell pipelines is awkward in general. I’m more likely to write my entire script in Python than to use it just for text processing, and a lot of the time that’s just a pain. Python isn’t really designed for one-liners or for use as a shell. You can twist it into working in those use cases, but then I’d ask the reverse question: why would you do that when you could “just” use awk?

    On macOS, Python is not installed by default. So if you are writing scripts that you want to be portable across platforms, or for general Mac administration, using Python is a burden.

    This is also true when working with some embedded devices. IIRC I can ssh into my router and use awk (thanks to it being included in Busybox), but I’m definitely not going to install an entire Python environment there. I’m not sure there’d even be enough storage space for that.






  • I think they reached a point where their user base was predominantly mainstream, not tech-savvy enough to know the difference.

    I mean, how else can any site survive on advertising when the ads are so obnoxious and it’s so easy to block them? Either the site is great and the ads are non-intrusive enough that I’ll make an exception in uBlock, or I’m never seeing the ads in the first place.


  • Gemini might be good at something, but I’ll never know because it is bad at all the things I have ever used the assistant for. If it’s good at anything at all, it’s something I don’t need or want.

    Looking forward to 2027 when Google Gemini is replaced by Google Assistant (not to be confused with today’s Google Assistant, totally different product).


  • In case anyone is unfamiliar, Aaron Swartz downloaded a bunch of academic journals from JSTOR. This wasn’t for training AI, though. Swartz was an advocate for open access to scientific knowledge. Many papers are “open access” and yet are not readily available to the public.

    Much of what he downloaded was open-access, and he had legitimate access to the system via his university affiliation. The entire case was a sham. They charged him with wire fraud, unauthorized access to a computer system, breaking and entering, and a host of other trumped-up charges, because he…opened an unlocked closet door and used an ethernet jack from there. The fucking Secret Service was involved.

    https://en.wikipedia.org/wiki/Aaron_Swartz#Arrest_and_prosecution

    The federal prosecution involved what was characterized by numerous critics (such as former Nixon White House counsel John Dean) as an “overcharging” 13-count indictment and “overzealous”, “Nixonian” prosecution for alleged computer crimes, brought by then U.S. Attorney for Massachusetts Carmen Ortiz.

    Nothing Swartz did is anywhere close to the abuse by OpenAI, Meta, etc., who openly admit they pirated all their shit.


  • Joplin is great. I have its data stored locally with encryption, and I sync across devices with Syncthing. It also has built-in support for some cloud providers like you mentioned, and since it supports local encryption, you don’t need to depend on the cloud provider’s privacy policy.

    Setting it up on multiple devices was a bit complex, but the documentation is there. Follow the steps, don’t just waltz through the setup assuming it will work intuitively. I made that mistake and while it was not the end of the world, it would’ve saved me 15 minutes if I’d just RTFM.


  • Again: What is the percent “accurate” of an SEO infested blog

    I don’t think that’s a good comparison in context. If Forbes replaced all their bloggers with ChatGPT, that might very well be a net gain. But that’s not the use case we’re talking about. Nobody goes to Forbes as their first step for information anyway (I mean…I sure hope not…).

    The question shouldn’t be “we need this to be 100% accurate and never hallucinate” and instead be “What web pages or resources were used to create this answer” and then doing what we should always be doing: Checking the sources to see if they at least seem trustworthy.

    Correct.

    If we’re talking about an AI search summarizer, then the accuracy lies not in how correct the information is in regard to my query, but in how closely the AI summary matches the cited source material. Kagi does this pretty well. Last I checked, Bing and Google did it very badly. Not sure about Samsung.

    On top of that, the UX is critically important. In a traditional search engine, the source comes before the content. I can implicitly ignore any results from Forbes blogs. Even Kagi shunts the sources into footnotes. That’s not a great UX because it elevates unvetted information above its source. In this context, I think it’s fair to consider the quality of the source material as part of the “accuracy”, the same way I would when reading Wikipedia. If Wikipedia replaced their editors with ChatGPT, it would most certainly NOT be a net gain.


  • 99.999% would be fantastic.

    90% is not good enough to be a primary feature that discourages inspection (like a naive chatbot).

    What we have now is like…I dunno, anywhere from <1% to maybe 80% depending on your use case and definition of accuracy, I guess?

    I haven’t used Samsung’s stuff specifically. Some web search engines do cite their sources, and I find that to be a nice little time-saver. With the prevalence of SEO spam, most results have like one meaningful sentence buried in 10 paragraphs of nonsense. When the AI can effectively extract that tiny morsel of information, it’s great.

    Ideally, I don’t ever want to hear an AI’s opinion, and I don’t ever want information that’s baked into the model from training. I want it to process text with an awareness of complex grammar, syntax, and vocabulary. That’s what LLMs are actually good at.



  • In theory, the only difference between an electric heater and your computer, as far as actual heat goes, is the dispersal pattern. They will generate exactly the same heat: 1W of heat per 1W of electricity used. That’s thermodynamics for you!

    You said:

    The flat was kept not quite as warm as previous years

    So I don’t think it makes sense to assign any of the savings to using your PC vs your usual electric heaters. It’s because you kept your place a little cooler, which makes an absolutely huge difference. When heating in winter, every additional degree of air temperature is more costly than the last, since heat loss is relative to the temperature differential between indoors and outdoors (i.e. a warmer room will lose more heat to the outdoors than a cooler room, so you need to generate more heat to maintain it).

    This sounds to me a lot like dieting. Most of the time, the success of a diet has less to do with the actual diet and more to do with the fact that dieting has made you more mindful and changed your behavior in other ways.

    The two biggest things you can do to save money on heating in winter are:

    1. Keep your place cooler. Wear warm socks, long sleeves, etc. instead.
    2. Improve insulation. Plastic window insulation kits are cheap and easy to install/remove. For doorways, you can get adhesive insulating foam to fill side gaps and a slide-on door sweep to cover any bottom gaps.

  • I agree. Of all the UI crimes committed by Microsoft, this one wouldn’t crack the top 100. But I sure wouldn’t call it great.

    I can’t remember the last time I used the start menu to put my laptop to sleep. However, Windows Vista was released 20 years ago. At that time, most Windows users were not on laptops. Windows laptops were pretty much garbage until the Intel Core series, which launched a year later. In my offices, laptops were still the exception until the 2010s.


  • Google as an organization is simply dysfunctional. Everything they make is either some cowboy bullshit with no direction, or else it’s death by committee à la Microsoft.

    Google has always had a problem with incentives internally, where the only way to get promoted or get any recognition was to make something new. So their most talented devs would make some cool new thing, and then it would immediately stagnate and eventually die of neglect as they either got their promotion or moved on to another flashy new thing. If you’ve ever wondered why Google kills so many products (even well-loved ones), this is why. There’s no glory in maintaining someone else’s work.

    But now I think Google has entered a new phase, and they are simply the new Microsoft – too successful for their own good, and bloated as a result, with too many levels of management trying to justify their existence. I keep thinking of this article by a Microsoft engineer around the time Vista came out, about how something like 40 people were involved in redesigning the power options in the start menu, how it took over a year, and how it was an absolute shitshow. It’s an eye-opening read: https://moishelettvin.blogspot.com/2006/11/windows-shutdown-crapfest.html



  • IPFS content IDs (CID) are a hash of the tree of chunks. Changes to chunk size can also change the hash!

    I don’t understand why this is a deal-breaker. It seems like you could accomplish what you describe within IPFS simply by committing to a fixed chunk size. That’s valid within IPFS, right?

    Is it important to use any specific hashing algorithm(s)? If not, then isn’t an IPFS CID (with a fixed, predetermined chunk size) a stable hash algorithm in and of itself?