• 0 Posts
  • 21 Comments
Joined 1 year ago
cake
Cake day: July 9th, 2023

help-circle

  • This kind of reminds me of Crispin Glover, from Back to the Future. He tried to negotiate a higher pay for the second movie, so the producers hired a different actor to play the role, but deliberately made the actor up to look like Glover. In response, Glover sued the producers and won. It set a critical precedent for Hollywood, about using someone’s likeness without consent.

    The article mentions they reached out to her two days before the launch - if she had said ‘OK,’ there’s no way they could have even recorded what they needed from her, let alone trained the model in time for the presentation. So they must have had a Scarlett Johansson voice ready to go. Other than training the model on movies (really not ideal for a high quality voice model), how would they have gotten the recordings they needed?

    If they hired a “random” voice actress, they might not run into issues. But if at any point they had a job listing, a discussion with a talent manager, or anything else where they mentioned wanting a “Scarlett Johansson sound-alike,” they might have dug themselves a nice hole here.

    Specifically regarding your question about hiring a voice actor that sounds like someone else - this is commonly done to replace people for cartoons. I don’t think it’s an issue if you are playing a character. But if you deliberately impersonate a person, there might be some trouble.




  • Considering that you are not using their software, was the laptop worth the premium you paid for it, vs buying from Clevo directly?

    I figured the hardware and software coming from the same vendor would yield the best results, and wanted to support a company that supports right-to-repair, and Linux in general. But ultimately I found Pop!_OS buggy and had performance issues, so I’m not using their OS, and their firmware is causing issues with my SSD, so I’d like to be off of it as well (but was told "there’s no process for reverting to the proprietary firmware“ for the specific model I have). I could have bought a Clevo directly, saving hundreds of dollars, and probably had a better working machine.


  • JoeyJoeJoeJr@lemmy.mltoLinux@lemmy.mlDell is so frustrating
    link
    fedilink
    arrow-up
    6
    arrow-down
    1
    ·
    3 months ago

    Would not recommend System76. I’ve had many issues with my machine (primarily software, related to their buggy custom firmware, and Pop!_OS, until I ditched that for stock Ubuntu). Their support has been terrible - rather similar to OP’s, actually. I’ve had the laptop for about 2.5 years, and I’m checking practically daily for something to replace it.



  • My usb-c ports can be a little touchy, too. The SD card slot is also really bad - the card has to be positioned perfectly to slide in, or it jams. I’m also upset that the usb-c port can only be used for charging after a full boot. It cannot be used to perform firmware updates, or even to do a ram test. This means day-to-day, usb-c can be used, but I have to keep track of the barrel charger, just in case. This, of course, was not specified on the product details page (nor, I think, that only one of the two usb-c ports could be used for charging - it’s possible I overlooked that, but still frustrating on an expensive laptop that lists usb-c charging as a feature).


  • JoeyJoeJoeJr@lemmy.mltoLinux@lemmy.mlLaptop companies: which one?
    link
    fedilink
    arrow-up
    33
    arrow-down
    1
    ·
    edit-2
    5 months ago

    I currently have a System76 laptop, and sincerely regret my purchase. When I purchased it, the Framework was not out yet - I wanted to support a company that supports right-to-repair, and figured since they controlled the hardware, firmware, and software (Pop!_OS), it would be a good, stable experience. It has not been, and support has generally been poor. I know other people have had better experiences than I have, but personally, I won’t be buying from them again.

    I haven’t personally used Purism, but former co-workers spoke really poorly of them. They were trying to buy a big batch for work, and said the build quality was awful. Additionally: https://youtu.be/wKegmu0V75s



  • I noted in another comment that SearXNG can’t do anything about the trackers that your browser can’t do, and solving this at the browser level is a much better solution, because it protects you everywhere, rather than just on the search engine.

    Routing over Tor is similar. Yes, you can route the search from your SearXNG instance to Google (or whatever upstream engine) over Tor, and hide your identity from Google. But then you click a link, and your IP connects to the IP of whatever site the results link to, and your ISP sees that. Knowing where you land can tell your ISP a lot about what you searched for. And the site you connected to knows your IP, so they get even more information - they know every action you took on the site, and everything you viewed. If you want to protect all of that, you should just use Tor on your computer, and protect every connection.

    This is the same argument for using Signal vs WhatsApp - yes, in WhatsApp the conversation may be E2E encrypted, but the metadata about who you’re chatting with, for how long, etc is all still very valuable to Meta.

    To reiterate/clarify what I’ve said elsewhere, I’m not making the case that people shouldn’t use SearXNG at all, only that their privacy claims are overstated, and if your goal is privacy, all the levels of security you would apply to SearXNG should be applied at your device level: Use a browser/extension to block trackers, use Tor to protect all your traffic, etc.


  • They are explicitly trying to move away from Google, and are looking for a new option because their current solution is forcing them to turn off ad-blocking. Sounds to me like they are looking for a private option. Plus, given the forum in which we are having the discussion (Lemmy), even if OP is not specifically concerned with privacy, it seems likely other users are.

    As for cookies, searxng can’t do any more than your browser (possibly with extensions) can do, and relying on your browser here is a much better solution, because it protects you on all sites, rather than just on your chosen search engine.

    “Trash mountain” results is a whole separate issue - you can certainly tune the results to your liking. But literally the second sentence of their GitHub headline is touting no tracking or profiling, so it seems worth bringing attention to the limitations, and that’s all I’m trying to do here.



  • It looks like a few people are recommending this, so just a quick note in case people are unaware:

    If you want to avoid being tracked, this is not a good solution. Searxng is a meta search engine, meaning it is effectively a proxy: you search on Searxng, it searches multiple sites and sends all the results back to you. If you use a public instance, you may be protected from the actual search engine*, because many people will use the same instance, and your queries will be mixed in with all of them. If you self host, however, all the searches will be your own - there is then no difference between using Searxng and just going to the site yourself.

    *The caveat with using the public instances is while you may be protected from the upstream engine, you have to trust the admins - nothing stops them from tracking you themselves (or passing your data on).

    Despite the claims in their docs, I would not consider this a privacy tool. If you are just looking for a good search engine, this may work, and it gives you flexibility and power to tune it yourself. But it’s probably not going to do anything good for your privacy, above and beyond what you can get from other meta search engines like Startpage and DuckDuckGo, or other “private” search engines like Brave.



  • This is approximately what I do as well, and would highly recommend. The one caveat I would add is while you are researching things you might want to do, take note of the subset of things you most want to do, and make sure you know what days/times they are open, if you need to book in advance, etc. I am very against having a hard schedule, but I also don’t want to travel somewhere only to miss the one thing I was really looking forward to because I decided “I’ll do that tomorrow,” only to find out it was closed the next day.

    An additional pro-tip: Make your first list of things you might want to do ahead of time, and name it after the place you are going, e.g. “New York.” Then while you’re traveling, make a second list of “favorites”, e.g. “New York Favorites.” Keep track of all the restaurants, activities, view points, etc that you enjoyed using that second list. Then whenever someone asks for recommendations for a particular location, you can just send them your favorites list.



  • I did a site:reddit.com search using my username and found ~50 comments that Reddit has undeleted but also hide from my own account. I could still edit and delete them.

    Perhaps you should re-read the post, and/or the comments here. The posts referenced are still live on Reddit, hence they can still be edited - OP is not talking about a cached view from the search engine.


  • it has its flaws.

    Yep yep. I was aware of some of what you pointed out - I think this might be a “perfect is the enemy of good” scenario, though. GitHub alone accounts for over 84% (based on the awesome-selfhosted-data repo):

    $ grep -r 'source_code_url' | cut -d ' ' -f 2 | cut -d '/' -f 3 | sort | uniq -c | sort -rn | head -n 15
       1068 github.com
         36 gitlab.com
          7 git.mills.io
          6 sourceforge.net
          6 framagit.org
          4 www.atlassian.com
          4 codeberg.org
          3 git.drupalcode.org
          3 git.cloudron.io
          2 repos.goffi.org
          2 git.tt-rss.org
          2 git.sr.ht
          2 cvsweb.openbsd.org
          1 yetishare.com
          1 www.wiz.cn
    
    $ python -c "print($(grep -r 'source_code_url' . | grep github.com | wc -l) / $(ls -1 | wc -l))"
    0.8422712933753943
    

    Adding in gitlab gets you to 87%:

    $ python -c "print($(grep -r 'source_code_url' . | grep -i -e github.com -e gitlab.com | wc -l) / $(ls -1 | wc -l))" 0.8706624605678234

    Also popularity != quality.

    True, but a thriving community generally means more resources, guides, etc, which can be important, especially for self-hosted solutions.

    In any case, the project is great, and much appreciated. Additionally, the enriched html version looks fantastic, and exposes most of the metadata* I’d want to see, regardless of how it’s sorted.

    *One other item to track, that I thought about after making my previous comment - number of contributors. It gives an additional data point on the size of the community, as well as an idea of how many people can be hit by busses before the continued development of the project gets called into question.


  • I would imagine the source for most projects is hosted on GitHub, or similar platforms? Perhaps you could consider forks, stars, and followers as “votes” and sort each sub category based on the votes. I would imagine that would be scriptable - the script could be included in the awesome list repo, and run periodically. It would be kind of interesting to tag “releases” and see how the sort order changes over time. If you wanted to get fancy, the sorting could probably happen as part of a CI task.

    If workable, the obvious benefit is you don’t have to exclude anything for subjective reasons, but it’s easier for readers of the list to quickly find the “most used” options.

    Just an idea off the top of my head. You may have already thought about it, and/or it may be full of holes.