Skip Navigation
Checksum for original files came out unreadable
  • What? The hash is just a number in a text file. Open the md5 with notepad?

    You aren't very clear... Are you saying it creates the hash, but then verifying fails? I don't know what you mean by unreadable

  • High Command Timeout & Read Error Rate on ST8000NM000A-2KE101 - Seagate Exos 7E8 8TB - Disk access is just too slow painfully
  • The read and seek error rates appear to be zero. Any errors would be recorded in the top 4 digits. Different manufacturers use the fields differently.

    Not sure about command timeout off the top of my head

    The ultraDMA CRC actually looks bad on the 1st one I think.

    Check some real documentation for seagate though to know what the fields mean

  • There is any app to download YouTube videos placed on a playlist?
  • Yt-dlp does pretty much everything. It can do a channels playlists, not sure about a custom saved one, but try feeding it a link to it? Then you can just keep re-running it to download any new additions

  • How do i make wget not logout itself?
  • Is it the website kicking it off? Can you run with --debug and post what it says when the request fails?

    I'll be back at my computer later and can send over an idea to try. I have a wgetrc file that has gotten around some issues before

  • Could someone please give me a walk through on how to crawl an entire web domain and scrape the images only?
  • WGET is awesome, I have scraped tons with it. So many options, you can even spoof all the request header info to get around sites that try to limit auto downloaders. Here is the manual: https://www.gnu.org/software/wget/manual/wget.html

    1. webp or any file extension will work. (note on webp, most sites actually have jpgs still, but convert and serve webp to save bandwidth if the browser says it accepts them. There is a header you can disable in firefox to not accept webp unless it is the only option:

    https://addons.mozilla.org/en-CA/firefox/addon/dont-accept-webp/

    Wget is not behaving identically to a browser so im unsure what this part of the request looks like or if it needs modification. If it isnt working let me know.

    1. 5 might be enough, but maybe not. Scroll down in my first link comments, they show how to set to infinite: "-l inf".

    For future scraping, look at the mirror command. It sets recursion to infinite and will make a full copy of the site. You can also use the --convert-links option, which changes all the links to point to the locally downloaded files. It then behaves the same as the real website.

    You cant go too deep unless you use --span-hosts, it can grab external files from different domains to make the mirrored site a true copy, but yea, you often don't need that. You also want to be more careful with recursive depth here - it can go too deep and you end up with too much data.

    1. I'm not sure about this. I think you can turn on logging, but I'm not sure what that gets you. I've used the no-clobber command to run wget again, without re-downloading existing files. This is handy for resuming or filling in gaps that were missed due to timeout, etc.

    Some sites also need to use the wait or random-wait command to avoid detection.

  • InitialsDiceBearhttps://github.com/dicebear/dicebearhttps://creativecommons.org/publicdomain/zero/1.0/„Initials” (https://github.com/dicebear/dicebear) by „DiceBear”, licensed under „CC0 1.0” (https://creativecommons.org/publicdomain/zero/1.0/)PL
    plunki @alien.top
    BOT
    Posts 0
    Comments 15