Results 1 to 6 of 6

Thread: html pages download from website

  1. #1
    Join Date
    Sep 2009
    Beans
    32

    html pages download from website

    does anyone know of a good website copier that runs on ubuntu and that downloads and saves html files so that i can open them without the need of the original program.

  2. #2
    Join Date
    Jan 2011
    Beans
    1,151

    Re: html pages download from website

    If you just want to download the html text code then I guess on command line

    Code:
    wget -r [url]
    example
    Code:
    wget -r google.com
    Website are built differently, depending on the website, they might thwart you efforts.

    All html code can be opened by any webbrowser
    (\ /)
    (O.o)
    (> <)
    This is Bunny. Copy Bunny into your signature to help him on his way to world domination.

  3. #3
    Join Date
    Sep 2009
    Beans
    32

    Re: html pages download from website

    Quote Originally Posted by idoitprone View Post
    If you just want to download the html text code then I guess on command line

    Code:
    wget -r [url]
    example
    Code:
    wget -r google.com
    Website are built differently, depending on the website, they might thwart you efforts.

    All html code can be opened by any webbrowser
    I mean i want to download at least a small part of a website. I only know the homapage url. So with wget i may not be able to solve my problem.

  4. #4
    Join Date
    Feb 2006
    Location
    uk
    Beans
    Hidden!

    Re: html pages download from website

    if you don't know the location of the data, how will your software?

    wget can open links recursively, and you can specify recursion depth, but if there are no links on the front page and no way to figure out where the content is, you wont get very far

  5. #5
    Join Date
    Sep 2009
    Beans
    32

    Re: html pages download from website

    Quote Originally Posted by aeiah View Post
    if you don't know the location of the data, how will your software?

    wget can open links recursively, and you can specify recursion depth, but if there are no links on the front page and no way to figure out where the content is, you wont get very far
    Could you please tell me how to make wget recursively download html pages ?

  6. #6
    Join Date
    Sep 2006
    Beans
    8,627
    Distro
    Ubuntu 14.04 Trusty Tahr

    Re: html pages download from website

    Quote Originally Posted by coffeecake View Post
    Could you please tell me how to make wget recursively download html pages ?
    It's described above in post #2. It is done with the help of the -r or --recursive option. Other options like --no-parent, --convert-links and --page-requisites might be useful to look at, too.

Tags for this Thread

Bookmarks

Posting Permissions

  • You may not post new threads
  • You may not post replies
  • You may not post attachments
  • You may not edit your posts
  •