does anyone have a list of about 2000 urls, i need them for a project but searching google and copying and pasting each url is taking forever. Any help would be grateful.
does anyone have a list of about 2000 urls, i need them for a project but searching google and copying and pasting each url is taking forever. Any help would be grateful.
So write a simple program to programmatically extract URLs from an HTML page.
Or maybe do the same from your browser cache.
If you dance barefoot on the broken glass of undefined behaviour, you've got to expect the occasional cut.
If at first you don't succeed, try writing your phone number on the exam paper.
In HTML, URLs for different webpages have
href="The URL"
For images and other items, it's
src="The URL"
Sometimes the href or src part doesn't have quotes and other times, there's no "http://www.examplewebsite.com/" beginning where "examplewebsite" is the website's URL. Mailto's should be ignored - those are for E-mails. This should help you get started. Using copy/paste, doing 2 every minute will take all day long to do for 2000.
i'd like to create a program to do it just have no idea what i'm doing maybe someone can hook me up or something? It would be appreciated beyond belief.
I bet you could write 2000 random words surrounded by www. and .com and get damn near 2000 links.
-Govtcheez
[email protected]
yea the point is i don't want to type all of that
> does anyone have a list of 2000 urls?
Um, just look in a Web directory.
for instance, http://dir.yahoo.com/
You should be able to find all sorts of sites there.
yea but i still have to copy and past it would be much better if i could just have a script or a list and not have to copy and paste it
Jesus Christ, since your lazy ass posted this thread, you could have done this 3 times.
-Govtcheez
[email protected]
Here....some people might get mad at me for basically giving you the apple here...but anyways...For my Internet Programming course this past semester, as the first half of our first project, we had to write a web client that would accept a web address, download that page, and then download all pages linked to that page.
The code is fairly well documented, so if you can read code, and my own documentation, you should be able to understand what is going on. It does not do exactly what you want to do, but you can make it do what you want to do by seriously changing only about 5 lines of code. It should compile fine using g++. If you can't understand the code, then you need to study socket programming a bit, and then come back....but honestly you should only need to change a few lines of code in the main() function.
Oh..the main() function is in client.cpp
It will not run in Windows...only in Linux (and possibly Mac)...I use the Unix Socket API.
This should give you a start...now go for it.
Oh...change the extension from .txt to .zip...it is a compressed zip folder with my code.
Like I said, my code doesn't do what you want to do, but it does something similar, and if you can understand code, you should be able to make it do what you want it to do by changing less than 5 lines.
He's probably going to ask you to write, compile, and release a windows-compatible version.
Better start a sourceforge group to help Captain Dumbass finish his homework.
-Govtcheez
[email protected]