Author |
Topic: Website Rankings (Read 423 times) |
|
TenaliRaman
Uberpuzzler
I am no special. I am only passionately curious.
Gender:
Posts: 1001
|
|
Website Rankings
« on: Nov 25th, 2003, 1:10pm » |
Quote Modify
|
(i thought first of putting it in CS section but then i thought this system should be known by everyone as a general information) You have created 10 web pages with one as your home page. How will you link them so as to maximise the average PR of your site? PR = page ranking we will follow google page ranking scheme here. our page ranking will be only based on backward links and no other factors. Assume that there are no links coming from other websites.
|
« Last Edit: Nov 25th, 2003, 1:11pm by TenaliRaman » |
IP Logged |
Self discovery comes when a man measures himself against an obstacle - Antoine de Saint Exupery
|
|
|
Jamie
Newbie
Gender:
Posts: 37
|
|
Re: Website Rankings
« Reply #1 on: Dec 1st, 2003, 4:16am » |
Quote Modify
|
I suspect you're going to have clarify this one a bit. For example, why not just link every page from every other? on Nov 25th, 2003, 1:10pm, TenaliRaman wrote:Assume that there are no links coming from other websites. |
| Hmmm... How is Google going to find our page in the first place?
|
|
IP Logged |
|
|
|
TenaliRaman
Uberpuzzler
I am no special. I am only passionately curious.
Gender:
Posts: 1001
|
|
Re: Website Rankings
« Reply #2 on: Dec 1st, 2003, 6:55am » |
Quote Modify
|
every inward coming link to a page increases the page's PR and every outgoing link reduces the page's PR. This is sort of the action - counteraction. The assumption basically is to simplify calculations a bit since incoming links from other sites would drastically increase your PR. Usually google does not find your page, you need to register with them. you might information on this from www.w3schools.com . [edit] As a general note to all, the google search engine and other stuff explained here. This i think is the official paper of google, http://www-db.stanford.edu/~backrub/google.html [/edit]
|
« Last Edit: Dec 1st, 2003, 7:01am by TenaliRaman » |
IP Logged |
Self discovery comes when a man measures himself against an obstacle - Antoine de Saint Exupery
|
|
|
aero_guy
Senior Riddler
Gender:
Posts: 513
|
|
Re: Website Rankings
« Reply #3 on: Dec 1st, 2003, 11:50am » |
Quote Modify
|
It seems (if I am understanding you) that in any contained system the average ranking will be zero. Every link counts as one point for the page being linked to and one against the one being linked from. If both the to and from pages are on your site than you will always have a zero average page ranking. For example, the average page ranking of the internet itself would be zero (actually it would be negative considering the amount of blind links).
|
|
IP Logged |
|
|
|
TenaliRaman
Uberpuzzler
I am no special. I am only passionately curious.
Gender:
Posts: 1001
|
|
Re: Website Rankings
« Reply #4 on: Dec 2nd, 2003, 10:31am » |
Quote Modify
|
yes aero, it would be zero if it had not been for something called as damping factor. The official paper which i had linked to in my last post gives a formula for calculating PR of a web page and it includes this damping factor. The formula gives the lowest possible value of (1-damping_factor) to any page.
|
|
IP Logged |
Self discovery comes when a man measures himself against an obstacle - Antoine de Saint Exupery
|
|
|
aero_guy
Senior Riddler
Gender:
Posts: 513
|
|
Re: Website Rankings
« Reply #5 on: Dec 2nd, 2003, 3:42pm » |
Quote Modify
|
From your link right under where they present the formula (along with damping factor): Note that the PageRanks form a probability distribution over web pages, so the sum of all web pages' PageRanks will be one. Therefore the average for ten self-contained pages will be .1 no matter how you link them.
|
« Last Edit: Dec 2nd, 2003, 3:44pm by aero_guy » |
IP Logged |
|
|
|
TenaliRaman
Uberpuzzler
I am no special. I am only passionately curious.
Gender:
Posts: 1001
|
|
Re: Website Rankings
« Reply #6 on: Dec 3rd, 2003, 1:54am » |
Quote Modify
|
hmm, mean = E(x) = sum over xP(x) (for discrete probabilities) and integral over xP(x) (for continuous probabilities) not necessarily (sum of probs)/10 (or am i making a mistake?)
|
|
IP Logged |
Self discovery comes when a man measures himself against an obstacle - Antoine de Saint Exupery
|
|
|
towr
wu::riddles Moderator Uberpuzzler
Some people are average, some are just mean.
Gender:
Posts: 13730
|
|
Re: Website Rankings
« Reply #7 on: Dec 3rd, 2003, 3:11am » |
Quote Modify
|
P(x) would have to be the chance of a certain webpage, where x would be the PR of that webpage.. I don't see a reason not to take every P(x) as 1/10, that is each page is weighed equally in the average. And since PR sums to 1 we get 1/10
|
|
IP Logged |
Wikipedia, Google, Mathworld, Integer sequence DB
|
|
|
TenaliRaman
Uberpuzzler
I am no special. I am only passionately curious.
Gender:
Posts: 1001
|
|
Re: Website Rankings
« Reply #8 on: Dec 3rd, 2003, 4:22am » |
Quote Modify
|
hmm, i could accept the argument (but if i do understand google's formula , the lowest possible value a page gets is 0.15 no??)
|
|
IP Logged |
Self discovery comes when a man measures himself against an obstacle - Antoine de Saint Exupery
|
|
|
towr
wu::riddles Moderator Uberpuzzler
Some people are average, some are just mean.
Gender:
Posts: 13730
|
|
Re: Website Rankings
« Reply #9 on: Dec 3rd, 2003, 6:07am » |
Quote Modify
|
Hmm.. it does look peculiar.. One the one hand they say all page rankings of all webpages must add up to one, and one the other hand the minimum seems (1-d) (where d is typically 0.85). The only solution for this I see is that the other terms may be negative.. But I guess I'll have to actually read the article to fidn out how it really works..
|
|
IP Logged |
Wikipedia, Google, Mathworld, Integer sequence DB
|
|
|
TenaliRaman
Uberpuzzler
I am no special. I am only passionately curious.
Gender:
Posts: 1001
|
|
Re: Website Rankings
« Reply #10 on: Dec 3rd, 2003, 8:06am » |
Quote Modify
|
Yes it is peculiar. if you note , the formula given is a recursive formula. If page A has link to page B and page B has link page A, then you would normally expect this formula to give a very high result (and it would if it had not been for that 'd'). I did some calculations and it seems to stabilise at one point but some of my calculations have given me average PR's well above 1(?).I might have probably made a mistake in either understanding the formula or in my calculations. I thought i understood it pretty well but i never gave the thought to the statement of the probability distribution. Hopefully you would get it and then explain it to me
|
|
IP Logged |
Self discovery comes when a man measures himself against an obstacle - Antoine de Saint Exupery
|
|
|
|