[Top][All Lists]
[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]
[igraph] betweenness running time estimate and
From: |
David Hunkins |
Subject: |
[igraph] betweenness running time estimate and |
Date: |
Sun, 13 Jul 2008 14:16:13 -0700 |
Thank you Gabor, Tamas and others very much for the igraph library.
I am analyzing some online photo sharing data donated to me by a large
photo sharing site.
In my first dataset I have about 5 million vertices and 5 million
edges. In the next dataset I'll get about 50 million vertices and 200
million edges. I am currently running analysis on a 2GB/2.0GHz Intel
Core Duo Macbook Pro. I am conditioning and loading the weighted,
directed graph data in Pajek format.
A couple of questions with which your user community may have some
experience.
First, I am having a trouble determining how long I should expect
various operations to take even with the 5M/5M dataset, and I am
wondering whether there is a simple metric (such as density of the
graph) that can be computed and then used to calculate order-of-
magnitude running times for the various other functions of interest.
Yesterday, for example I ran betweenness.estimate(graph, cutoff=2) and
it hasn't completed after 24 hours (and the Macbook is not paging at
all, just getting a little 'warm').
Second, do you know anybody who has used commercial (or otherwise)
scalable computing resources to run analyses on datasets of this size?
I fear that, even if I can melt my laptops to solve the problem for
the 5M/5M dataset, I will need other resources when I go to 50M/200M
dataset.
Thanks for any assistance you and the community can provide. Kösz!
David Hunkins
- [igraph] betweenness running time estimate and,
David Hunkins <=