The Sci2 Tool's database functionality greatly improves the volume of data which can be loaded into and analyzed by the tool. Whereas most scientometric tools available as of March 2010 require powerful computing resources to perform large scale analyses each time a network needs to be extracted, the pre-loaded database runs network extraction algorithms quickly and allows the users to format their own custom queries. This functionality has some front-heavy memory requirements in order to initially load the data into the database, the upper limits of which can be seen in Tables 3.1 and 3.2.
Entries | Load | Merge | Merge | Match | Extract | Extract | Extract | Extract | Extract | Extract | Extract | Author |
---|---|---|---|---|---|---|---|---|---|---|---|---|
50 | 5 | 1 | 1 | 1 | 1 | 1 | 1 | 1 | 1 | 10 | 4 | 2 |
500 | 10 | 4 | 4 | 1 | 1 | 1 | 1 | 1 | 1 | 20 | 4 | 5 |
5000 | 100 | 48 | 180 | 3 | 1 | 1 | 3 | 18 | 2 | 3300 | 20 | 35 |
10000 | 210 | 120 | 480 | 5 | 2 | 1 | 10 | 30 | 7 |
| 60 | 100 |
20000 | 400 | 300 | 1200 | 15 | 5 | 1 | 22 | 70 | 20 |
| 164 | 250 |
Table 3.1: The number of seconds to perform each action on a dataset of the given size, on a computer with 12GB of memory
Entries | Load | Merge | Merge | Match | Extract | Extract | Extract | Extract | Extract | Extract | Extract | Author |
---|---|---|---|---|---|---|---|---|---|---|---|---|
50 | 50 | 55 | 9 | 4 | 3 | 1 | 1 | 1 | 1 | 1 | 40 | 13 |
500 | 500 | 420 | 52 | 13 | 3 | 1 | 1 | 1 | 2 | 1 | 75 | 13 |
5000 | 5000 | 3600 | 1080 | 480 | 25 | 1 | 1 | 12 | 90 | 2 |
| 180 |
Table 3.2: The number of seconds to perform each action on a dataset of the given size, on a computer with 1.2GB of memory
Attachments:



