This page last changed on Feb 04, 2011 by dapolley.

The Sci2 Tool's database functionality greatly improves the volume of data which can be loaded into and analyzed by the tool. Whereas most scientometric tools available as of March 2010 require powerful computing resources to perform large scale analyses each time a network needs to be extracted, the pre-loaded database runs network extraction algorithms quickly and allows the users to format their own custom queries. This functionality has some front-heavy memory requirements in order to initially load the data into the database, the upper limits of which can be seen in Tables 3.1 and 3.2.

Entries Load
Merge
People
Merge
Journals
Match
Extract
Authors
Extract
Documents
Extract
Co-Authors
Extract
Document
Citation
(with outer)
Extract
Author
Citation
Extract
Document
Co-Citation
Extract
Document
Co-Citation
(core)
Author
Co-Citation
50
5
1
1
1
1
1
1
1
1
10
4
2
500
10
4
4
1
1
1
1
1
1
20
4
5
5000
100
48
180
3
1
1
3
18
2
3300
20
35
10000
210
120
480
5
2
1
10
30
7
  60
100
20000
400
300
1200
15
5
1
22
70
20
  164
250

Table 3.1: The number of seconds to perform each action on a dataset of the given size, on a computer with 12GB of memory

Entries Load
Merge
People
Merge
Journals
Match
Extract
Authors
Extract
Documents
Extract
Co-Authors
Extract
Document
Citation
(with outer)
Extract
Author
Citation
Extract
Document
Co-Citation
Extract
Document
Co-Citation
(core)
Author
Co-Citation
50
50
55
9
4
3
1
1
1
1
1
40
13
500
500
420
52
13
3
1
1
1
2
1
75
13
5000
5000
3600
1080
480
25
1
1
12
90
2
  180

Table 3.2: The number of seconds to perform each action on a dataset of the given size, on a computer with 1.2GB of memory


Document generated by Confluence on May 31, 2011 15:16