It just takes a lot of time to run reclink on a large dataset.
But I have three quite good computers.
I thought maybe I can do it quickly by dividing the data into three pieces and run each subdata on each computer.
But I can't be sure if it will speed it up because I don't know how reclink actually works internally.
Which dataset should I divide by three to speed this up? Master data or using data?
If the number of company names is 100K, then does reclink take roughly half the time than when there are 200K company names?
But I have three quite good computers.
I thought maybe I can do it quickly by dividing the data into three pieces and run each subdata on each computer.
But I can't be sure if it will speed it up because I don't know how reclink actually works internally.
Which dataset should I divide by three to speed this up? Master data or using data?
If the number of company names is 100K, then does reclink take roughly half the time than when there are 200K company names?
Comment