Depending on your ultimate objective this can be approached in many ways. We have used Robocopy which is embedded in Windows operating systems and then had our IT operating systems folk run a command which tells us the size and volume, number of files on particular drives. We then run a command to split files on the drives, based on Last Accessed Date (must be touched in the last 3 years) into two sections and have users focus on the last 3 years for initial migration.
We have used Duplicate File Finder (free download) to find and delete duplicates as well as have users clean up on their own:
We then use ShareGate to migrate documents.
After implementation, drives are set to read only and an assessment is made of what is left and of true value to migrate (based on business value, records retention etc). This backlog is then done over a period of 1 year.
Rachel Larson, ermm, CIP
Analyst III, EDRM
Governance, Risk and Compliance
Business Technology Services
2121 11th Street West
Saskatoon, SK S7M 1J3
*please consider the environment before printing this email
This email and any files transmitted with it are personal and confidential, and are solely for the use of the individual or entity addressed. Therefore, if you are not the intended recipient, please delete this email and any files transmitted with it (without making any copies) and advise the author immediately.
I asked our SharePoint Consultant, Pramit John for suggestions. He said: 'We use AvePoint over here and it’s been pretty good. I think even Quest (www.quest.com ) has tools to do the file share migration. For analysis I would definitely break the source into different logical sections. Use PowerShell to do the analysis and export the data into a database table and use a BI tool to do the analysis. As Robert mentioned, even I have seen this issue in Excel where it crashes due to the huge volume of records, typically in millions, or it trims the content'.
Please Check out Adlib Software, Adlib is helping to address a gap in the market with a unique file analytics approach, which leverages the creation of standardized high-definition content and classification with built-in repeatable intelligence. Industry adoption of file analysis technologies like Adlib Elevate is growing and predicted to jump from 10% to 50% by 2020.