Uploaded image for project: 'Picard'
  1. Picard
  2. PICARD-1432

Application performance while clustering large libraries



    • Type: Bug
    • Status: Closed
    • Priority: Normal
    • Resolution: Fixed
    • Affects Version/s: 2.0.4
    • Fix Version/s: None
    • Component/s: None
    • Labels:
    • Environment:
      Windows 10 1603, Core i7, 8GB, SSD


      When working with over 3000 files, the system is unresponsive but is working on whatever task at hand. Knowing that the application performance is poor but still accomplishes what it needs to, I am currently working with 42k files and it is processing about 500-750 an hour to perform clustering. The CPU is only taxed at 17-20% and the applications memory is under 1.5GB.
      Performance degrades considerably when working with too many files at once. If I had worked on the 3k files alone, it would have completed in a few hours, but because I have 42k files, the same 3000 files of processing has been running for 7 hours.

      There are a couple things I think could be adjusted.

      • - Provide an option to enable more threads to be used.
      • - Seems the application is limiting itself to never go above 20% CPU utilization. In 15 hours of churning, looking back at perfmon logs, the application seems to cap at 20%.
      • - Provide an interrupt/cancel button that stops the pending workstream. Say for example someone selects all clusters and chooses scan. There should be a way to stop the scan and select a smaller number of files.


          Issue Links



              • Assignee:
                ScarFraags Jason Condo
              • Votes:
                2 Vote for this issue
                4 Start watching this issue


                • Created:


                  Version Package