Join PrimeGrid
Returning Participants
Community
Leader Boards
Results
Other
drummers-lowrise
|
Message boards :
Number crunching :
Summer Paralympics Challenge
Author |
Message |
Michael Goetz Volunteer moderator Project administrator
 Send message
Joined: 21 Jan 10 Posts: 13955 ID: 53948 Credit: 392,876,449 RAC: 184,453
                               
|
Welcome to Summer Paralympics Challenge
To participate in the Challenge, please select only the Extended Sierpinski Problem (LLR) (aka ESP-LLR) project in your PrimeGrid preferences section. The challenge will begin 2 September 2016 18:00 UTC and end 7 September 2016 18:00 UTC. Application builds are available for Linux , Windows and MacIntel 32 bit and 64 bit. Intel CPU's with AVX capabilities will be significantly faster than the ones without (including all AMD CPUs), as this instruction set allows for more computing power.
ATTENTION: The primality program LLR is CPU intensive; so, it is vital to have a stable system with good cooling. It does not tolerate "even the slightest of errors." Please see this post for more details on how you can "stress test" your computer. WU's may take around 6 hours on the fastest/newest computers and more than a day on slower/older computers. If your computer is highly overclocked, please consider "stress testing" it. Sieving is an excellent alternative for computers that are not able to LLR. :)
Highly overclocked Haswell or Skylake (i.e., Intel Core i7, i5, and i3 -4xxx or better) computers running the application will see fastest times. Note that ESP is now running the latest, brand new FMA3 version of LLR which takes full advantage of the new Haswell features. It's faster than the previous LLR app and draws more power and produces more heat. If you have a Haswell or Skylake CPU, especially if it's overclocked or has overclocked memory, and haven't run the new FMA3 LLR before, we strongly suggest running it before the challenge while you are monitoring the temperatures.
Please, please, please make sure your machines are up to the task.
Time zone converter:
The World Clock - Time Zone Converter
NOTE: The countdown clock on the front page uses the host computer time. Therefore, if your computer time is off, so will the countdown clock. For precise timing, use the UTC Time in the data section to the left of the countdown clock.
Scoring Information
Scores will be kept for individuals and teams. Only work units issued AFTER 2 September 2016 18:00 UTC and received BEFORE 7 Septemberl 2016 18:00 UTC will be considered for credit. We will use the same scoring method as for BOINC credit.
Therefore, each completed WU will earn a unique score based on its n value. The higher the n, the higher the score. A quorum of 2 is NOT needed to award Challenge score - i.e. no double checker. Therefore, each returned result will earn a Challenge score. Please note that if the result is eventually declared invalid, the score will be removed.
At the Conclusion of the Challenge
We kindly ask users "moving on" to ABORT their WU's instead of DETACHING, RESETTING, or PAUSING.
ABORTING WU's allows them to be recycled immediately; thus a much faster "clean up" to the end of an LLR Challenge. DETACHING, RESETTING, and PAUSING WU's causes them to remain in limbo until they EXPIRE. Therefore, we must wait until WU's expire to send them out to be completed.
Please consider either completing what's in the queue or ABORTING them. Thank you. :)
About the Extended Sierpinski Project
Waclaw Franciszek Sierpinski (14 March 1882 - 21 October 1969), a Polish mathematician, was known for outstanding contributions to set theory, number theory, theory of functions and topology. It is in number theory where we find the Sierpinski problem.
Basically, the Sierpinski problem is "What is the smallest Sierpinski number" and the prime Sierpinski problem is "What is the smallest 'prime' Sierpinski number?"
First we look at Proth numbers (named after the French mathematician François Proth). A Proth number is a number of the form k*2^n+1 where k is odd, n is a positive integer, and 2^n>k.
A Sierpinski number is an odd k such that the Proth number k*2^n+1 is not prime for all n. For example, 3 is not a Sierpinski number because n=2 produces a prime number (3*2^2+1=13). In 1962, John Selfridge proved that 78,557 is a Sierpinski number...meaning he showed that for all n, 78557*2^n+1 was not prime.
Most number theorists believe that 78,557 is the smallest Sierpinski number, but it hasn't yet been proven. In order to prove it, it has to be shown that every single k less than 78,557 is not a Sierpinski number, and to do that, some n must be found that makes k*2^n+1 prime.
The smallest proven 'prime' Sierpinski number is 271,129. In order to prove it, it has to be shown that every single 'prime' k less than 271,129 is not a Sierpinski number, and to do that, some n must be found that makes k*2^n+1 prime.
Should both of these problems be solved, k = 78557 will be established as the smallest Sierpinski number, and k = 271129 will be established as the smallest prime Sierpinski number. However, this would not prove that k = 271129 is the second provable Sierpinski number. Since the prime Sierpinski problem is testing all prime k's for 78557 < k < 271129, all that's needed is to test the composite k's for 78557 < k < 271129. Thus, the Extended Sierpinski Problem is established.
The following k's remain for each project:
Sierpinski problem (SoB) Prime Sierpinski problem (PSP) Extended Sierpinski Problem (ESP)
10223 10223* 91549
21181 22699* 99739
22699 67607* 131179
24737 79309 163187
55459 79817 193997
67607 152267 200749
156511 202705
168451 209611
222113 227723
225931 229673
237019 238411
'*being tested by Seventeen or Bust
Additional Information
For more information about Sierpinski, Sierpinski number, and the Sierpinsk problem, please see these resources:
What is LLR?
The Lucas-Lehmer-Riesel (LLR) test is a primality test for numbers of the form N = k*2^n − 1, with 2^n > k. Also, LLR is a program developed by Jean Penne that can run the LLR-tests. It includes the Proth test to perform +1 tests and PRP to test non base 2 numbers. See also:
(Edouard Lucas: 1842-1891, Derrick H. Lehmer: 1905-1991, Hans Riesel: 1929-2014).
____________
My lucky number is 75898524288+1
| |
|
Michael Goetz Volunteer moderator Project administrator
 Send message
Joined: 21 Jan 10 Posts: 13955 ID: 53948 Credit: 392,876,449 RAC: 184,453
                               
|
Two days until the challenge starts! Some thoughts about ESP:
Looks like my Haswell can do the ESP tasks in roughly 15 hours.
PSP has always had very large tasks, and SoB's tasks are positively huge. But ESP has always been the "small" Sierpinski tasks. Probably shouldn't think about it that way any more. The tests are closing in on 3 million digits and aren't that much smaller than the 321 tasks. ESP overtook TRP a while ago (TRP has a LOT more Ks and therefore progresses much more slowly.)
Unlike all those other projects, ESP is relatively new. It only started about 4 or 5 years ago.
____________
My lucky number is 75898524288+1 | |
|
|
Looks like my Haswell can do the ESP tasks in roughly 15 hours.
Would you like benchmarks posted once we start completing tasks? I expect my Skylakes to be around 9-10 hrs per unit. | |
|
RafaelVolunteer tester
 Send message
Joined: 22 Oct 14 Posts: 909 ID: 370496 Credit: 530,936,231 RAC: 418,923
                        
|
Looks like my Haswell can do the ESP tasks in roughly 15 hours.
Would you like benchmarks posted once we start completing tasks? I expect my Skylakes to be around 9-10 hrs per unit.
http://www.primegrid.com/results.php?hostid=504305
Most WU are just shy of 10h on my Skylake box. Keep in mind, tasks run slower the more WU you have running at once (though general throughput is still best with 4 at a time). | |
|
Michael Goetz Volunteer moderator Project administrator
 Send message
Joined: 21 Jan 10 Posts: 13955 ID: 53948 Credit: 392,876,449 RAC: 184,453
                               
|
Looks like my Haswell can do the ESP tasks in roughly 15 hours.
Would you like benchmarks posted once we start completing tasks? I expect my Skylakes to be around 9-10 hrs per unit.
It will be useful for people planning on entering the challenge. It certainly can't hurt.
____________
My lucky number is 75898524288+1 | |
|
|
Looks like my Haswell can do the ESP tasks in roughly 15 hours.
Would you like benchmarks posted once we start completing tasks? I expect my Skylakes to be around 9-10 hrs per unit.
http://www.primegrid.com/results.php?hostid=504305
Most WU are just shy of 10h on my Skylake box. Keep in mind, tasks run slower the more WU you have running at once (though general throughput is still best with 4 at a time).
I expect to have slightly better times due to the greater stock frequencies. We will see. | |
|
mackerel Volunteer tester
 Send message
Joined: 2 Oct 08 Posts: 2639 ID: 29980 Credit: 568,393,769 RAC: 1,834
                              
|
For Haswell and later quad core systems, ram speed is as important as CPU clock speed. My faster Skylake quad does 960k FFT units in around 9 hours. This particular system has dual rank dual channel 3200 ram. Another system that has a similar ram configuration but only 3000 speed is closer to 10 hours at that FFT size. Expect times to go up a lot if you're running slower ram. | |
|
|
For Haswell and later quad core systems, ram speed is as important as CPU clock speed. My faster Skylake quad does 960k FFT units in around 9 hours. This particular system has dual rank dual channel 3200 ram. Another system that has a similar ram configuration but only 3000 speed is closer to 10 hours at that FFT size. Expect times to go up a lot if you're running slower ram.
Any opinion on 3 vs 4 units at a time and throughput? I would expect faster ram to favor 4 units, but enough to make a difference? | |
|
|
I have dual channel dual rank but it it is rated at 2666 MHz and is limited by the mobo which supports up to 2133 MHz.
I am eager to get going for some benchmark results this Friday. | |
|
mackerel Volunteer tester
 Send message
Joined: 2 Oct 08 Posts: 2639 ID: 29980 Credit: 568,393,769 RAC: 1,834
                              
|
I can't claim to have tested every possible scenario, but in general you can look at the combined CPU demand compared against the available ram bandwidth. More CPU power will lead to more throughput, up to such a point where ram becomes the ultimate bottleneck. In such a scenario, going 3 cores instead of 4 might not cost you throughput, but you have an improvement in unit time.
As a very rough guide, for a quad core processor (Haswell+), having dual channel ram at a nominal marketing speed equal to the CPU clock would be practically ram unlimited, or processor limited. In my case I have CPU at 4200 and ram at 3200, and that is still good for a high level of efficiency. Certainly there is a major performance hit of tens of % if I run the same ram at standard 2133. | |
|
RafaelVolunteer tester
 Send message
Joined: 22 Oct 14 Posts: 909 ID: 370496 Credit: 530,936,231 RAC: 418,923
                        
|
For Haswell and later quad core systems, ram speed is as important as CPU clock speed. My faster Skylake quad does 960k FFT units in around 9 hours. This particular system has dual rank dual channel 3200 ram. Another system that has a similar ram configuration but only 3000 speed is closer to 10 hours at that FFT size. Expect times to go up a lot if you're running slower ram.
Any opinion on 3 vs 4 units at a time and throughput? I would expect faster ram to favor 4 units, but enough to make a difference?
Wanna know for yourself? Download Prime95 and benchmark yourself. It won't give you a time estimate, but at least it's very precise to measure throughput. Just download it and in the "prime.txt", put the following lines (which are the custom settings I'm using):
OnlyBench5678=0
MinBenchFFT=960
MaxBenchFFT=960
BenchTime=20
BenchHyperthreads=0
BenchMultithreads=0
Then, fire the app and Options -> Benchmark. It'll show you iter/s for a 1~however many cores load. You can use it to compare how much slower things become when adding more WUs and decide if it really is worth adding that extra core or not.
For reference, I do 351.8 iter/s with a single core. With 2 cores, I 697.9 iter/s, which is almost a perfect scaling; I'll blame the little loss on windows's process scheduling. Anyhow, 2 cores are totally worth it. Then, 3 cores: 973.7. And that's already a bad sign, that's only ~2.77 scaling where 3 was expected. Still, it's worth having that 3rd core. And then! 4 cores: 1033.6.
Yeah, that's right, 1033.6, aka 2.94 scaling where 4 (or at the very least, 3.something) was expected. I'll leave it up to you to decide if it's worth putting that extra core just for the sake of a tiny boost in throughput for the challenge, or if you should do something else with it (say, a sieve project or perhaps use if to feed GPUs).
Now, this is for MY machine. Compare it to yours to see how you do.
*Note: that "960" is the FTT size of most ESP tasks. Change it if you want benchmarks for another project | |
|
mackerel Volunteer tester
 Send message
Joined: 2 Oct 08 Posts: 2639 ID: 29980 Credit: 568,393,769 RAC: 1,834
                              
|
It won't give you a time estimate, but at least it's very precise to measure throughput.
I think you can work out the time estimate from the benchmark results. The number of iterations is the "n" of the unit being tested. Leading edge for ESP is about 9.46M.
So: 9.46M / (iters/s) * cores = estimated unit time in seconds
For the numbers given previously by Rafael, that works out to be 7.5, 7.5, 8.1, 10.2 hours for running 1, 2, 3, 4 units respectively. Does that correlate?
My concern here is that LLR times might be slightly different from Prime95 times. Also do note there will be variation between the Prime95 benchmark runs, so you might want to try multiple times and see. I believe there were changes in 28.9 which might make it run faster than 28.7, and 28.7 might more closely resemble how we run LLR. I haven't looked at this in depth. | |
|
|
For the numbers given previously by Rafael, that works out to be 7.5, 7.5, 8.1, 10.2 hours for running 1, 2, 3, 4 units respectively.
The above predictions suggest that running all four cores will maximize the number of units tested per day!
With three units running simultaneously and taking up 8.1 hrs as a set, one will complete (24 / 8.1) x 3 or 8.89 units per day.
Proceeding in a similar fashion but now with four units running simultaneously and taking 10.2 hrs as a set, we have (24 / 10.2) x 4 or 9.41 units per day.
In other words, the overall gain is about 0.5 units per day. | |
|
mackerel Volunteer tester
 Send message
Joined: 2 Oct 08 Posts: 2639 ID: 29980 Credit: 568,393,769 RAC: 1,834
                              
|
If you were running forever, yes. As a challenge has a fixed finishing time, you might need to micro-manage workunit lengths (buy running fewer cores) near the end. | |
|
|
If you were running forever, yes. As a challenge has a fixed finishing time, you might need to micro-manage workunit lengths (buy running fewer cores) near the end.
No question there is a juggling act at the end.
The difference I compute below translates to 2.5 units over the course of the five-day challenge for a single quadcore CPU and which will mushroom further with more cores in the mix from a single user. | |
|
RafaelVolunteer tester
 Send message
Joined: 22 Oct 14 Posts: 909 ID: 370496 Credit: 530,936,231 RAC: 418,923
                        
|
For the numbers given previously by Rafael, that works out to be 7.5, 7.5, 8.1, 10.2 hours for running 1, 2, 3, 4 units respectively. Does that correlate?
Pretty much. Running 4 at a time, I was getting 10h on point. Ignoring the leading edge thing (and the fact that some WU have a smaller FTT), I'd say it's a good enough estimate. | |
|
Michael Goetz Volunteer moderator Project administrator
 Send message
Joined: 21 Jan 10 Posts: 13955 ID: 53948 Credit: 392,876,449 RAC: 184,453
                               
|
I can pretty much guarantee that you will be seeing tasks with different FFT sizes and possibly significantly different run times during this challenge. Planning your task strategy down to the minute probably won't work very well since later tasks may take longer to run than you anticipate.
Since ESP doesn't have many Ks to search, n, and therefore size, increases fairly quickly.
____________
My lucky number is 75898524288+1 | |
|
mackerel Volunteer tester
 Send message
Joined: 2 Oct 08 Posts: 2639 ID: 29980 Credit: 568,393,769 RAC: 1,834
                              
|
Planning to the hour might be good enough. I've had past challenges where I knew I would be late by some hours for a last set if I ran all cores blazing, but reducing the number of active cores near the end would reduce it enough for a fighting chance.
Also... it remains to be seen just how much the leading edge can be pushed. Lower work (resends or whatever) isn't critical, since it will be shorter. Time taken should scale with the n value, at least until we hit the next FFT size. Maybe tasks at the end will be some low % longer than now. | |
|
yank  Send message
Joined: 14 May 07 Posts: 111 ID: 8367 Credit: 11,474,819,218 RAC: 0
                    
|
ESP-LLR work units:
If I remember it correctly one has to turn off the HT and select only two cores to have good results working this work-unit type.
What percent of the CPU should be set?
____________
| |
|
RafaelVolunteer tester
 Send message
Joined: 22 Oct 14 Posts: 909 ID: 370496 Credit: 530,936,231 RAC: 418,923
                        
|
ESP-LLR work units:
If I remember it correctly one has to turn off the HT and select only two cores to to have good results working this work-unit type.
What percent of the CPU should be set?
HT is bad and should be turned off; that part is pretty accurate.
The 2 cores thing, however, is not. If you'd look at my example, you'd see that having 3 cores (out of my 4 available) would strike the best balance between throughput and core usage. In my case, that would be 75% CPU usage.
But your results may vary. You machine might have better (or worse) scaling than mine. Benchmark it yourself so you can get a good grasp as to how to better set your PC. | |
|
yank  Send message
Joined: 14 May 07 Posts: 111 ID: 8367 Credit: 11,474,819,218 RAC: 0
                    
|
So 75% of the computer cores would be a good starting point? | |
|
RafaelVolunteer tester
 Send message
Joined: 22 Oct 14 Posts: 909 ID: 370496 Credit: 530,936,231 RAC: 418,923
                        
|
So 75% of the computer cores would be a good starting point?
Again, depends on your PC. 75% is good for mine because I have very fast RAM; if I had potato RAM, 50% would have been better. If I had a 6 core CPU, maybe 67% would be smarter. Run the benchmark (instructions are in an earlier post) and you can easily figure what what works best for you. | |
|
|
I am going to run 100% all cores to the wall.
Sure glad I stopped in here, I have been doing a couple trial runs and I have 2 systems pretty much the same and one was taking a couple hours longer today than the other system. I thought something was wrong.
Finding out there are different size WUs , is one less thing to figure out.
Last couple days I had a bad stick of memory and one GPU that has to go for a RMA.
GPU 7 day before the warranty runs out!!
Good luck everyone, hope someone finds a prime or two. | |
|
|
e5-1650 Sandybridge hexcores (i7-39300K) around 15-18 hours at 4.3ghz
x5660 Westmere (I7-970) no AVX, around 33 hours at 3.9ghz
Not sure what my e5-2670 will do, some where in-between those two above.
I'll guess 25 hours. | |
|
Dave  Send message
Joined: 13 Feb 12 Posts: 3171 ID: 130544 Credit: 2,233,831,976 RAC: 576,376
                           
|
At the other end of the scale anyone know if a Celeron N2840 could squeeze a unit or 2 in 432000 secs? | |
|
mackerel Volunteer tester
 Send message
Joined: 2 Oct 08 Posts: 2639 ID: 29980 Credit: 568,393,769 RAC: 1,834
                              
|
Assuming it has the same IPC as first gen Core i3/5/7 CPUs which also didn't AVX, ball park estimate would be 60 hours a unit.
I would suggest trying it regardless, and after letting it run an hour or so see what the estimated time says. | |
|
|
At the other end of the scale anyone know if a Celeron N2840 could squeeze a unit or 2 in 432000 secs?
I am tempted to throw into the mix my two Dell laptops with a duo core 2.67GHz P9600 chip .... they did just fine with the ESP and TRP Sieves and I would not mind seeing one or two units complete from each with ESP LLR.
Edit: http://ark.intel.com/products/37266/Intel-Core2-Duo-Processor-P9600-6M-Cache-2_66-GHz-1066-MHz-FSB | |
|
Michael Goetz Volunteer moderator Project administrator
 Send message
Joined: 21 Jan 10 Posts: 13955 ID: 53948 Credit: 392,876,449 RAC: 184,453
                               
|
I am tempted to throw into the mix my two Dell laptops with a duo core 2.67GHz P9600 chip .... they did just fine with the ESP and TRP Sieves and I would not mind seeing one or two units complete from each with ESP LLR.
This would be an excellent time to remind anyone contemplating using a laptop with an Intel AVX-capable CPU (i.e., the laptops that will perform best in this challenge) for LLR processing that laptops, even so-called "gaming" laptops, were not designed with this kind of computation in mind and may have difficulty dealing with the heat produced. This is especially true if the laptop isn't brand new and therefore likely has dust insulating the cooling surfaces and/or less than perfect thermal paste.
In the best case scenario the computer will simply slow down rather than let the temperatures get too high. Worst case is that the extended run time at high temperatures causes some component to fail.
Personally, I never use a mobile device for crunching, unless I'm willing to accept that it may shorten the life of the device. Since I'm not willing to accept a dead computer as the price for a few credits, I essentially never crunch on a laptop. And certainly not with any LLR tasks.
If you are going to use a laptop for this challenge, please make sure it has a lot of cooling. I recommend a good laptop cooling stand that blows air into the vents on the bottom of the laptop as well as a powerful desk fan blowing air across the top of the keyboard. Never, ever crunch with the laptop closed -- this blocks convection from 50% of the laptop's surface area..
____________
My lucky number is 75898524288+1 | |
|
|
Thanks, Michael. | |
|
|
How to join? Or do i join automatically? Automatically i mean that all computed wu's after challenge started are part of Summer Paralympics challenge? | |
|
Michael Goetz Volunteer moderator Project administrator
 Send message
Joined: 21 Jan 10 Posts: 13955 ID: 53948 Credit: 392,876,449 RAC: 184,453
                               
|
How to join? Or do i join automatically? Automatically i mean that all computed wu's after challenge started are part of Summer Paralympics challenge?
There is no explicit "joining" required. Any ESP-LLR tasks that were both sent to your computer after the start of the challenge and also returned back to the server before the challenge ends will count towards your challenge score.
Tasks sent to you BEFORE the challenge starts will NOT count.
Tasks returned AFTER the challenge ends will NOT count.
____________
My lucky number is 75898524288+1 | |
|
|
How to join? Or do i join automatically? Automatically i mean that all computed wu's after challenge started are part of Summer Paralympics challenge?
There is no explicit "joining" required. Any ESP-LLR tasks that were both sent to your computer after the start of the challenge and also returned back to the server before the challenge ends will count towards your challenge score.
Tasks sent to you BEFORE the challenge starts will NOT count.
Tasks returned AFTER the challenge ends will NOT count.
Thank you. | |
|
|
Minutes to go!!
____________
| |
|
Michael Goetz Volunteer moderator Project administrator
 Send message
Joined: 21 Jan 10 Posts: 13955 ID: 53948 Credit: 392,876,449 RAC: 184,453
                               
|
If anyone along the Eastern seaboard of the United States sees what appears to be a huge dust storm heading their way, that is NOT Hurricane Ermine. I was blowing the dust out of my computer. Carry on...
____________
My lucky number is 75898524288+1 | |
|
RafaelVolunteer tester
 Send message
Joined: 22 Oct 14 Posts: 909 ID: 370496 Credit: 530,936,231 RAC: 418,923
                        
|
Let's go folks, time to knock a few K out of this conjecture. | |
|
Sysadm@Nbg Volunteer moderator Volunteer tester Project scientist
 Send message
Joined: 5 Feb 08 Posts: 1217 ID: 18646 Credit: 859,513,949 RAC: 202,534
                      
|
and here we go ...
____________
Sysadm@Nbg
my current lucky number: 113856050^65536 + 1
PSA-PRPNet-Stats-URL: http://u-g-f.de/PRPNet/
| |
|
mackerel Volunteer tester
 Send message
Joined: 2 Oct 08 Posts: 2639 ID: 29980 Credit: 568,393,769 RAC: 1,834
                              
|
Go Go Go! I'm still finishing off a build to add but the running systems are away. | |
|
|
Nice. My new system ist sitting naked on the desktop and crunching away - didn't get the case shipped in time :-) | |
|
Michael Goetz Volunteer moderator Project administrator
 Send message
Joined: 21 Jan 10 Posts: 13955 ID: 53948 Credit: 392,876,449 RAC: 184,453
                               
|
1708 tasks sent out in the first 5 minutes.
____________
My lucky number is 75898524288+1 | |
|
Dave  Send message
Joined: 13 Feb 12 Posts: 3171 ID: 130544 Credit: 2,233,831,976 RAC: 576,376
                           
|
Should have been 1709 - just 1 remote PC that will be about half an hour late to the party. Athlon64 topping out @ 80C but more cool flowing through the house now. | |
|
|
For the numbers given previously by Rafael, that works out to be 7.5, 7.5, 8.1, 10.2 hours for running 1, 2, 3, 4 units respectively.
The above predictions suggest that running all four cores will maximize the number of units tested per day!
With three units running simultaneously and taking up 8.1 hrs as a set, one will complete (24 / 8.1) x 3 or 8.89 units per day.
Proceeding in a similar fashion but now with four units running simultaneously and taking 10.2 hrs as a set, we have (24 / 10.2) x 4 or 9.41 units per day.
In other words, the overall gain is about 0.5 units per day.
Something very strange is going on with about 45 minutes worth of processing.
The i3 6100 (3.7 GHz) and i3 6300 (3.8 GHz) are showing about a 10-hr runtime per unit per core (both physical cores running) and which is what I expected before the start of the challenge.
The i7 6700k (4.0 GHz) is showing about a 20-hr runtime per unit per core with all four physical cores running.
My i7 3630QM (3.2 GHz) is showing around 18.5 hrs runtime per unit per core with three of the four physical cores running (the other physical core is dedicated to PPR12M and Collatz).
No HT with CPU usage set at 50% for all of the Skylake systems.
Does not compute (excuse the pun) with the performance of the i7 6700k. | |
|
Michael Goetz Volunteer moderator Project administrator
 Send message
Joined: 21 Jan 10 Posts: 13955 ID: 53948 Credit: 392,876,449 RAC: 184,453
                               
|
About 7300 tasks sent out in the first hour.
About 16-18 hours running 4 cores on a stock speed Haswell i5-4670K.
____________
My lucky number is 75898524288+1 | |
|
mackerel Volunteer tester
 Send message
Joined: 2 Oct 08 Posts: 2639 ID: 29980 Credit: 568,393,769 RAC: 1,834
                              
|
The i7 6700k (4.0 GHz) is showing about a 20-hr runtime per unit per core with all four physical cores running.
No HT with CPU usage set at 50% for all of the Skylake systems.
I don't understand the last line if you can clarify? I would also try using CPU-z to check the clocks of CPU and ram are where you expect them to be, and there are no other programs running and stealing CPU cycles.
My Skylake systems are more or less within expectations:
4.2 GHz core, 3200/2R ram, 9 hour estimate
4.2 GHz core, 3000/2R ram, 10 hour estimate
4.0 GHz core, 2666/1R ram, 12 hour estimate
| |
|
Michael Goetz Volunteer moderator Project administrator
 Send message
Joined: 21 Jan 10 Posts: 13955 ID: 53948 Credit: 392,876,449 RAC: 184,453
                               
|
Does not compute (excuse the pun) with the performance of the i7 6700k.
I'd make sure (with Task Manager if on Windows or the equivalent elsewhere) that you are in fact only running 4 tasks. It sounds like you have other tasks running (maybe not ESP, or even PrimeGrid, or even BOINC). Look for something else that's sucking up CPU.
____________
My lucky number is 75898524288+1 | |
|
|
Does not compute (excuse the pun) with the performance of the i7 6700k.
I'd make sure (with Task Manager if on Windows or the equivalent elsewhere) that you are in fact only running 4 tasks. It sounds like you have other tasks running (maybe not ESP, or even PrimeGrid, or even BOINC). Look for something else that's sucking up CPU.
I rebooted as a precautionary measure but no luck. According to the task manager, I am using only about 55% of the CPU but I have similar usage with my i3 systems which are behaving according to expectations.
Edit: According to http://www.primegrid.com/results.php?hostid=520564&offset=0&show_names=0&state=1&appid=, I have a TRP Sieve unit lurking around and which is not showing up in the BOINC Manager (I made sure all such tasks completed before starting on the challenge).
Also, can the variable FFT account for a factor of 2 in the runtimes? | |
|
|
The i7 6700k (4.0 GHz) is showing about a 20-hr runtime per unit per core with all four physical cores running.
No HT with CPU usage set at 50% for all of the Skylake systems.
I don't understand the last line if you can clarify? I would also try using CPU-z to check the clocks of CPU and ram are where you expect them to be, and there are no other programs running and stealing CPU cycles.
My Skylake systems are more or less within expectations:
4.2 GHz core, 3200/2R ram, 9 hour estimate
4.2 GHz core, 3000/2R ram, 10 hour estimate
4.0 GHz core, 2666/1R ram, 12 hour estimate
I am referring to the BOINC Manager and where I have set CPU usage to 50% so as to effectively turn off HT. The times you list are what I would have expected (somewhere around 10-ish). | |
|
Michael Goetz Volunteer moderator Project administrator
 Send message
Joined: 21 Jan 10 Posts: 13955 ID: 53948 Credit: 392,876,449 RAC: 184,453
                               
|
Also, can the variable FFT account for a factor of 2 in the runtimes?
No. Of all the ESP tasks sent out during the challenge, the difference between the shortest and the longest tasks is about 12%.
____________
My lucky number is 75898524288+1 | |
|
|
Tho i doubt any results are in yet, were will they get posted? , isnt there a graph/chart/table etc with the results? Sorry if ive missed the post | |
|
|
http://www.primegrid.com/challenge/2016_6/top_users.html
I wonder of a who delivery first unit completed and validated. If the rules are working so it must be in the current fastest cpu, computer or some system..most suitable for calculating primegrid.. or no..?:-) | |
|
mackerel Volunteer tester
 Send message
Joined: 2 Oct 08 Posts: 2639 ID: 29980 Credit: 568,393,769 RAC: 1,834
                              
|
I am referring to the BOINC Manager and where I have set CPU usage to 50% so as to effectively turn off HT. The times you list are what I would have expected (somewhere around 10-ish).
What confused me was you said no HT earlier. It is on or off regardless if you use all threads.
So, 6700k HT on, running 4 tasks? Still look in task manager. How do the 8 CPU blocks look? Typical windows behaviour would have all 8 being partially used. In my experience Windows doesn't spread them perfectly, and you will get around 10% reduction in performance compared to HT off, or HT on with manual affinity set. With manual affinity set to either odd or even numbered threads, you can recover that performance. Best is still to turn off HT. Affinity is a bit of a pain to set and still not 100% reliable as I've had cases where two threads end up on one core with another core idle... | |
|
mackerel Volunteer tester
 Send message
Joined: 2 Oct 08 Posts: 2639 ID: 29980 Credit: 568,393,769 RAC: 1,834
                              
|
I wonder of a who delivery first unit completed and validated. If the rules are working so it must be in the current fastest cpu, computer or some system..most suitable for calculating primegrid.. or no..?:-)
On recent past long unit LLR challenges I've done pretty well, usually appearing within the first or 2nd updates. Trick is you need a fast CPU *and* fast ram to go with it. Note optimising a system for most units in the challenge is different from fastest to return. Also by luck if you get a "fast" unit that could shave a lot of time off too. | |
|
Michael Goetz Volunteer moderator Project administrator
 Send message
Joined: 21 Jan 10 Posts: 13955 ID: 53948 Credit: 392,876,449 RAC: 184,453
                               
|
http://www.primegrid.com/challenge/2016_6/top_users.html
I wonder of a who delivery first unit completed and validated. If the rules are working so it must be in the current fastest cpu, computer or some system..most suitable for calculating primegrid.. or no..?:-)
The fastest completed ESP task in the database was done by Rafael's Skylake i5-6600K in just under 8 hours.
There's also a number of tasks completed in less than 9 hours. Of the top 25 fastest completed ESP tasks, all were done by Skylake i5-6600K and i7-6700K CPUs.
(The fastest AMD CPU comes nowhere close: just under 29 hours on an FX-8350.)
____________
My lucky number is 75898524288+1 | |
|
|
I am referring to the BOINC Manager and where I have set CPU usage to 50% so as to effectively turn off HT. The times you list are what I would have expected (somewhere around 10-ish).
What confused me was you said no HT earlier. It is on or off regardless if you use all threads.
So, 6700k HT on, running 4 tasks? Still look in task manager. How do the 8 CPU blocks look? Typical windows behaviour would have all 8 being partially used. In my experience Windows doesn't spread them perfectly, and you will get around 10% reduction in performance compared to HT off, or HT on with manual affinity set. With manual affinity set to either odd or even numbered threads, you can recover that performance. Best is still to turn off HT. Affinity is a bit of a pain to set and still not 100% reliable as I've had cases where two threads end up on one core with another core idle...
I have gone through the same steps between my two i3 Skylakes and single i7 Skylake and, as a result, I would have expected similar behaviour between the three systems (ie all around the expected runtimes or all of them way off). I will wait the 20 hrs for them to complete and will then see how the next set of four behave (not that I expect any difference). If I get similar behaviour with the new set, I will abort them and have the machine return to TRP Sieve.
I also installed Process Lasso (as an afterthought) which looks after the affinity but no difference.
Edit: The machine has 2x8Gb 2400 MHz RAM which is dual rank and dual channel. On other projects (ex AP27 and TRP Sieve), this system consistently outperformed the i3 6100 and i3 6300 (as expected due to the greater stock speed). | |
|
RafaelVolunteer tester
 Send message
Joined: 22 Oct 14 Posts: 909 ID: 370496 Credit: 530,936,231 RAC: 418,923
                        
|
The fastest completed ESP task in the database was done by Rafael's Skylake i5-6600K in just under 8 hours.
There's also a number of tasks completed in less than 9 hours. Of the top 25 fastest completed ESP tasks, all were done by Skylake i5-6600K and i7-6700K CPUs.
O.O
I feel so important right now....
Still, that's a bit surprising. While 4263mhz core/cache with 3146mhz dual channel dual rank RAM is blazing fast, I would expect other systems (say, 3200mhz 4.3 owners with 8mb of cache) to be even faster, even if marginally. I guess running 3 WU at a time rather than 4 really speeds things up.... | |
|
Michael Goetz Volunteer moderator Project administrator
 Send message
Joined: 21 Jan 10 Posts: 13955 ID: 53948 Credit: 392,876,449 RAC: 184,453
                               
|
The fastest completed ESP task in the database was done by Rafael's Skylake i5-6600K in just under 8 hours.
There's also a number of tasks completed in less than 9 hours. Of the top 25 fastest completed ESP tasks, all were done by Skylake i5-6600K and i7-6700K CPUs.
O.O
I feel so important right now....
Still, that's a bit surprising. While 4263mhz core/cache with 3146mhz dual channel dual rank RAM is blazing fast, I would expect other systems (say, 3200mhz 4.3 owners with 8mb of cache) to be even faster, even if marginally. I guess running 3 WU at a time rather than 4 really speeds things up....
Typically, the first task to come in on a challenge is from someone that not only has a fast computer, but also is running very few tasks -- sometimes just one task. Of course, for a challenge, it's not getting the fastest task, or the first task. It's doing the most tasks.
____________
My lucky number is 75898524288+1 | |
|
mackerel Volunteer tester
 Send message
Joined: 2 Oct 08 Posts: 2639 ID: 29980 Credit: 568,393,769 RAC: 1,834
                              
|
Still, that's a bit surprising. While 4263mhz core/cache with 3146mhz dual channel dual rank RAM is blazing fast, I would expect other systems (say, 3200mhz 4.3 owners with 8mb of cache) to be even faster, even if marginally. I guess running 3 WU at a time rather than 4 really speeds things up....
As the numbers earlier showed, running 4 cores still gives a benefit to throughput, but individual unit times will be longer than 3 or fewer.
From the earlier data another half hour or so could be shaved off by going down to two cores. To go faster still, going single core will make the ram effectively unlimited at that point. Then concentrate on core overclocking. Past testing didn't show any noticeably benefit from 6MB or 8MB cache for multiple big units, but it might just make a difference here. A single unit could fit in 8MB of L3 cache and get more of a boost from that. I might have to try this after the challenge...
I think an overclocked Haswell-E system could be competitive also, if not too many cores are used. | |
|
|
As the numbers earlier showed, running 4 cores still gives a benefit to throughput, but individual unit times will be longer than 3 or fewer.
Is running fewer tasks significant on far slower machines? I currently have a quad core N3530 running 4 tasks. A little extrapolation shows that they will take about 125 hours to finish. If I reduce this to 3 in order to speed things up, should I abort the 4th task rather than suspend it in order to clear it from memory? | |
|
|
Edit: The machine has 2x8Gb 2400 MHz RAM which is dual rank and dual channel. On other projects (ex AP27 and TRP Sieve), this system consistently outperformed the i3 6100 and i3 6300 (as expected due to the greater stock speed).
Problem solved: according to CPU-Z, the RAM is single channel. Someone is going to get a mouthful (dealer for the RAM). | |
|
Michael Goetz Volunteer moderator Project administrator
 Send message
Joined: 21 Jan 10 Posts: 13955 ID: 53948 Credit: 392,876,449 RAC: 184,453
                               
|
As the numbers earlier showed, running 4 cores still gives a benefit to throughput, but individual unit times will be longer than 3 or fewer.
Is running fewer tasks significant on far slower machines? I currently have a quad core N3530 running 4 tasks. A little extrapolation shows that they will take about 125 hours to finish. If I reduce this to 3 in order to speed things up, should I abort the 4th task rather than suspend it in order to clear it from memory?
There's so many variables that affect performance, it's impossible to accurately predict what a very different type of computer will do. So the only way to know how your system will run 3 vs 4 tasks is to try it.
As for the other question, suspending a task is sufficient. You do not need to abort it. If it's suspended it won't be affecting the speed of the tasks that are running.
____________
My lucky number is 75898524288+1 | |
|
|
Thanks, Michael. 6 hours in, 4.8% complete. I'll see what the next 6 hours produce. | |
|
|
Edit: The machine has 2x8Gb 2400 MHz RAM which is dual rank and dual channel. On other projects (ex AP27 and TRP Sieve), this system consistently outperformed the i3 6100 and i3 6300 (as expected due to the greater stock speed).
Problem solved: according to CPU-Z, the RAM is single channel. Someone is going to get a mouthful (dealer for the RAM).
Hmmm. How many dimm slots and how many are populated? That shouldn't be a function of the memory, it should be the motherboard. And I'm not aware of any motherboard that supports a 6700k that doesn't support dual-channel memory. | |
|
|
Edit: The machine has 2x8Gb 2400 MHz RAM which is dual rank and dual channel. On other projects (ex AP27 and TRP Sieve), this system consistently outperformed the i3 6100 and i3 6300 (as expected due to the greater stock speed).
Problem solved: according to CPU-Z, the RAM is single channel. Someone is going to get a mouthful (dealer for the RAM).
Hmmm. How many dimm slots and how many are populated? That shouldn't be a function of the memory, it should be the motherboard. And I'm not aware of any motherboard that supports a 6700k that doesn't support dual-channel memory.
There are four slots with the proper two being populated. I know for certain the motherboard (MSI Z170A SLI PLUS) supports dual channel DDR4 memory. The sticks I received are not what I ordered (dual channel vs single channel, respectively). I will check tomorrow the product code on the RAM packaging for further confirmation before ripping into the dealer.
I used CPU-Z on my i3 systems and it does report dual channel. I could exchange sticks so as to get the extra processing power of the i7 6700k but I would rather leave things as they are right now. My mistake for not checking the RAM upon receipt. | |
|
|
[Edit: Sorry, getting in the middle of an important discussion here.]
Like Van Z, I'm surprised that you can have two sticks in there, in the proper slots, and have it function as single-channel. Like Van Z (again), I thought the mobo should sort that out if everything else is in order.
Separately: Let's try to hope that someone knocks a k off during the challenge. In GFN-land, the tradition is for it to happen about 12 hours before (well, not a k... but you know what I mean, finding a GFN prime), but we don't seem to have any such luck with other projects. | |
|
|
Do you mean dual rank?
If you have 2 dimm slots populated and it is reporting a single channel, that might be location-related.
Are they in slots 2 and 4?
https://us.msi.com/Motherboard/support/Z170A-SLI-PLUS.html#down-manual
I used CPU-Z on my i3 systems and it does report dual channel. I could exchange sticks so as to get the extra processing power of the i7 6700k but I would rather leave things as they are right now. My mistake for not checking the RAM upon receipt.
| |
|
Michael Goetz Volunteer moderator Project administrator
 Send message
Joined: 21 Jan 10 Posts: 13955 ID: 53948 Credit: 392,876,449 RAC: 184,453
                               
|
It looks like the next time the leaderboards are generated (at 02:30 UTC) there will be at least 3 results that have been returned.
____________
My lucky number is 75898524288+1 | |
|
Michael Goetz Volunteer moderator Project administrator
 Send message
Joined: 21 Jan 10 Posts: 13955 ID: 53948 Credit: 392,876,449 RAC: 184,453
                               
|
As of right now, three results have been returned, from two computers. Both are Haswell i7-6700K CPUs.
The first was returned by Steve of team Aggie the Pew, at 02:21:14.
23 seconds later DeleteNull of team SETI.Germany returned the second result, and 112 seconds after that he returned the third result.
Congrats on returning the first results! For what it's worth, my Haswell's tasks are at 50% right now.
____________
My lucky number is 75898524288+1 | |
|
|
Zimmer,
I cannot thank you enough! It was indeed a case of a misplaced stick (slots 4 and 3). I just did the change to slots 4 and 2 and CPU-Z is now reporting dual channel and the newly downloaded units are showing the expected runtime during the first five minutes of execution.
I owe you a beer!!!
Do you mean dual rank?
If you have 2 dimm slots populated and it is reporting a single channel, that might be location-related.
Are they in slots 2 and 4?
https://us.msi.com/Motherboard/support/Z170A-SLI-PLUS.html#down-manual
I used CPU-Z on my i3 systems and it does report dual channel. I could exchange sticks so as to get the extra processing power of the i7 6700k but I would rather leave things as they are right now. My mistake for not checking the RAM upon receipt.
| |
|
|
I have two e5-2683 V4 cpus with 16 cores each, 40mb cache, no HT. And only two wu are way ahead of the other wu. Very dissapointed with this performance. Single channel memory, it's a HP DL360 Gen8 server.
____________
| |
|
|
I like reading about all the speedy computers out there and what you're doing to make them so fast. I have two rather slower AMD based machines under Linux Mint and have been doing some case mods prior to this latest challenge, and am pleased that the A8 quad-core box with a new hole at the top and extra fan in it is now looking like completing these WUs in 54 hours, so I'm hoping to get 8 through from there by the end of the challenge period. My other is an AMD Athlon dual-core which I've installed a huge fan inside and some ducting so it blows directly on the RAM and CPU cooler fins, plus an extra exhaust fan too, and that's looking like it'll complete in 105 hours, so should also get two WUs through by the end, just. It might not sound much, but I'm a very happy bunny! :-) | |
|
mackerel Volunteer tester
 Send message
Joined: 2 Oct 08 Posts: 2639 ID: 29980 Credit: 568,393,769 RAC: 1,834
                              
|
I have two e5-2683 V4 cpus with 16 cores each, 40mb cache, no HT. And only two wu are way ahead of the other wu. Very dissapointed with this performance. Single channel memory, it's a HP DL360 Gen8 server.
Am I reading that right, two E5-2683 V4 in one system, no HT, for 32 cores. And a single memory channel? Is that per socket or total? Either way, it is still going to cripple performance. For big LLR tasks, get quad channel per socket.
I'm running an ES E5-2583 V3 with quad channel 2133, and that's still going to be memory bandwidth limited. It is a bit annoying the Xeon ram speed is locked so I can't use faster ram even with X99 chipset. If it helps for comparison, with the CPU at 2.3 GHz all cores running, I'm expecting to do 14 units every 28 hours or so as the system is currently configured. | |
|
|
I have two e5-2683 V4 cpus with 16 cores each, 40mb cache, no HT. And only two wu are way ahead of the other wu. Very dissapointed with this performance. Single channel memory, it's a HP DL360 Gen8 server.
Am I reading that right, two E5-2683 V4 in one system, no HT, for 32 cores. And a single memory channel? Is that per socket or total? Either way, it is still going to cripple performance. For big LLR tasks, get quad channel per socket.
I'm running an ES E5-2583 V3 with quad channel 2133, and that's still going to be memory bandwidth limited. It is a bit annoying the Xeon ram speed is locked so I can't use faster ram even with X99 chipset. If it helps for comparison, with the CPU at 2.3 GHz all cores running, I'm expecting to do 14 units every 28 hours or so as the system is currently configured.
16 per socket. It only has one stick of ram in one socket. I think that might be the problem. Each cpu should have memory per socket.
CPU-Z TXT Report
-------------------------------------------------------------------------
Binaries
-------------------------------------------------------------------------
CPU-Z version 1.77.0.x64
Processors
-------------------------------------------------------------------------
Number of processors 2
Number of threads 32
APICs
-------------------------------------------------------------------------
Processor 0
-- Core 0
-- Thread 0 0
-- Core 1
-- Thread 0 2
-- Core 2
-- Thread 0 4
-- Core 3
-- Thread 0 6
-- Core 4
-- Thread 0 8
-- Core 5
-- Thread 0 10
-- Core 6
-- Thread 0 12
-- Core 7
-- Thread 0 14
-- Core 8
-- Thread 0 16
-- Core 9
-- Thread 0 18
-- Core 10
-- Thread 0 20
-- Core 11
-- Thread 0 22
-- Core 12
-- Thread 0 24
-- Core 13
-- Thread 0 26
-- Core 14
-- Thread 0 28
-- Core 15
-- Thread 0 30
Processor 1
-- Core 0
-- Thread 0 32
-- Core 1
-- Thread 0 34
-- Core 2
-- Thread 0 36
-- Core 3
-- Thread 0 38
-- Core 4
-- Thread 0 40
-- Core 5
-- Thread 0 42
-- Core 6
-- Thread 0 44
-- Core 7
-- Thread 0 46
-- Core 8
-- Thread 0 48
-- Core 9
-- Thread 0 50
-- Core 10
-- Thread 0 52
-- Core 11
-- Thread 0 54
-- Core 12
-- Thread 0 56
-- Core 13
-- Thread 0 58
-- Core 14
-- Thread 0 60
-- Core 15
-- Thread 0 62
Timers
-------------------------------------------------------------------------
ACPI timer 3.580 MHz
HPET timer 14.318 MHz
Perf timer 2.048 MHz
Sys timer 1.000 KHz
Processors Information
-------------------------------------------------------------------------
Processor 1 ID = 0
Number of cores 16 (max 16)
Number of threads 16 (max 32)
Name Intel Xeon v4
Codename Broadwell-E/EP
Specification Intel(R) Xeon(R) CPU E5-2683 v4 @ 2.10GHz
Package (platform ID) Socket 2011 LGA (0x0)
CPUID 6.F.1
Extended CPUID 6.4F
Core Stepping M0
Technology 14 nm
TDP Limit 120.0 Watts
Tjmax 102.0 °C
Core Speed 2496.7 MHz
Multiplier x Bus Speed 25.0 x 99.9 MHz
Stock frequency 2100 MHz
Instructions sets MMX, SSE, SSE2, SSE3, SSSE3, SSE4.1, SSE4.2, EM64T, VT-x, AES, AVX, AVX2, FMA3, TSX
L1 Data cache 16 x 32 KBytes, 8-way set associative, 64-byte line size
L1 Instruction cache 16 x 32 KBytes, 8-way set associative, 64-byte line size
L2 cache 16 x 256 KBytes, 8-way set associative, 64-byte line size
L3 cache 40 MBytes, 20-way set associative, 64-byte line size
FID/VID Control yes
Turbo Mode supported, enabled
Max non-turbo ratio 21x
Max turbo ratio 30x
Max efficiency ratio 12x
Max Power 240 Watts
Min Power 64 Watts
O/C bins none
Ratio 1 core 30x
Ratio 2 cores 30x
Ratio 3 cores 28x
Ratio 4 cores 27x
Ratio 5 cores 26x
Ratio 6 cores 26x
Ratio 7 cores 26x
Ratio 8 cores 26x
Ratio 9 cores 26x
Ratio 10 cores 26x
Ratio 11 cores 26x
Ratio 12 cores 26x
Ratio 13 cores 26x
Ratio 14 cores 26x
Ratio 15 cores 26x
Ratio 16 cores 26x
Ratio 17 cores 26x
Ratio 18 cores 26x
Ratio 19 cores 26x
Ratio 20 cores 26x
Ratio 21 cores 26x
Ratio 22 cores 26x
Ratio 23 cores 26x
Ratio 24 cores 26x
TSC 2098.1 MHz
APERF 2596.6 MHz
MPERF 2097.9 MHz
IA Voltage Mode PCU adaptive
IA Voltage Offset 0 mV
GT Voltage Mode PCU adaptive
GT Voltage Offset 0 mV
LLC/Ring Voltage Mode PCU adaptive
LLC/Ring Voltage Offset 0 mV
Agent Voltage Mode PCU adaptive
Agent Voltage Offset 0 mV
Temperature 0 89 degC (192 degF) (Core #0)
Temperature 1 89 degC (192 degF) (Core #1)
Temperature 2 88 degC (190 degF) (Core #2)
Temperature 3 90 degC (194 degF) (Core #3)
Temperature 4 90 degC (194 degF) (Core #4)
Temperature 5 89 degC (192 degF) (Core #5)
Temperature 6 89 degC (192 degF) (Core #6)
Temperature 7 88 degC (190 degF) (Core #7)
Temperature 8 88 degC (190 degF) (Core #8)
Temperature 9 89 degC (192 degF) (Core #9)
Temperature 10 88 degC (190 degF) (Core #10)
Temperature 11 90 degC (194 degF) (Core #11)
Temperature 12 89 degC (192 degF) (Core #12)
Temperature 13 87 degC (188 degF) (Core #13)
Temperature 14 89 degC (192 degF) (Core #14)
Temperature 15 87 degC (188 degF) (Core #15)
Temperature 16 92 degC (197 degF) (Package)
Voltage 0 0.90 Volts (VID)
Voltage 1 +0.00 Volts (IA Offset)
Voltage 2 +0.00 Volts (GT Offset)
Voltage 3 +0.00 Volts (LLC/Ring Offset)
Voltage 4 +0.00 Volts (System Agent Offset)
Power 0 118.99 W (Package)
Power 1 n.a. (IA Cores)
Power 2 n.a. (Uncore)
Power 3 16.91 W (DRAM)
Clock Speed 0 2496.73 MHz (Core #0)
Clock Speed 1 2496.73 MHz (Core #1)
Clock Speed 2 2496.73 MHz (Core #2)
Clock Speed 3 2496.73 MHz (Core #3)
Clock Speed 4 2496.73 MHz (Core #4)
Clock Speed 5 2496.73 MHz (Core #5)
Clock Speed 6 2496.73 MHz (Core #6)
Clock Speed 7 2496.73 MHz (Core #7)
Clock Speed 8 2496.73 MHz (Core #8)
Clock Speed 9 2496.73 MHz (Core #9)
Clock Speed 10 2496.73 MHz (Core #10)
Clock Speed 11 2496.73 MHz (Core #11)
Clock Speed 12 2496.73 MHz (Core #12)
Clock Speed 13 2496.73 MHz (Core #13)
Clock Speed 14 2496.73 MHz (Core #14)
Clock Speed 15 2496.73 MHz (Core #15)
Processor 2 ID = 1
Number of cores 16 (max 16)
Number of threads 16 (max 32)
Name Intel Xeon v4
Codename Broadwell-E/EP
Specification Intel(R) Xeon(R) CPU E5-2683 v4 @ 2.10GHz
Package (platform ID) Socket 2011 LGA (0x0)
CPUID 6.F.1
Extended CPUID 6.4F
Core Stepping M0
Technology 14 nm
TDP Limit 120.0 Watts
Tjmax 102.0 °C
Core Speed 2596.6 MHz
Multiplier x Bus Speed 26.0 x 99.9 MHz
Stock frequency 2100 MHz
Instructions sets MMX, SSE, SSE2, SSE3, SSSE3, SSE4.1, SSE4.2, EM64T, VT-x, AES, AVX, AVX2, FMA3, TSX
L1 Data cache 16 x 32 KBytes, 8-way set associative, 64-byte line size
L1 Instruction cache 16 x 32 KBytes, 8-way set associative, 64-byte line size
L2 cache 16 x 256 KBytes, 8-way set associative, 64-byte line size
L3 cache 40 MBytes, 20-way set associative, 64-byte line size
FID/VID Control yes
Turbo Mode supported, enabled
Max non-turbo ratio 21x
Max turbo ratio 30x
Max efficiency ratio 12x
Max Power 240 Watts
Min Power 64 Watts
O/C bins none
Ratio 1 core 30x
Ratio 2 cores 30x
Ratio 3 cores 28x
Ratio 4 cores 27x
Ratio 5 cores 26x
Ratio 6 cores 26x
Ratio 7 cores 26x
Ratio 8 cores 26x
Ratio 9 cores 26x
Ratio 10 cores 26x
Ratio 11 cores 26x
Ratio 12 cores 26x
Ratio 13 cores 26x
Ratio 14 cores 26x
Ratio 15 cores 26x
Ratio 16 cores 26x
Ratio 17 cores 26x
Ratio 18 cores 26x
Ratio 19 cores 26x
Ratio 20 cores 26x
Ratio 21 cores 26x
Ratio 22 cores 26x
Ratio 23 cores 26x
Ratio 24 cores 26x
TSC 2097.5 MHz
APERF 2597.1 MHz
MPERF 2097.6 MHz
IA Voltage Mode PCU adaptive
IA Voltage Offset 0 mV
GT Voltage Mode PCU adaptive
GT Voltage Offset 0 mV
LLC/Ring Voltage Mode PCU adaptive
LLC/Ring Voltage Offset 0 mV
Agent Voltage Mode PCU adaptive
Agent Voltage Offset 0 mV
Temperature 0 85 degC (185 degF) (Core #0)
Temperature 1 85 degC (185 degF) (Core #1)
Temperature 2 85 degC (185 degF) (Core #2)
Temperature 3 91 degC (195 degF) (Core #3)
Temperature 4 87 degC (188 degF) (Core #4)
Temperature 5 86 degC (186 degF) (Core #5)
Temperature 6 86 degC (186 degF) (Core #6)
Temperature 7 84 degC (183 degF) (Core #7)
Temperature 8 84 degC (183 degF) (Core #8)
Temperature 9 85 degC (185 degF) (Core #9)
Temperature 10 84 degC (183 degF) (Core #10)
Temperature 11 84 degC (183 degF) (Core #11)
Temperature 12 84 degC (183 degF) (Core #12)
Temperature 13 84 degC (183 degF) (Core #13)
Temperature 14 85 degC (185 degF) (Core #14)
Temperature 15 84 degC (183 degF) (Core #15)
Temperature 16 91 degC (195 degF) (Package)
Voltage 0 0.92 Volts (VID)
Voltage 1 +0.00 Volts (IA Offset)
Voltage 2 +0.00 Volts (GT Offset)
Voltage 3 +0.00 Volts (LLC/Ring Offset)
Voltage 4 +0.00 Volts (System Agent Offset)
Power 0 78.94 W (Package)
Power 1 n.a. (IA Cores)
Power 2 n.a. (Uncore)
Power 3 7.86 W (DRAM)
Clock Speed 0 2596.60 MHz (Core #0)
Clock Speed 1 2496.73 MHz (Core #1)
Clock Speed 2 2596.60 MHz (Core #2)
Clock Speed 3 2596.60 MHz (Core #3)
Clock Speed 4 2596.60 MHz (Core #4)
Clock Speed 5 2596.60 MHz (Core #5)
Clock Speed 6 2596.60 MHz (Core #6)
Clock Speed 7 2596.60 MHz (Core #7)
Clock Speed 8 2596.60 MHz (Core #8)
Clock Speed 9 2496.73 MHz (Core #9)
Clock Speed 10 2596.60 MHz (Core #10)
Clock Speed 11 2596.60 MHz (Core #11)
Clock Speed 12 2596.60 MHz (Core #12)
Clock Speed 13 2596.60 MHz (Core #13)
Clock Speed 14 2596.60 MHz (Core #14)
Clock Speed 15 2596.60 MHz (Core #15) | |
|
Michael Goetz Volunteer moderator Project administrator
 Send message
Joined: 21 Jan 10 Posts: 13955 ID: 53948 Credit: 392,876,449 RAC: 184,453
                               
|
16 per socket. It only has one stick of ram in one socket. I think that might be the problem. Each cpu should have memory per socket.
Actually, you want 4 sticks per socket (8 total), if the motherboard supports quad channel.
Temperature 0 89 degC (192 degF) (Core #0)
...
Temperature 16 92 degC (197 degF) (Package)
I don't run any Xeon systems, but those temperatures seem dangerously high. From Intel, the package temperature limit is 84 degrees, and you're way above that.
If you put in the optimal memory configuration, the number of instructions executed per second will significantly increase, which will increase the power consumed, which will increase temperatures even more.
Since Intel chips are usually pretty good about not literally going up in smoke (they slow themselves down rather than overheating), I suspect that in addition to having the wrong memory configuration, your CPU cooling system is severely underperforming for some reason (perhaps even completely broken) and the CPU is therefore throttling itself so that it quite literally does not self destruct. This may be a more significant problem than the memory, and certainly the system appears to be at risk of catastrophic failure.
____________
My lucky number is 75898524288+1 | |
|
|
Thanks for the heads up. just increased the cooling to max. Let's see what happens temp wise.
update:
got them down to 60 C.
____________
| |
|
Michael Goetz Volunteer moderator Project administrator
 Send message
Joined: 21 Jan 10 Posts: 13955 ID: 53948 Credit: 392,876,449 RAC: 184,453
                               
|
After 1 day:
Challenge: Summer Paralympics
App: 20 (ESP-LLR)
(As of 2016-09-03 18:07:47 UTC)
15990 tasks have been sent out. [CPU/GPU/anonymous_platform: 15990 (100%) / 0 (0%) / 0 (0%)]
Of those tasks that have been sent out:
468 (3%) came back with some kind of an error. [468 (3%) / 0 (0%) / 0 (0%)]
1590 (10%) have returned a successful result. [1590 (10%) / 0 (0%) / 0 (0%)]
13932 (87%) are still in progress. [13932 (87%) / 0 (0%) / 0 (0%)]
Of the tasks that have been returned successfully:
1231 (77%) are pending validation. [1231 (77%) / 0 (0%) / 0 (0%)]
342 (22%) have been successfully validated. [342 (22%) / 0 (0%) / 0 (0%)]
0 (0%) were invalid. [0 (0%) / 0 (0%) / 0 (0%)]
17 (1%) are inconclusive. [17 (1%) / 0 (0%) / 0 (0%)]
The current leading edge (i.e., latest work unit for which work has actually been sent out to a host) is n=9704126. The leading edge was at n=9464166 at the beginning of the challenge. Since the challenge started, the leading edge has advanced 2.54% as much as it had prior to the challenge!
Almost 16 thousand tasks were sent out in the first day. We normally do about 300 per day. We're sending them out at about 200 per hour. That's not the surge at the start of the challenge; that's 200 per hour being sent out every hour right now.
Great work everyone. Let's keep it up and see if we can't knock off one of the remaining Ks!
____________
My lucky number is 75898524288+1 | |
|
Michael Goetz Volunteer moderator Project administrator
 Send message
Joined: 21 Jan 10 Posts: 13955 ID: 53948 Credit: 392,876,449 RAC: 184,453
                               
|
Thanks for the heads up. just increased the cooling to max. Let's see what happens temp wise.
update:
got them down to 60 C.
That's MUCH better. 60 is a normal temperature. 90 is time to panic.
____________
My lucky number is 75898524288+1 | |
|
|
A while back I noticed that a dual xeon 2011 with single rank memory in quad channel was slower than my other system just like it that had dual rank memory in quad channel mode.
I did install dual rank quad channel in the slower system and it now run the same speed pretty much.
My question is, would it be better off with single rank in quad channel or dual rank in dual channel? Maybe it would be a wash?
This would be a single CPU socket 2011 running i7-3939k or E5-1650, pretty much those CPUs are the same. | |
|
mackerel Volunteer tester
 Send message
Joined: 2 Oct 08 Posts: 2639 ID: 29980 Credit: 568,393,769 RAC: 1,834
                              
|
My question is, would it be better off with single rank in quad channel or dual rank in dual channel? Maybe it would be a wash?
In my previous testing, dual rank was roughly equivalent to 20% higher clocked single rank. So quad channel single rank should be much faster than the same speed ram in dual channel dual rank. The relationship might not be linear, as it depends on the balance of CPU and ram relative to each other, so see that 20% as a maximum case. | |
|
|
Thank you for the reply. I think I will just keep my single rank Gskill 4x4gb 1866 Sniper mem in quad channel.
I had emailed Gskill about their memory and if it was single or dual rank.
Their answer was all their memory 4gb and under is single rank and 8gb sticks and larger are dual rank.
He said that has been industry standard for a while for desktop memory.
And I maybe able to find some older dual rank memory on ebay or somewhere
Whats the chances of someone finding a prime in this challenge? | |
|
Michael Goetz Volunteer moderator Project administrator
 Send message
Joined: 21 Jan 10 Posts: 13955 ID: 53948 Credit: 392,876,449 RAC: 184,453
                               
|
Whats the chances of someone finding a prime in this challenge?
Mathematically, it's a pretty low probability, but since Scott's on vacation, finding a prime is almost a certainty. :)
____________
My lucky number is 75898524288+1 | |
|
mackerel Volunteer tester
 Send message
Joined: 2 Oct 08 Posts: 2639 ID: 29980 Credit: 568,393,769 RAC: 1,834
                              
|
I had emailed Gskill about their memory and if it was single or dual rank.
Their answer was all their memory 4gb and under is single rank and 8gb sticks and larger are dual rank.
I found whoever does G.Skill support to be clueless about their own products when I tried contacting them before. Their lower latency high speed Trident Z/Ripjaws V modules at 8GB are single rank. It is a trend for newer modules to be higher density, hence lower rank. Higher latency 8GB modules still seem to be dual rank, but you need to check on a case by case basis. | |
|
|
Looks like my Haswell can do the ESP tasks in roughly 15 hours.
Would you like benchmarks posted once we start completing tasks? I expect my Skylakes to be around 9-10 hrs per unit.
It will be useful for people planning on entering the challenge. It certainly can't hurt.
Some benchmarks now that a reasonable sample size is available:
(1) Skylake i7 6700k @ stock 4.0 GHz (4 physical cores, 2400 MHz, dual rank/channel): 11.78 hrs/unit
(2) Skylake i3 6300 @ stock 3.8 GHz (2 physical cores, 2666 MHz, dual rank/channel): 10.33 hrs/unit (mobo limit @ 2133 MHz)
(3) Skylake i3 6100 @ stock 3.7 GHz (2 physical cores, 2666 MHz, dual rank/channel): 10.76 hrs/unit (mobo limit @ 2133 MHz)
(4) Ivy Bridge i7 3630QM @ 2.4-3.2 GHz (3/4 physical cores): 19.56 hrs/unit | |
|
mackerel Volunteer tester
 Send message
Joined: 2 Oct 08 Posts: 2639 ID: 29980 Credit: 568,393,769 RAC: 1,834
                              
|
Taking just the Skylake CPUs for now... all run dual channel ram, all cores active, HT off
i7-6700k @ 4.2 GHz, 3200 dual rank, 9.1 hours
i3-6100 @ 3.7 GHz, 3000 single rank, 9.3 hours
i5-6600k @ 4.2 GHz, 3000 dual rank, 10.0 hours
i7-6700k @ 4.0 GHz, 2666 single rank, 12.2 hours
Anthony, if the 2666 ram is the Crucial 4GB modules I think that was talked about in the past, they're single rank regardless what it says on their web page. I have it in my 4th system above as the mobo didn't like other faster ram I got for it. | |
|
|
Anthony, if the 2666 ram is the Crucial 4GB modules I think that was talked about in the past, they're single rank regardless what it says on their web page. I have it in my 4th system above as the mobo didn't like other faster ram I got for it.
Yes, they are indeed the Crucial Ballistix Tactical Series (2x4 Gb, DDR4). | |
|
mackerel Volunteer tester
 Send message
Joined: 2 Oct 08 Posts: 2639 ID: 29980 Credit: 568,393,769 RAC: 1,834
                              
|
Absent proof otherwise, I'm not aware of any DDR4 4GB modules that are dual rank. Some but not all 8GB modules are dual rank, so still have to be careful there.
Results for the remainder of my farm:
i3-4360 @ 3.7 GHz, 1600 dual rank, 11.3 hours
i5-5675C @ 3.5 GHz, 1600 dual rank, 11.7 hours
i5-4570S @ 3.2 GHz, 2400 dual rank, 14.0 hours
i3-4150T @ 3.0 GHz, 1600 single rank, 14.0 hours
E5-2583 v3 ES @ 2.3 GHz, 2133 single rank - quad channel, 27.7 hours
I would further caution, in previous comparisons like this I would manually pick only a certain FFT size for comparison, but I'm lazy here and this includes all results including smaller FFT ones. Having said that, with a small modification I think I could make it more representative of bigger FFTs. BRB! | |
|
mackerel Volunteer tester
 Send message
Joined: 2 Oct 08 Posts: 2639 ID: 29980 Credit: 568,393,769 RAC: 1,834
                              
|
tldr: I need to turn some cores off on the Xeon.
My idea to get 960k FFT indicative times without manually sorting them was to replace the average function with median. On the assumption the PC gets more 960k FFT units than smaller ones, the median will return a value within the 960k units. The effect of this varied by up to 2.5% depending on the specific PC, so I'm not going back to edit the earlier results.
It did make me think, the Xeon seems rather slow at double the unit time of an i3 at 3 GHz. Ram is limiting, but how much? I first looked at old benchmark data, but took new data using Prime95 since I replaced the ram with lower latency at the start of the challenge. The conclusion was clear: it was worse than I thought. Throughput peaked with 12 cores of 14 running, and I was losing 2% at 14. Going back the other way, to balance throughput and unit times I should be running 9 or 10 cores, each giving 95% or 97% relative to 12 cores, but with shorter estimated unit times of 19.1 and 20.8 hours, compared to 24.2 of 12, and 28.9 of 14 (assuming n=9.45M at start of challenge). The 14 core estimated time of 28.9h isn't far off the 28.4h median, and you would expect the median to be on the lower side of 960k units due to the presence of shorter units. Further complicating matters is I have partial units on the go, and a certain amount of time left in the challenge to optimise. #firstworldproblems | |
|
Michael Goetz Volunteer moderator Project administrator
 Send message
Joined: 21 Jan 10 Posts: 13955 ID: 53948 Credit: 392,876,449 RAC: 184,453
                               
|
tldr: I need to turn some cores off on the Xeon.
My idea to get 960k FFT indicative times without manually sorting them was to replace the average function with median. On the assumption the PC gets more 960k FFT units than smaller ones, the median will return a value within the 960k units. The effect of this varied by up to 2.5% depending on the specific PC, so I'm not going back to edit the earlier results.
Although some of the smaller-K tasks at the beginning of the challenge had an 864K FFT, most of the Ks at the beginning and everything right now is 960K (except for resends, of course.) So FFTs aren't going to make much of a difference.
However... N will have have a large effect. There's very few Ks remaining, and a lot of tasks are going out the door, and N is increasing. With FFT size factored out of the equation, task duration is directly proportional to N, and N has increased 3.26% since the beginning of the challenge, and it will continue to increase as the challenge keeps sending out tasks. Be careful when comparing task times from different parts of the challenge.
____________
My lucky number is 75898524288+1 | |
|
mackerel Volunteer tester
 Send message
Joined: 2 Oct 08 Posts: 2639 ID: 29980 Credit: 568,393,769 RAC: 1,834
                              
|
For the times given earlier, we are including units since start of challenge. I wont have the database stats of course, but simply looking at the times from several systems then ball park 30% of units were noticeably "shorter" than others, implicitly smaller FFT units than 960k. That's not insignificant if trying to work out an average looking forward.
I actually forgot to consider increasing n in my calculation on how many cores to turn off on the Xeon to get units out in time, but I decided to play safe regardless and sacrifice a few units in potential throughput to guarantee the others will make it in time. So it wouldn't have made a difference anyway unless I really push the limits of micro management, which I'm too lazy to do. | |
|
Michael Goetz Volunteer moderator Project administrator
 Send message
Joined: 21 Jan 10 Posts: 13955 ID: 53948 Credit: 392,876,449 RAC: 184,453
                               
|
For the times given earlier, we are including units since start of challenge. I wont have the database stats of course, but simply looking at the times from several systems then ball park 30% of units were noticeably "shorter" than others, implicitly smaller FFT units than 960k. That's not insignificant if trying to work out an average looking forward.
It's more like 5.5%. Of all the tasks sent out so far in the challenge, 1168 have the 864K FFT and 19940 have the 960K FFT.
(The standard "your mileage may vary" caveat applies: different CPUs use different FFTs sometimes, so you may be seeing different FFT sizes than the "official" FFT sizes.)
____________
My lucky number is 75898524288+1 | |
|
|
I have four servers with dual Xeons. Some are E5-2400 series and one is E5-2600 v4 series. The i3-6100 with 8gb of dual rank outperforms all these servers by 9 o 10 hours n this challenge. The prior challenge these severs did well, I placed 3rd overall, including a dual Interlagos computer with 12 cores each and 256 gb of memory. I would stick with i7-6700 cpus or any i7 that is better. I am considering two more H170 motherboards with i7-6700 and dual rank memory.
What configuration would you suggest to use with an Intel cpu so that I can use the 256gb DDR3 with and then I could retire the AMD computer?
____________
| |
|
Michael Goetz Volunteer moderator Project administrator
 Send message
Joined: 21 Jan 10 Posts: 13955 ID: 53948 Credit: 392,876,449 RAC: 184,453
                               
|
After about 2 days:
Challenge: Summer Paralympics
App: 20 (ESP-LLR)
(As of 2016-09-04 17:17:53 UTC)
21829 tasks have been sent out. [CPU/GPU/anonymous_platform: 21829 (100%) / 0 (0%) / 0 (0%)]
Of those tasks that have been sent out:
1199 (5%) came back with some kind of an error. [1199 (5%) / 0 (0%) / 0 (0%)]
5012 (23%) have returned a successful result. [5012 (23%) / 0 (0%) / 0 (0%)]
15618 (72%) are still in progress. [15618 (72%) / 0 (0%) / 0 (0%)]
Of the tasks that have been returned successfully:
3207 (64%) are pending validation. [3207 (64%) / 0 (0%) / 0 (0%)]
1727 (34%) have been successfully validated. [1727 (34%) / 0 (0%) / 0 (0%)]
4 (0%) were invalid. [4 (0%) / 0 (0%) / 0 (0%)]
74 (1%) are inconclusive. [74 (1%) / 0 (0%) / 0 (0%)]
The current leading edge (i.e., latest work unit for which work has actually been sent out to a host) is n=9785742. The leading edge was at n=9464166 at the beginning of the challenge. Since the challenge started, the leading edge has advanced 3.40% as much as it had prior to the challenge!
____________
My lucky number is 75898524288+1 | |
|
mackerel Volunteer tester
 Send message
Joined: 2 Oct 08 Posts: 2639 ID: 29980 Credit: 568,393,769 RAC: 1,834
                              
|
Sieve is relatively undemanding, and scales well with any CPU power. That also applies to smaller LLR units. Bigger LLR on the other hand, quickly overwhelm the CPU cache and become limited by ram performance.
For DDR3 and using for LLR, I propose Broadwell. This has a unique trick in the desktop parts, in the form of 128MB L4 cache which negates the need for fast ram, at least for any LLR task we're likely to see in its useful lifetime. There are only two CPUs in the range: i5-5675C and i7-5775C. Double check mobo compatibility as older Haswell boards wont support it. These CPU have a drawback in that they don't overclock well. Even piling on the volts I could only get to 4.1 GHz stable, and for a balance of power consumption I run my i5-5675C at 3.5 GHz only, somewhat above the 3.1 GHz base. If LLR is the main use, I don't think getting the i7 is worth it.
If the DDR3 is high speed (2400) then a regular high clocked Haswell should still do well. Again check mobo compatibility with high speed ram.
In a quick check, the pricing (in UK at least) has gone silly for Broadwell parts and they cost more than Skylake now. I'd rather just resell the DDR3 ram and get more Skylake instead. | |
|
|
Sieve is relatively undemanding, and scales well with any CPU power. That also applies to smaller LLR units. Bigger LLR on the other hand, quickly overwhelm the CPU cache and become limited by ram performance.
For DDR3 and using for LLR, I propose Broadwell. This has a unique trick in the desktop parts, in the form of 128MB L4 cache which negates the need for fast ram, at least for any LLR task we're likely to see in its useful lifetime. There are only two CPUs in the range: i5-5675C and i7-5775C. Double check mobo compatibility as older Haswell boards wont support it. These CPU have a drawback in that they don't overclock well. Even piling on the volts I could only get to 4.1 GHz stable, and for a balance of power consumption I run my i5-5675C at 3.5 GHz only, somewhat above the 3.1 GHz base. If LLR is the main use, I don't think getting the i7 is worth it.
If the DDR3 is high speed (2400) then a regular high clocked Haswell should still do well. Again check mobo compatibility with high speed ram.
In a quick check, the pricing (in UK at least) has gone silly for Broadwell parts and they cost more than Skylake now. I'd rather just resell the DDR3 ram and get more Skylake instead.
THANKS!
____________
| |
|
Michael Goetz Volunteer moderator Project administrator
 Send message
Joined: 21 Jan 10 Posts: 13955 ID: 53948 Credit: 392,876,449 RAC: 184,453
                               
|
In case anyone is interested, these are the smallest and the largest numbers tested so far during the challenge:
Smallest: 2,819,774 digits
Largest: 2,956,511 digits
(That's counting tasks that have been sent out, regardless of whether they've been completed.)
I suspect we'll surpass three million digits before the challenge ends.
____________
My lucky number is 75898524288+1 | |
|
Michael Goetz Volunteer moderator Project administrator
 Send message
Joined: 21 Jan 10 Posts: 13955 ID: 53948 Credit: 392,876,449 RAC: 184,453
                               
|
After 3 days:
Challenge: Summer Paralympics
App: 20 (ESP-LLR)
(As of 2016-09-05 18:54:28 UTC)
28354 tasks have been sent out. [CPU/GPU/anonymous_platform: 28354 (100%) / 0 (0%) / 0 (0%)]
Of those tasks that have been sent out:
1863 (7%) came back with some kind of an error. [1863 (7%) / 0 (0%) / 0 (0%)]
9806 (35%) have returned a successful result. [9806 (35%) / 0 (0%) / 0 (0%)]
16685 (59%) are still in progress. [16685 (59%) / 0 (0%) / 0 (0%)]
Of the tasks that have been returned successfully:
5248 (54%) are pending validation. [5248 (54%) / 0 (0%) / 0 (0%)]
4400 (45%) have been successfully validated. [4400 (45%) / 0 (0%) / 0 (0%)]
24 (0%) were invalid. [24 (0%) / 0 (0%) / 0 (0%)]
134 (1%) are inconclusive. [134 (1%) / 0 (0%) / 0 (0%)]
The current leading edge (i.e., latest work unit for which work has actually been sent out to a host) is n=9878496. The leading edge was at n=9464166 at the beginning of the challenge. Since the challenge started, the leading edge has advanced 4.38% as much as it had prior to the challenge!
____________
My lucky number is 75898524288+1 | |
|
Michael Goetz Volunteer moderator Project administrator
 Send message
Joined: 21 Jan 10 Posts: 13955 ID: 53948 Credit: 392,876,449 RAC: 184,453
                               
|
After 4 days:
Challenge: Summer Paralympics
App: 20 (ESP-LLR)
(As of 2016-09-06 18:42:34 UTC)
33116 tasks have been sent out. [CPU/GPU/anonymous_platform: 33116 (100%) / 0 (0%) / 0 (0%)]
Of those tasks that have been sent out:
3585 (11%) came back with some kind of an error. [3585 (11%) / 0 (0%) / 0 (0%)]
14399 (43%) have returned a successful result. [14399 (43%) / 0 (0%) / 0 (0%)]
15132 (46%) are still in progress. [15132 (46%) / 0 (0%) / 0 (0%)]
Of the tasks that have been returned successfully:
6675 (46%) are pending validation. [6675 (46%) / 0 (0%) / 0 (0%)]
7503 (52%) have been successfully validated. [7503 (52%) / 0 (0%) / 0 (0%)]
47 (0%) were invalid. [47 (0%) / 0 (0%) / 0 (0%)]
174 (1%) are inconclusive. [174 (1%) / 0 (0%) / 0 (0%)]
The current leading edge (i.e., latest work unit for which work has actually been sent out to a host) is n=9923535. The leading edge was at n=9464166 at the beginning of the challenge. Since the challenge started, the leading edge has advanced 4.85% as much as it had prior to the challenge!
____________
My lucky number is 75898524288+1 | |
|
Michael Goetz Volunteer moderator Project administrator
 Send message
Joined: 21 Jan 10 Posts: 13955 ID: 53948 Credit: 392,876,449 RAC: 184,453
                               
|
With 1 day to go, it's time for the usual end of challenge reminder:
At the Conclusion of the Challenge
After the challenge ends, we would prefer users "moving on" to finish those tasks they have downloaded, if not then please ABORT the WU's instead of DETACHING, RESETTING, or PAUSING.
ABORTING WU's allows them to be recycled immediately; thus a much faster "clean up" to the end of a Challenge. DETACHING, RESETTING, and PAUSING WU's causes them to remain in limbo until they EXPIRE. Therefore, we must wait until WU's expire to send them out to be completed.
____________
My lucky number is 75898524288+1 | |
|
Michael Goetz Volunteer moderator Project administrator
 Send message
Joined: 21 Jan 10 Posts: 13955 ID: 53948 Credit: 392,876,449 RAC: 184,453
                               
|
Whats the chances of someone finding a prime in this challenge?
Mathematically, it's a pretty low probability, but since Scott's on vacation, finding a prime is almost a certainty. :)
We did, in fact find a prime -- a PPS-Mega. No ESP prime yet, but there's still a few hours to go.
____________
My lucky number is 75898524288+1 | |
|
Michael Goetz Volunteer moderator Project administrator
 Send message
Joined: 21 Jan 10 Posts: 13955 ID: 53948 Credit: 392,876,449 RAC: 184,453
                               
|
In case anyone is interested, these are the smallest and the largest numbers tested so far during the challenge:
Smallest: 2,819,774 digits
Largest: 2,956,511 digits
(That's counting tasks that have been sent out, regardless of whether they've been completed.)
I suspect we'll surpass three million digits before the challenge ends.
It looks like we're going to fall just a little bit short. As the challenge winds down, people have, as usual, been aborting tasks they won't start prior to the end of the challenge. As a result, a lot of tasks being sent out have been resends, so the leading edge has mostly stalled over the last 12 or so hours.
Here's where we stand right now:
Largest task sent out so far: 2,993,464
Leading edge 'n': 9,944,054
____________
My lucky number is 75898524288+1 | |
|
Michael Goetz Volunteer moderator Project administrator
 Send message
Joined: 21 Jan 10 Posts: 13955 ID: 53948 Credit: 392,876,449 RAC: 184,453
                               
|
The challenge is over!
Challenge: Summer Paralympics
App: 20 (ESP-LLR)
(As of 2016-09-07 18:00:03 UTC)
36407 tasks have been sent out. [CPU/GPU/anonymous_platform: 36407 (100%) / 0 (0%) / 0 (0%)]
Of those tasks that have been sent out:
6947 (19%) came back with some kind of an error. [6947 (19%) / 0 (0%) / 0 (0%)]
20234 (56%) have returned a successful result. [20234 (56%) / 0 (0%) / 0 (0%)]
9226 (25%) are still in progress. [9226 (25%) / 0 (0%) / 0 (0%)]
Of the tasks that have been returned successfully:
6709 (33%) are pending validation. [6709 (33%) / 0 (0%) / 0 (0%)]
13162 (65%) have been successfully validated. [13162 (65%) / 0 (0%) / 0 (0%)]
83 (0%) were invalid. [83 (0%) / 0 (0%) / 0 (0%)]
280 (1%) are inconclusive. [280 (1%) / 0 (0%) / 0 (0%)]
The current leading edge (i.e., latest work unit for which work has actually been sent out to a host) is n=9944054. The leading edge was at n=9464166 at the beginning of the challenge. Since the challenge started, the leading edge has advanced 5.07% as much as it had prior to the challenge!
What a fantastic challenge! over 20 thousand tasks were completed -- that's more than 13 times what we would normally do in 5 days. Thank you all!
While we didn't find any primes, you never know when the next prime will pop up. With these numbers being just shy of 3 million digits, a new ESP prime would be the 17th largest prime ever discovered. So please feel free to continue with ESP if you want. The next prime is out there somewhere.
____________
My lucky number is 75898524288+1 | |
|
Michael Goetz Volunteer moderator Project administrator
 Send message
Joined: 21 Jan 10 Posts: 13955 ID: 53948 Credit: 392,876,449 RAC: 184,453
                               
|
With the challenge over, it's time to start the cleanup. I expect the cleanup to take between 18 and 36 days.
Cleanup:
Sep-7: Summer Paralympics: 6976 tasks outstanding; 6876 affecting individual (293) scoring positions; 6268 affecting team (67) scoring positions.
Sep-8: Summer Paralympics: 5849 tasks outstanding; 5596 affecting individual (287) scoring positions; 4107 affecting team (64) scoring positions.
____________
My lucky number is 75898524288+1 | |
|
Michael Goetz Volunteer moderator Project administrator
 Send message
Joined: 21 Jan 10 Posts: 13955 ID: 53948 Credit: 392,876,449 RAC: 184,453
                               
|
I rarely say this, but we actually could use a little help with the cleanup. ESP isn't a particularly popular project, and there's about 1500 cleanup tasks waiting to be sent out. That's probably 5 day's worth, and that's assuming no more tasks get cancelled. This is one of those rare times when you actually can help with a cleanup.
____________
My lucky number is 75898524288+1 | |
|
|
Michael,
I can commit for about 225 units. If you have the means to send me units needing confirmation, please do so. I already have enough pending credit (157k) to flip to Amethyst (1M) but I am still crunching ESP LLR so as to flip to Ruby (2M) in about ten days' tiime.
Edit: I just checked and all eleven of my cores are running units needing confirmation. | |
|
|
I'll fire some up and get the word out for you. | |
|
Michael Goetz Volunteer moderator Project administrator
 Send message
Joined: 21 Jan 10 Posts: 13955 ID: 53948 Credit: 392,876,449 RAC: 184,453
                               
|
FYI, in case you didn't read it yet, there's a possibility we may suspend the TRP sieve much sooner than expected. I've given the official 30 day warning, just in case.
If you want to work on that badge instead, feel free to do so. The ESP cleanup will take care of itself, and a few extra days don't make much of a difference in the big picture. The TRP-Sieve badge, however, may have a limited number of days left, so if that's important to you, please do that instead.
____________
My lucky number is 75898524288+1 | |
|
|
I rarely say this, but we actually could use a little help with the cleanup. ESP isn't a particularly popular project, and there's about 1500 cleanup tasks waiting to be sent out. That's probably 5 day's worth, and that's assuming no more tasks get cancelled. This is one of those rare times when you actually can help with a cleanup.
I'll help clean up ESP for a few days. Ran up to Sapphire in TRP-SV last challenge and am happy with that.
By the way, Zune-san, in case you're reading. :) Next time I'll make sure all my other sub-projects are cancelled and perhaps bother with turning off HT too... can't have you snatching first place from me like that at the last second like that again. ;) That being said, my idle hardware can change at any time, so we'll see what's happening in four weeks. :D | |
|
tng Send message
Joined: 29 Aug 10 Posts: 465 ID: 66603 Credit: 45,703,545,082 RAC: 23,212,924
                                                   
|
I rarely say this, but we actually could use a little help with the cleanup. ESP isn't a particularly popular project, and there's about 1500 cleanup tasks waiting to be sent out. That's probably 5 day's worth, and that's assuming no more tasks get cancelled. This is one of those rare times when you actually can help with a cleanup.
I'll help clean up ESP for a few days. Ran up to Sapphire in TRP-SV last challenge and am happy with that.
By the way, Zune-san, in case you're reading. :) Next time I'll make sure all my other sub-projects are cancelled and perhaps bother with turning off HT too... can't have you snatching first place from me like that at the last second like that again. ;) That being said, my idle hardware can change at any time, so we'll see what's happening in four weeks. :D
In four weeks, temperatures here should be significantly lower. You'd better bring all that you have.
____________
| |
|
Scott Brown Volunteer moderator Project administrator Volunteer tester Project scientist
 Send message
Joined: 17 Oct 05 Posts: 2382 ID: 1178 Credit: 17,956,227,805 RAC: 11,913,199
                                                
|
I rarely say this, but we actually could use a little help with the cleanup. ESP isn't a particularly popular project, and there's about 1500 cleanup tasks waiting to be sent out. That's probably 5 day's worth, and that's assuming no more tasks get cancelled. This is one of those rare times when you actually can help with a cleanup.
I'll help clean up ESP for a few days. Ran up to Sapphire in TRP-SV last challenge and am happy with that.
By the way, Zune-san, in case you're reading. :) Next time I'll make sure all my other sub-projects are cancelled and perhaps bother with turning off HT too... can't have you snatching first place from me like that at the last second like that again. ;) That being said, my idle hardware can change at any time, so we'll see what's happening in four weeks. :D
In four weeks, temperatures here should be significantly lower. You'd better bring all that you have.
Indeed! :)
| |
|
|
In four weeks, temperatures here should be significantly lower. You'd better bring all that you have.
Thankfully temperature isn't an issue for me; climate controlled lab. :) There are a lot of systems I could bring online, but tend to focus on the more powerful ones instead. Probably at least 300 more cores of various sorts sitting around... really just a question of if I can feed them without blowing a breaker or something. :p Now, if I were a borger, like I'm sure some are, it would be a frighteningly different picture! <grin>
| |
|
Michael Goetz Volunteer moderator Project administrator
 Send message
Joined: 21 Jan 10 Posts: 13955 ID: 53948 Credit: 392,876,449 RAC: 184,453
                               
|
Back to talking about ESP: :)
Cleanup:
Sep-7: Summer Paralympics: 6976 tasks outstanding; 6876 affecting individual (293) scoring positions; 6268 affecting team (67) scoring positions.
Sep-8: Summer Paralympics: 5849 tasks outstanding; 5596 affecting individual (287) scoring positions; 4107 affecting team (64) scoring positions.
Sep-9: Summer Paralympics: 4766 tasks outstanding; 4557 affecting individual (283) scoring positions; 3370 affecting team (60) scoring positions.
____________
My lucky number is 75898524288+1 | |
|
Michael Goetz Volunteer moderator Project administrator
 Send message
Joined: 21 Jan 10 Posts: 13955 ID: 53948 Credit: 392,876,449 RAC: 184,453
                               
|
I've moved several posts over to the thread that's actually about the possibility of suspending TRP-Sieve. It's better to have the discussion over there.
The thread is located here: http://www.primegrid.com/forum_thread.php?id=7000
____________
My lucky number is 75898524288+1 | |
|
|
The fastest completed ESP task in the database was done by Rafael's Skylake i5-6600K in just under 8 hours.
There's also a number of tasks completed in less than 9 hours. Of the top 25 fastest completed ESP tasks, all were done by Skylake i5-6600K and i7-6700K CPUs.
O.O
I feel so important right now....
Still, that's a bit surprising. While 4263mhz core/cache with 3146mhz dual channel dual rank RAM is blazing fast, I would expect other systems (say, 3200mhz 4.3 owners with 8mb of cache) to be even faster, even if marginally. I guess running 3 WU at a time rather than 4 really speeds things up....
http://www.primegrid.com/result.php?resultid=738677941
I did a task with an FFT of 960K (I believe Rafael's was of the lower FFT size) in 9.9 hours. It was running with one other ESP-LLR task for about 85% of the time, and when that other task completed I suspended everything apart from the task in question. I think for about 10% of the time (a complete guess) I may have been running my GPU on PPS-Sieve. If it had have run 100% on its own then I know the time saved would have still only been marginal, but maybe it would have completed in just a bit over 9 hours? Who knows, just thinking aloud about a Haswell 4790K@4.4, and 2133 RAM, vs. Rafael there. I think my L3 cache must have made a significant difference, considering my RAM is about a gigahert slower and my CPU transistors and pathways are twice the size.
Not meant to be a boast, or a competition post - again, simply thinking aloud. I realise that it's not often that any of us will simply run one task for about 10 hours, anyway, aside from pure benchmarking it's a waste of good CPU time.
All about balancing and priorities, isn't it... | |
|
mackerel Volunteer tester
 Send message
Joined: 2 Oct 08 Posts: 2639 ID: 29980 Credit: 568,393,769 RAC: 1,834
                              
|
Comparing unit times has lots of different factors...
The n has advanced quite a lot between start and end of challenge, of the order 5% and times will scale accordingly. So a computer doing 10 hours a unit at the start would take 10.5 hours at the end.
When running fewer cores, ram impact is also reduced significantly. Even further than that, if running 1 task on an 8MB cache CPU, it doesn't even need to touch ram significantly. I've yet to see if that yields a significant boost. Cache quantity doesn't matter once you run two or more units, you're back to hitting ram then.
If ram isn't limiting, then performance scales with CPU clock. Skylake is about 14% faster than Haswell in that respect, so a 4.4 GHz Haswell would be roughly equivalent to a 3.86 GHz Skylake. | |
|
|
There's no doubt that I would prefer a Skylake to my Haswell, and I'm not trying to suggest otherwise.
Results on paper, and results tested on one system vs. one other system do tend to suggest that most benchmarks are unique and/or guesswork, though, don't they? None of the benchmarks quoted by anyone on here ever add up to my own experience.
I'm not sure what you mean by cache not mattering that much, when in another thread you've said that certain Broadwells might be worth buying because of their L4 cache. I realise that those Iris Pro CPUs are unique, but make your mind up... does on-chip cache matter or not? I don't want to start a fight but clearly individual benchmarking doesn't really mean THAT much, and Intel probably put more cache on their more expensive chips because they can be faster. Maybe the best policy is to hire 20,000 Core 2 Duos for the next CPU challenge, with each completing one task within the deadline. Doesn't really mean much though, does it? And best not bother spending two days wasting time benching them, either.
But I do want a Skydrake. Skylake. | |
|
mackerel Volunteer tester
 Send message
Joined: 2 Oct 08 Posts: 2639 ID: 29980 Credit: 568,393,769 RAC: 1,834
                              
|
I've worked out a lot of the factors affecting LLR performance. Not all of them, but enough of them. I can't predict to the nearest second what any unit is going to run at, but with some hardware info and at least one reference point, I can scale that reasonably for Intel platforms since Sandy Bridge. Without sufficient info, you can't do that. At a minimum you'd need to know CPU clock, number of concurrent tasks for how many cores/threads, ram speed, channels and ranks. If you don't factor all that in, things can and will look different. There will be other factors not considered that could also have an influence, but the ones I mentioned should cover the majority of it.
On processor cache, the full statement should be that the difference between 6MB and 8MB L3 cache of most desktop i5 and i7 CPUs is insignificant when running more than one core on a large LLR task, like ESP in this challenge. Once you exceed the cache size, it doesn't matter how much you have as the bottleneck is elsewhere. There are some edge cases where it may make a difference, like for example running one task per core of MEGA, or one task per CPU of ESP, where in theory the 8MB will have a nice benefit over 6MB.
Desktop Broadwell was an odd ball with 128MB L4 cache, which essentially replaces the ram as far as LLR tasks we do are concerned. How many people bought desktop Broadwell? I feel like the only one, and without any indication of more of the same from Intel, I am treating that as the exception not the rule. Maybe some high end mobile users will get that benefit too, but I wouldn't want to crunch on a laptop regardless. | |
|
Michael Goetz Volunteer moderator Project administrator
 Send message
Joined: 21 Jan 10 Posts: 13955 ID: 53948 Credit: 392,876,449 RAC: 184,453
                               
|
Cleanup:
Sep-7: Summer Paralympics: 6976 tasks outstanding; 6876 affecting individual (293) scoring positions; 6268 affecting team (67) scoring positions.
Sep-8: Summer Paralympics: 5849 tasks outstanding; 5596 affecting individual (287) scoring positions; 4107 affecting team (64) scoring positions.
Sep-9: Summer Paralympics: 4766 tasks outstanding; 4557 affecting individual (283) scoring positions; 3370 affecting team (60) scoring positions.
Sep-10: Summer Paralympics: 4004 tasks outstanding; 3452 affecting individual (279) scoring positions; 2316 affecting team (57) scoring positions.
Sep-11: Summer Paralympics: 3254 tasks outstanding; 2813 affecting individual (273) scoring positions; 1784 affecting team (51) scoring positions.
Sep-12: Summer Paralympics: 2347 tasks outstanding; 1910 affecting individual (252) scoring positions; 847 affecting team (41) scoring positions.
Sep-13: Summer Paralympics: 1899 tasks outstanding; 1475 affecting individual (240) scoring positions; 647 affecting team (38) scoring positions.
____________
My lucky number is 75898524288+1 | |
|
|
I have squeezed seven of the nine tasks I had suspended due to TRP Sieve and the other two will be allowed to continue later today (they are about 70% complete). I realize it is a drop in the bucket but every unit helps. | |
|
Michael Goetz Volunteer moderator Project administrator
 Send message
Joined: 21 Jan 10 Posts: 13955 ID: 53948 Credit: 392,876,449 RAC: 184,453
                               
|
Cleanup:
Sep-7: Summer Paralympics: 6976 tasks outstanding; 6876 affecting individual (293) scoring positions; 6268 affecting team (67) scoring positions.
Sep-8: Summer Paralympics: 5849 tasks outstanding; 5596 affecting individual (287) scoring positions; 4107 affecting team (64) scoring positions.
Sep-9: Summer Paralympics: 4766 tasks outstanding; 4557 affecting individual (283) scoring positions; 3370 affecting team (60) scoring positions.
Sep-10: Summer Paralympics: 4004 tasks outstanding; 3452 affecting individual (279) scoring positions; 2316 affecting team (57) scoring positions.
Sep-11: Summer Paralympics: 3254 tasks outstanding; 2813 affecting individual (273) scoring positions; 1784 affecting team (51) scoring positions.
Sep-12: Summer Paralympics: 2347 tasks outstanding; 1910 affecting individual (252) scoring positions; 847 affecting team (41) scoring positions.
Sep-13: Summer Paralympics: 1899 tasks outstanding; 1475 affecting individual (240) scoring positions; 647 affecting team (38) scoring positions.
Sep-14: Summer Paralympics: 1535 tasks outstanding; 1116 affecting individual (222) scoring positions; 486 affecting team (32) scoring positions.
Sep 15: Summer Paralympics: 1210 tasks outstanding; 831 affecting individual (202) scoring positions; 338 affecting team (28) scoring positions.
Sep-16: Summer Paralympics: 948 tasks outstanding; 639 affecting individual (184) scoring positions; 118 affecting team (23) scoring positions.
Sep-17: Summer Paralympics: 851 tasks outstanding; 564 affecting individual (172) scoring positions; 108 affecting team (22) scoring positions.
Sep-18: Summer Paralympics: 803 tasks outstanding; 535 affecting individual (170) scoring positions; 98 affecting team (21) scoring positions.
Sep-19: Summer Paralympics: 718 tasks outstanding; 461 affecting individual (164) scoring positions; 85 affecting team (19) scoring positions.
Sep-20: Summer Paralympics: 447 tasks outstanding; 255 affecting individual (116) scoring positions; 45 affecting team (11) scoring positions.
Sep-21: Summer Paralympics: 361 tasks outstanding; 202 affecting individual (99) scoring positions; 36 affecting team (8) scoring positions.
Sep-22: Summer Paralympics: 300 tasks outstanding; 151 affecting individual (86) scoring positions; 24 affecting team (6) scoring positions.
Sep-23: Summer Paralympics: 169 tasks outstanding; 78 affecting individual (59) scoring positions; 12 affecting team (4) scoring positions.
Sep-24: Summer Paralympics: 133 tasks outstanding; 63 affecting individual (47) scoring positions; 11 affecting team (4) scoring positions.
Sep-25: Summer Paralympics: 111 tasks outstanding; 50 affecting individual (39) scoring positions; 10 affecting team (4) scoring positions.
Sep-26: Summer Paralympics: 96 tasks outstanding; 46 affecting individual (36) scoring positions; 9 affecting team (3) scoring positions.
Sep-27: Summer Paralympics: 77 tasks outstanding; 38 affecting individual (32) scoring positions; 6 affecting team (2) scoring positions.
Sep-28: Summer Paralympics: 58 tasks outstanding; 24 affecting individual (21) scoring positions; 5 affecting team (2) scoring positions.
Sep-29: Summer Paralympics: 46 tasks outstanding; 20 affecting individual (17) scoring positions; 4 affecting team (2) scoring positions.
Sep-30: Summer Paralympics: 34 tasks outstanding; 13 affecting individual (12) scoring positions; 3 affecting team (2) scoring positions.
Oct-1: Summer Paralympics: 21 tasks outstanding; 9 affecting individual (8) scoring positions; 1 affecting team (1) scoring positions.
Oct-2: Summer Paralympics: 19 tasks outstanding; 8 affecting individual (7) scoring positions; 1 affecting team (1) scoring positions.
Oct-3: Summer Paralympics: 12 tasks outstanding; 7 affecting individual (6) scoring positions; 0 affecting team (0) scoring positions.
Oct-4: Summer Paralympics: 9 tasks outstanding; 4 affecting individual (4) scoring positions; 0 affecting team (0) scoring positions.
Oct-5: Summer Paralympics: 8 tasks outstanding; 3 affecting individual (3) scoring positions; 0 affecting team (0) scoring positions.
____________
My lucky number is 75898524288+1 | |
|
RogerVolunteer developer Volunteer tester
 Send message
Joined: 27 Nov 11 Posts: 1138 ID: 120786 Credit: 268,668,824 RAC: 0
                    
|
The results are final!
Top 3 individuals:
1: zunewantan
2: fastgeek
3: Scott Brown
Top 3 teams:
1: Aggie The Pew
2: Czech National Team
3: [H]ard|OCP
Congratulations to the winners, and well done to everyone who participated.
See you at the next challenge!
____________
| |
|
Message boards :
Number crunching :
Summer Paralympics Challenge |