We've been doing some benchmarks with Elasticsearch on the 5-node PicoCluster. We were able to push 1,000,000 tweets and 3495.1316MB in just under 9 hours. The tweets are small, around 1KB each, but have a lot of fields and are pretty complicated. That's pretty good considering that the SD cards are not very good a small reads and writes.
Pushed 1000 docs in 34.457 secs at 29.021679 docs/sec 3.4431486 MB/sec, 980000 total docs in 520 minutes 33 seconds at 31.37682 docs/second 98.0% complete
Pushed 1000 docs in 36.632 secs at 27.298536 docs/sec 3.5004349 MB/sec, 981000 total docs in 521 minutes 9 seconds at 31.372042 docs/second 98.1% complete
Pushed 1000 docs in 34.607 secs at 28.89589 docs/sec 3.5908194 MB/sec, 982000 total docs in 521 minutes 44 seconds at 31.369303 docs/second 98.2% complete
Pushed 1000 docs in 30.67 secs at 32.605152 docs/sec 3.3349895 MB/sec, 983000 total docs in 522 minutes 15 seconds at 31.370514 docs/second 98.299995% complete
Pushed 1000 docs in 31.243 secs at 32.007168 docs/sec 3.431964 MB/sec, 984000 total docs in 522 minutes 46 seconds at 31.37115 docs/second 98.4% complete
Pushed 1000 docs in 28.858 secs at 34.652435 docs/sec 3.4087648 MB/sec, 985000 total docs in 523 minutes 15 seconds at 31.374163 docs/second 98.5% complete
Pushed 1000 docs in 29.598 secs at 33.786068 docs/sec 3.4104357 MB/sec, 986000 total docs in 523 minutes 44 seconds at 31.376436 docs/second 98.6% complete
Pushed 1000 docs in 32.356 secs at 30.90617 docs/sec 3.4084692 MB/sec, 987000 total docs in 524 minutes 17 seconds at 31.375952 docs/second 98.7% complete
Pushed 1000 docs in 37.807 secs at 26.450129 docs/sec 3.4255342 MB/sec, 988000 total docs in 524 minutes 55 seconds at 31.370039 docs/second 98.799995% complete
Pushed 1000 docs in 33.404 secs at 29.936535 docs/sec 3.4184904 MB/sec, 989000 total docs in 525 minutes 28 seconds at 31.36852 docs/second 98.9% complete
Pushed 1000 docs in 34.465 secs at 29.014942 docs/sec 3.4793549 MB/sec, 990000 total docs in 526 minutes 2 seconds at 31.36595 docs/second 99.0% complete
Pushed 1000 docs in 30.792 secs at 32.475967 docs/sec 3.4305592 MB/sec, 991000 total docs in 526 minutes 33 seconds at 31.367033 docs/second 99.1% complete
Pushed 1000 docs in 29.749 secs at 33.614574 docs/sec 3.4574842 MB/sec, 992000 total docs in 527 minutes 3 seconds at 31.369146 docs/second 99.2% complete
Pushed 1000 docs in 32.825 secs at 30.464584 docs/sec 3.370614 MB/sec, 993000 total docs in 527 minutes 36 seconds at 31.368208 docs/second 99.299995% complete
Pushed 1000 docs in 37.048 secs at 26.99201 docs/sec 3.451209 MB/sec, 994000 total docs in 528 minutes 13 seconds at 31.36309 docs/second 99.4% complete
Pushed 1000 docs in 35.307 secs at 28.322996 docs/sec 3.3885374 MB/sec, 995000 total docs in 528 minutes 48 seconds at 31.35971 docs/second 99.5% complete
Pushed 1000 docs in 37.64 secs at 26.567482 docs/sec 3.4242926 MB/sec, 996000 total docs in 529 minutes 26 seconds at 31.35403 docs/second 99.6% complete
Pushed 1000 docs in 28.108 secs at 35.57706 docs/sec 3.4203835 MB/sec, 997000 total docs in 529 minutes 54 seconds at 31.357765 docs/second 99.7% complete
Pushed 1000 docs in 28.886 secs at 34.618847 docs/sec 3.4412107 MB/sec, 998000 total docs in 530 minutes 23 seconds at 31.360725 docs/second 99.8% complete
Pushed 1000 docs in 40.074 secs at 24.953835 docs/sec 3.4108858 MB/sec, 999000 total docs in 531 minutes 3 seconds at 31.352667 docs/second 99.9% complete
Pushed 1000 docs in 0.0 secs at Infinity docs/sec 3.4554148 MB/sec, 1000000 total docs in 531 minutes 3 seconds at 31.38405 docs/second 100.0% complete
Pushed 1000000 total docs and 3495.1316MB in 531 minutes 3 seconds at 31.38405 per second,
Wednesday, December 31, 2014
Thursday, December 18, 2014
10GB Terasort Benchmark on 5-node Raspberry PI Cluster - 2H 52m 56s
We just set a new record for the 10GB terasort on a 5-node PicoCluster! We cut over an our off the benchmark time bringing the total to under 3 hours! Pretty amazing!
Hadoop job_201412181311_0002 on master User: hadoopJob Name: TeraSort
Job File: hdfs://pi0:54310/tmp/hadoop-hadoop/mapred/staging/hadoop/.staging/job_201412181311_0002/job.xml
Submit Host: pi0
Submit Host Address: 10.1.10.120
Job-ACLs: All users are allowed
Job Setup: Successful
Status: Succeeded
Started at: Thu Dec 18 14:54:20 MST 2014
Finished at: Thu Dec 18 17:47:16 MST 2014
Finished in: 2hrs, 52mins, 56sec
Job Cleanup: Successful
Kind | % Complete | Num Tasks | Pending | Running | Complete | Killed | Failed/Killed Task Attempts | |
---|---|---|---|---|---|---|---|---|
map | 100.00% | 80 | 0 | 0 | 80 | 0 | 0 / 0 | |
reduce | 100.00% | 80 | 0 | 0 | 80 | 0 | 0 / 0 |
Counter | Map | Reduce | Total | |
---|---|---|---|---|
Map-Reduce Framework | Spilled Records | 0 | 0 | 300,000,000 |
Map output materialized bytes | 0 | 0 | 10,200,038,400 | |
Reduce input records | 0 | 0 | 100,000,000 | |
Virtual memory (bytes) snapshot | 0 | 0 | 46,356,074,496 | |
Map input records | 0 | 0 | 100,000,000 | |
SPLIT_RAW_BYTES | 8,800 | 0 | 8,800 | |
Map output bytes | 0 | 0 | 10,000,000,000 | |
Reduce shuffle bytes | 0 | 0 | 10,200,038,400 | |
Physical memory (bytes) snapshot | 0 | 0 | 32,931,528,704 | |
Map input bytes | 0 | 0 | 10,000,000,000 | |
Reduce input groups | 0 | 0 | 100,000,000 | |
Combine output records | 0 | 0 | 0 | |
Reduce output records | 0 | 0 | 100,000,000 | |
Map output records | 0 | 0 | 100,000,000 | |
Combine input records | 0 | 0 | 0 | |
CPU time spent (ms) | 0 | 0 | 27,827,080 | |
Total committed heap usage (bytes) | 0 | 0 | 32,344,113,152 | |
File Input Format Counters | Bytes Read | 0 | 0 | 10,000,144,320 |
FileSystemCounters | HDFS_BYTES_READ | 10,000,153,120 | 0 | 10,000,153,120 |
FILE_BYTES_WRITTEN | 20,404,679,750 | 10,204,290,230 | 30,608,969,980 | |
FILE_BYTES_READ | 10,265,248,834 | 10,200,000,960 | 20,465,249,794 | |
HDFS_BYTES_WRITTEN | 0 | 10,000,000,000 | 10,000,000,000 | |
File Output Format Counters | Bytes Written | 0 | 0 | 10,000,000,000 |
Job Counters | Launched map tasks | 0 | 0 | 80 |
Launched reduce tasks | 0 | 0 | 80 | |
SLOTS_MILLIS_REDUCES | 0 | 0 | 28,079,434 | |
Total time spent by all reduces waiting after reserving slots (ms) | 0 | 0 | 0 | |
SLOTS_MILLIS_MAPS | 0 | 0 | 22,051,330 | |
Total time spent by all maps waiting after reserving slots (ms) | 0 | 0 | 0 | |
Rack-local map tasks | 0 | 0 | 30 | |
Data-local map tasks | 0 | 0 | 50 |
Map Completion Graph - close
Reduce Completion Graph - close
Go back to JobTracker
This is Apache Hadoop release 1.2.1
Thursday, December 4, 2014
The Daily WTF: The Robot Guys
Not really Big Data, but pretty funny :)
http://thedailywtf.com/articles/the-robot-guys
- Craig
http://thedailywtf.com/articles/the-robot-guys
- Craig
Subscribe to:
Posts (Atom)