aboutsummaryrefslogtreecommitdiff
path: root/src/main/java
diff options
context:
space:
mode:
authortivrfoa <lescoutinhovr@gmail.com>2024-01-27 15:41:00 -0300
committerGitHub <noreply@github.com>2024-01-27 19:41:00 +0100
commitd9604d9258ce29d955d89ce3f6d72dfd5119a42a (patch)
tree1d05054959a84e020f3c616f728a9dd465043f25 /src/main/java
parenta304f80710940fca3b283e9816b92c46986871d0 (diff)
Use LinkedBlockingQueue to process results - based on thomaswue (#603)
/** * Solution based on thomaswue solution, commit: * commit d0a28599c293d3afe3291fc3cf169a7b25ae9ae6 * Author: Thomas Wuerthinger <thomas.wuerthinger@oracle.com> * Date: Sun Jan 21 20:13:48 2024 +0100 * * Changes: * 1) Use LinkedBlockingQueue to store partial results, that * will then be merged into the final map later. * As different chunks finish at different times, this allows * to process them as they finish, instead of joining the * threads sequentially. * This change seems more useful for the 10k dataset, as the * runtime difference of each chunk is greater. * 2) Use only 4 threads if the file is >= 14GB. * This showed much better results on my local test, but I only * run with 200 million rows (because of limited RAM), and I have * no idea how it will perform on the 1brc HW. */
Diffstat (limited to 'src/main/java')
-rw-r--r--src/main/java/dev/morling/onebrc/CalculateAverage_tivrfoa.java431
1 files changed, 431 insertions, 0 deletions
diff --git a/src/main/java/dev/morling/onebrc/CalculateAverage_tivrfoa.java b/src/main/java/dev/morling/onebrc/CalculateAverage_tivrfoa.java
new file mode 100644
index 0000000..a1b4844
--- /dev/null
+++ b/src/main/java/dev/morling/onebrc/CalculateAverage_tivrfoa.java
@@ -0,0 +1,431 @@
+/*
+ * Copyright 2023 The original authors
+ *
+ * Licensed under the Apache License, Version 2.0 (the "License");
+ * you may not use this file except in compliance with the License.
+ * You may obtain a copy of the License at
+ *
+ * http://www.apache.org/licenses/LICENSE-2.0
+ *
+ * Unless required by applicable law or agreed to in writing, software
+ * distributed under the License is distributed on an "AS IS" BASIS,
+ * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+ * See the License for the specific language governing permissions and
+ * limitations under the License.
+ */
+package dev.morling.onebrc;
+
+import java.io.IOException;
+import java.nio.ByteBuffer;
+import java.nio.ByteOrder;
+import java.nio.channels.FileChannel;
+import java.nio.charset.StandardCharsets;
+import java.nio.file.Path;
+import java.nio.file.StandardOpenOption;
+import java.util.*;
+import java.util.concurrent.LinkedBlockingQueue;
+
+/**
+ * Solution based on thomaswue solution, commit:
+ * commit d0a28599c293d3afe3291fc3cf169a7b25ae9ae6
+ * Author: Thomas Wuerthinger <thomas.wuerthinger@oracle.com>
+ * Date: Sun Jan 21 20:13:48 2024 +0100
+ *
+ * Changes:
+ * 1) Use LinkedBlockingQueue to store partial results, that
+ * will then be merged into the final map later.
+ * As different chunks finish at different times, this allows
+ * to process them as they finish, instead of joining the
+ * threads sequentially.
+ * This change seems more useful for the 10k dataset, as the
+ * runtime difference of each chunk is greater.
+ * 2) Use only 4 threads if the file is >= 14GB.
+ * This showed much better results on my local test, but I only
+ * run with 200 million rows (because of limited RAM), and I have
+ * no idea how it will perform on the 1brc HW.
+ */
+public class CalculateAverage_tivrfoa {
+ private static final String FILE = "./measurements.txt";
+ private static LinkedBlockingQueue<List<Result>> partialResultQueue;
+ private static int C = 10_000;
+ private static final int MIN_TEMP = -999;
+ private static final int MAX_TEMP = 999;
+
+ // Holding the current result for a single city.
+ private static class Result {
+ long lastNameLong, secondLastNameLong;
+ long[] name;
+ int count;
+ short min, max;
+ long sum;
+
+ private Result() {
+ this.min = MAX_TEMP;
+ this.max = MIN_TEMP;
+ }
+
+ public String toString() {
+ return round(((double) min) / 10.0) + "/" + round((((double) sum) / 10.0) / count) + "/" + round(((double) max) / 10.0);
+ }
+
+ private static double round(double value) {
+ return Math.round(value * 10.0) / 10.0;
+ }
+
+ // Accumulate another result into this one.
+ private void add(Result other) {
+ if (other.min < min) {
+ min = other.min;
+ }
+ if (other.max > max) {
+ max = other.max;
+ }
+ sum += other.sum;
+ count += other.count;
+ }
+
+ public String calcName() {
+ ByteBuffer bb = ByteBuffer.allocate(name.length * Long.BYTES).order(ByteOrder.nativeOrder());
+ bb.asLongBuffer().put(name);
+ byte[] array = bb.array();
+ int i = 0;
+ while (array[i++] != ';')
+ ;
+ return new String(array, 0, i - 1, StandardCharsets.UTF_8);
+ }
+ }
+
+ private static final class SolveChunk extends Thread {
+ private long chunkStart, chunkEnd;
+
+ public SolveChunk(long chunkStart, long chunkEnd) {
+ this.chunkStart = chunkStart;
+ this.chunkEnd = chunkEnd;
+ }
+
+ @Override
+ public void run() {
+ try {
+ partialResultQueue.put(parseLoop(chunkStart, chunkEnd));
+ }
+ catch (Exception e) {
+ e.printStackTrace();
+ System.exit(1);
+ }
+ }
+ }
+
+ public static void main(String[] args) throws Exception {
+ boolean runTrick = true;
+ for (var arg : args) {
+ if (arg.equals("--worker")) {
+ runTrick = false;
+ break;
+ }
+ }
+ if (runTrick) {
+ spawnWorker();
+ return;
+ }
+ final int cpus = Runtime.getRuntime().availableProcessors();
+ final long[] chunks = getSegments(cpus);
+ final int workers = chunks.length - 1;
+ partialResultQueue = new LinkedBlockingQueue<>(workers);
+ final SolveChunk[] threads = new SolveChunk[workers];
+ for (int i = 0; i < workers; i++) {
+ threads[i] = new SolveChunk(chunks[i], chunks[i + 1]);
+ threads[i].start();
+ }
+ final TreeMap<String, Result> ret = new TreeMap<>();
+ for (int i = 0; i < workers; ++i) {
+ accumulateResults(ret, partialResultQueue.take());
+ }
+ System.out.println(ret);
+ System.out.close();
+ }
+
+ private static void spawnWorker() throws IOException {
+ ProcessHandle.Info info = ProcessHandle.current().info();
+ ArrayList<String> workerCommand = new ArrayList<>();
+ info.command().ifPresent(workerCommand::add);
+ info.arguments().ifPresent(args -> workerCommand.addAll(Arrays.asList(args)));
+ workerCommand.add("--worker");
+ new ProcessBuilder()
+ .command(workerCommand)
+ .inheritIO()
+ .redirectOutput(ProcessBuilder.Redirect.PIPE)
+ .start()
+ .getInputStream()
+ .transferTo(System.out);
+ }
+
+ private static void accumulateResults(TreeMap<String, Result> result, List<Result> newResult) {
+ for (Result r : newResult) {
+ String name = r.calcName();
+ Result current = result.putIfAbsent(name, r);
+ if (current != null) {
+ current.add(r);
+ }
+ }
+ }
+
+ // Main parse loop.
+ private static ArrayList<Result> parseLoop(long chunkStart, long chunkEnd) {
+ ArrayList<Result> ret = new ArrayList<>(C);
+ Result[] results = new Result[1 << 17];
+ Scanner scanner = new Scanner(chunkStart, chunkEnd);
+ long word = scanner.getLong();
+ long pos = findDelimiter(word);
+ while (scanner.hasNext()) {
+ long nameAddress = scanner.pos();
+ long hash = 0;
+
+ // Search for ';', one long at a time.
+ if (pos != 0) {
+ pos = Long.numberOfTrailingZeros(pos) >>> 3;
+ scanner.add(pos);
+ word = mask(word, pos);
+ hash = word;
+
+ int number = scanNumber(scanner);
+ long nextWord = scanner.getLong();
+ long nextPos = findDelimiter(nextWord);
+
+ Result existingResult = results[hashToIndex(hash, results)];
+ if (existingResult != null && existingResult.lastNameLong == word) {
+ word = nextWord;
+ pos = nextPos;
+ record(existingResult, number);
+ continue;
+ }
+
+ scanner.setPos(nameAddress + pos);
+ }
+ else {
+ scanner.add(8);
+ hash = word;
+ long prevWord = word;
+ word = scanner.getLong();
+ pos = findDelimiter(word);
+ if (pos != 0) {
+ pos = Long.numberOfTrailingZeros(pos) >>> 3;
+ scanner.add(pos);
+ word = mask(word, pos);
+ hash ^= word;
+
+ Result existingResult = results[hashToIndex(hash, results)];
+ if (existingResult != null && existingResult.lastNameLong == word && existingResult.secondLastNameLong == prevWord) {
+ int number = scanNumber(scanner);
+ word = scanner.getLong();
+ pos = findDelimiter(word);
+ record(existingResult, number);
+ continue;
+ }
+ }
+ else {
+ scanner.add(8);
+ hash ^= word;
+ while (true) {
+ word = scanner.getLong();
+ pos = findDelimiter(word);
+ if (pos != 0) {
+ pos = Long.numberOfTrailingZeros(pos) >>> 3;
+ scanner.add(pos);
+ word = mask(word, pos);
+ hash ^= word;
+ break;
+ }
+ else {
+ scanner.add(8);
+ hash ^= word;
+ }
+ }
+ }
+ }
+
+ // Save length of name for later.
+ int nameLength = (int) (scanner.pos() - nameAddress);
+ int number = scanNumber(scanner);
+
+ // Final calculation for index into hash table.
+ int tableIndex = hashToIndex(hash, results);
+ outer: while (true) {
+ Result existingResult = results[tableIndex];
+ if (existingResult == null) {
+ existingResult = newEntry(results, nameAddress, tableIndex, nameLength, scanner);
+ ret.add(existingResult);
+ }
+ // Check for collision.
+ int i = 0;
+ int namePos = 0;
+ for (; i < nameLength + 1 - 8; i += 8) {
+ if (namePos >= existingResult.name.length || existingResult.name[namePos++] != scanner.getLongAt(nameAddress + i)) {
+ tableIndex = (tableIndex + 31) & (results.length - 1);
+ continue outer;
+ }
+ }
+
+ int remainingShift = (64 - (nameLength + 1 - i) << 3);
+ if (((existingResult.lastNameLong ^ (scanner.getLongAt(nameAddress + i) << remainingShift)) == 0)) {
+ record(existingResult, number);
+ break;
+ }
+ else {
+ // Collision error, try next.
+ tableIndex = (tableIndex + 31) & (results.length - 1);
+ }
+ }
+
+ word = scanner.getLong();
+ pos = findDelimiter(word);
+ }
+ return ret;
+ }
+
+ private static int scanNumber(Scanner scanPtr) {
+ scanPtr.add(1);
+ long numberWord = scanPtr.getLong();
+ int decimalSepPos = Long.numberOfTrailingZeros(~numberWord & 0x10101000);
+ int number = convertIntoNumber(decimalSepPos, numberWord);
+ scanPtr.add((decimalSepPos >>> 3) + 3);
+ return number;
+ }
+
+ private static void record(Result existingResult, int number) {
+ if (number < existingResult.min) {
+ existingResult.min = (short) number;
+ }
+ if (number > existingResult.max) {
+ existingResult.max = (short) number;
+ }
+ existingResult.sum += number;
+ existingResult.count++;
+ }
+
+ private static int hashToIndex(long hash, Result[] results) {
+ int hashAsInt = (int) (hash ^ (hash >>> 28));
+ int finalHash = (hashAsInt ^ (hashAsInt >>> 17));
+ return (finalHash & (results.length - 1));
+ }
+
+ private static long mask(long word, long pos) {
+ return (word << ((7 - pos) << 3));
+ }
+
+ // Special method to convert a number in the ascii number into an int without branches created by Quan Anh Mai.
+ private static int convertIntoNumber(int decimalSepPos, long numberWord) {
+ int shift = 28 - decimalSepPos;
+ // signed is -1 if negative, 0 otherwise
+ long signed = (~numberWord << 59) >> 63;
+ long designMask = ~(signed & 0xFF);
+ // Align the number to a specific position and transform the ascii to digit value
+ long digits = ((numberWord & designMask) << shift) & 0x0F000F0F00L;
+ // Now digits is in the form 0xUU00TTHH00 (UU: units digit, TT: tens digit, HH: hundreds digit)
+ // 0xUU00TTHH00 * (100 * 0x1000000 + 10 * 0x10000 + 1) =
+ // 0x000000UU00TTHH00 + 0x00UU00TTHH000000 * 10 + 0xUU00TTHH00000000 * 100
+ long absValue = ((digits * 0x640a0001) >>> 32) & 0x3FF;
+ long value = (absValue ^ signed) - signed;
+ return (int) value;
+ }
+
+ private static long findDelimiter(long word) {
+ long input = word ^ 0x3B3B3B3B3B3B3B3BL;
+ long tmp = (input - 0x0101010101010101L) & ~input & 0x8080808080808080L;
+ return tmp;
+ }
+
+ private static Result newEntry(Result[] results, long nameAddress, int hash, int nameLength, Scanner scanner) {
+ Result r = new Result();
+ results[hash] = r;
+ long[] name = new long[(nameLength / Long.BYTES) + 1];
+ int pos = 0;
+ int i = 0;
+ for (; i < nameLength + 1 - Long.BYTES; i += Long.BYTES) {
+ name[pos++] = scanner.getLongAt(nameAddress + i);
+ }
+
+ if (pos > 0) {
+ r.secondLastNameLong = name[pos - 1];
+ }
+
+ int remainingShift = (64 - (nameLength + 1 - i) << 3);
+ long lastWord = (scanner.getLongAt(nameAddress + i) << remainingShift);
+ r.lastNameLong = lastWord;
+ name[pos] = lastWord >> remainingShift;
+ r.name = name;
+ return r;
+ }
+
+ private static long[] getSegments(int cpus) throws IOException {
+ try (var fileChannel = FileChannel.open(Path.of(FILE), StandardOpenOption.READ)) {
+ long fileSize = fileChannel.size();
+ int numberOfChunks = cpus / 2;
+ if (fileSize < (int) 14e9) {
+ C = 500;
+ numberOfChunks = cpus;
+ }
+ long segmentSize = (fileSize + numberOfChunks - 1) / numberOfChunks;
+ long[] chunks = new long[numberOfChunks + 1];
+ long mappedAddress = fileChannel.map(FileChannel.MapMode.READ_ONLY, 0, fileSize, java.lang.foreign.Arena.global()).address();
+ chunks[0] = mappedAddress;
+ long endAddress = mappedAddress + fileSize;
+ Scanner s = new Scanner(mappedAddress, mappedAddress + fileSize);
+ for (int i = 1; i < numberOfChunks; ++i) {
+ long chunkAddress = mappedAddress + i * segmentSize;
+ // Align to first row start.
+ while (chunkAddress < endAddress && (s.getLongAt(chunkAddress++) & 0xFF) != '\n')
+ ;
+ chunks[i] = Math.min(chunkAddress, endAddress);
+ }
+ chunks[numberOfChunks] = endAddress;
+ return chunks;
+ }
+ }
+
+ private static class Scanner {
+
+ private static final sun.misc.Unsafe UNSAFE = initUnsafe();
+
+ private static sun.misc.Unsafe initUnsafe() {
+ try {
+ java.lang.reflect.Field theUnsafe = sun.misc.Unsafe.class.getDeclaredField("theUnsafe");
+ theUnsafe.setAccessible(true);
+ return (sun.misc.Unsafe) theUnsafe.get(sun.misc.Unsafe.class);
+ }
+ catch (NoSuchFieldException | IllegalAccessException e) {
+ throw new RuntimeException(e);
+ }
+ }
+
+ long pos, end;
+
+ public Scanner(long start, long end) {
+ this.pos = start;
+ this.end = end;
+ }
+
+ boolean hasNext() {
+ return pos < end;
+ }
+
+ long pos() {
+ return pos;
+ }
+
+ void add(long delta) {
+ pos += delta;
+ }
+
+ long getLong() {
+ return UNSAFE.getLong(pos);
+ }
+
+ long getLongAt(long pos) {
+ return UNSAFE.getLong(pos);
+ }
+
+ void setPos(long l) {
+ this.pos = l;
+ }
+ }
+} \ No newline at end of file