Improve Huffman decoding speed using table-based implementation. #35
Add this suggestion to a batch that can be applied as a single commit.
This suggestion is invalid because no changes were made to the code.
Suggestions cannot be applied while the pull request is closed.
Suggestions cannot be applied while viewing a subset of changes.
Only one suggestion per line can be applied in a batch.
Add this suggestion to a batch that can be applied as a single commit.
Applying suggestions on deleted lines is not supported.
You must change the existing code in this line in order to create a valid suggestion.
Outdated suggestions cannot be applied.
This suggestion has been applied or marked resolved.
Suggestions cannot be applied from pending reviews.
Suggestions cannot be applied on multi-line comments.
Suggestions cannot be applied while the pull request is queued to merge.
Suggestion cannot be applied right now. Please check back later.
This is heavily inspired by the work done by @tatsuhiro-t for nghttp2, and ported to Python. In some micro-benchmarks focusing on decoding primarily ASCII-text, the Huffman decoder in the form of this unrolled state machine performs 4x faster than the naive Python version used elsewhere in the HPACK module, decompressing a 1kB Huffman string in ~600usec compared to the naive version's 2.4msec.
This is without question a micro-optimisation, but given that the table is not in flux anymore there's little advantage to using the naive Huffman decoder.
Enormous gratitude is owed to @tatsuhiro-t for his work on this decoder.