Skip to content

About the width and something else #848

Answered by auphelia
wanglu4042 asked this question in Q&A
Discussion options

You must be logged in to vote

Hi @wanglu4042 ,

If you have a 2-bit quantized network, you will not need transformations that are specialized for 1-bit compute.

For the observation during throughput test:

  • FINN implements the network in a dataflow style. That means that each layer is implemented individually and the data is pushed through the layers like a pipeline. The throughput is thus determined by the slowest layer and not necessarily influenced by the number of layers. By increasing the parallelism per layer you might be able to achieve higher throughput. Please have a look at this tutorial about folding factors.

Replies: 1 comment 2 replies

Comment options

You must be logged in to vote
2 replies
@wanglu4042
Comment options

@auphelia
Comment options

Answer selected by auphelia
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Category
Q&A
Labels
None yet
2 participants