A 14-bit A to D conversion has to be accomplished for a specific design. In our
ID: 1799490 • Letter: A
Question
A 14-bit A to D conversion has to be accomplished for a specific design. In our design kit, we have two 8-bit microprocessors that implement Successive Approximation for A to D Conversions. In our case. we will have to connect the two processors in parallel to perform the 14-bit conversion. If the processors are driven by a 1 Mhz oscillator, how long will it take to do the conversion? Assume that the second processor uses it's 6 low order bits and Successive Approximation of the bits starts with the MSB. Also assume that the signals have been conditioned before they are inputted into the A to D microprocessors.Explanation / Answer
17.6 hours to perform the 14-bit conversion