Broadcasters normally specify electromagnetic field strength in dBµ, which is dB above 1 µV/m. Receiver manufacturers typically specify sensitivity in dBf, which is dB above 1 femtowatt (10−15 W). I wondered how the two were related. Originally I used an antenna analysis program to find out. But it's more straightforward to directly use electromagnetics.
The following equation gives the power an antenna captures from an incident electromagnetic wave:
P = E²λ²g / 4πη
where P is power in watts, E is the electric field intensity in volts/meter, λ is wavelength in meters, g is isotropic antenna gain, and η is the impedance of free space.
Substituting numerical constants and converting to dB yields
dBf = dBµ + 10log(λ²g) - 6.76
At 98 MHz, λ = 3.06. For a halfwave dipole, g = 1.64. Therefore, at the middle of the FM broadcast band
dBf = dBµ + 5.1 + G
where G is antenna gain in dBd. G includes conductor, dielectric, mismatch, balun, and feedline losses. The expression is accurate to within 1 dB over 88–108 MHz.
60-dBµ FCC service contours are for a receive dipole 9 meters above ground. That length of RG-6 feedline yields 64.5 dBf. This is close to the 65 dBf IEEE 185-1975 specifies for several receiver measurements.