No, the statement that radio waves get four times weaker every time the distance is doubled does not apply to laser beams. The behavior of laser beams differs from that of radio waves due to their different propagation characteristics.
The intensity of a laser beam follows an inverse square law, which states that the intensity decreases inversely with the square of the distance from the source. In other words, if the distance from the laser source is doubled, the intensity of the laser beam will decrease to one-fourth (1/2^2) of its original intensity. This is because the energy of the laser beam spreads out over a larger area as it propagates.
Radio waves, on the other hand, do not strictly follow the inverse square law for intensity reduction. The behavior of radio waves depends on various factors, including the frequency and the presence of obstacles or other environmental conditions. While the intensity of radio waves does decrease with distance, the rate of decrease can vary depending on the specific circumstances.
It's important to note that the behavior of any electromagnetic wave, including lasers and radio waves, can be influenced by factors such as absorption, scattering, and diffraction, which may introduce additional complexities to their propagation patterns.