It's fine if they account for the number of tests they have made when they calculate their significance levels. If they just kept on trying different options until they ended up with p < 0.05 it's almost guaranteed to be noise.
Ah that's not too bad. Though to be fair you also need the data size. That's only what one in a 1k chance (or is it 10k? Too lazy to count it out). If their dataset is small or they automated testing cofactors there's still a decent chance of false probability.