Did you read the contest rules and FAQ? (http://www.netflixprize.com/rules, http://www.netflixprize.com/faq -- and if you didn't, what are you doing throwing around words like "corrupt"?) It looks to me like they're being very transparent and explained things very clearly. You have a bunch of data in the test set; these are partitioned randomly into two equally-sized subsets, "quiz" and "test". You submit predictions for both subsets, but are only told how you did on "quiz". Netflix provides an MD5 checksum for the judging file that defines what the partition is; this file will be made available "at the end of the Contest". So this will be verifiable by anyone soon.
Also, don't brush overfitting aside. In the (paraphrased) words of one machine learning researcher, "life is a battle against entropy. In the same way, machine learning research is a battle against overfitting." Any data whose test results you use to select your algorithm or to adjust your algorithm's parameters is not properly considered part of the test set; after your optimization that data will provide an optimistically-biased estimate of your true error. Since competitors could get regular updates on their performance on the quiz dataset, one must assume that they were attempting to optimize this performance, and so quiz set performance was not be a good estimate of their true error. You can only get a good estimate of the true error of a method by testing against data that has played no part in its development.
Also, don't brush overfitting aside. In the (paraphrased) words of one machine learning researcher, "life is a battle against entropy. In the same way, machine learning research is a battle against overfitting." Any data whose test results you use to select your algorithm or to adjust your algorithm's parameters is not properly considered part of the test set; after your optimization that data will provide an optimistically-biased estimate of your true error. Since competitors could get regular updates on their performance on the quiz dataset, one must assume that they were attempting to optimize this performance, and so quiz set performance was not be a good estimate of their true error. You can only get a good estimate of the true error of a method by testing against data that has played no part in its development.