I’ve hired a lot of people in my life. I’ve been in management or partial-management positions since I was 23.
And I can tell you that if you insist that every person hired – whether it be receptionist, file clerk or janitor – have a college degree, all that means is that the very minute those degreed individuals find a better job more aligned to their skills and interests, they will be gone often without even giving notice.
And then you will be stuck.
Does it really make any sense that every job requires a degree? Especially since as from what I’ve seen most colleges teach very little useful (that is to say, you get out of education what you put into it, and most 18-year-old kids have no idea what the hell is going on).