Context Rot: How Increasing Input Tokens Impacts LLM Performance