Thinking Labs, led by Mira Murati, is tackling nondeterminism in large language models, aiming for consistent outputs regardless of server load. Researchers believe current LLMs like ChatGPT and Gemini suffer from batch invariance issues, causing inconsistent answers. Their solution involves redesigning GPU kernels to ensure deterministic behavior, crucial for enterprise AI, scientific research, and model fine-tuning.