Francois Chollet is leaving Google
Departure and career plans
- The Keras creator is leaving Google to start a new company with a friend; no move to another major lab.
- They remain US‑based for now, but are positive about the AI scene in Paris.
- Some see Google’s blog farewell as unusual and possibly a “soft launch” for the new venture.
Keras, TensorFlow, PyTorch, JAX
- Many recall early Keras (on Theano/TF1) as transformative: easy, Pythonic, and critical to deep learning’s “takeoff,” especially vs Theano, Caffe, Torch7.
- Common criticism: the abstraction was “too easy for basics, too hard for custom work” (custom losses, RNN variants, bespoke training loops), pushing researchers toward raw TensorFlow and then PyTorch.
- PyTorch is widely seen as the current default: better flexibility, LLM tooling, multi‑GPU support, performance, ecosystem, and community momentum.
- JAX is praised as powerful yet under‑appreciated; there are claims Google uses it heavily internally and that TensorFlow is losing ground.
Multi‑backend Keras and production use
- Several see the 2018–2019 folding of Keras into TensorFlow as the moment Keras “died,” and believe it accelerated PyTorch adoption.
- The Keras author clarifies they did not decide that merger; it was a higher‑level TensorFlow leadership decision and, in hindsight, likely a mistake.
- Keras is now standalone and multi‑backend again (TF, JAX, PyTorch), with explicit emphasis on backward compatibility and “progressive disclosure of complexity.”
- Users report Keras still running reliably in production (often since ~2018–2019) for vision and recommendation workloads; others report heavy technical debt and migration to PyTorch.
- The author lists many large companies using Keras; skeptics counter that this reflects legacy rather than current research leadership.
ARC‑AGI benchmark and AI progress
- Extensive debate around the ARC benchmark and a recent $1M prize:
- Some describe strong results (e.g., systems using GPT‑4‑generated synthetic tasks) as “gaming” a benchmark that was supposed to resist brute‑force and big‑data memorization.
- Others argue this is legitimate progress in problem‑solving and test‑time fine‑tuning, not a hack.
- Concerns are raised that human baselines measured via Mechanical Turk underestimate motivated human performance.
- The organizer plans ARC 2 with tasks that are harder to brute‑force yet similar human difficulty, and sees ARC as a high‑leverage path toward AGI‑relevant research.
- They expect ARC to be solved within a few years, see that solution as a stepping stone (not AGI itself), and maintain skepticism about an “intelligence explosion,” citing diminishing returns and the need to separate intelligence from autonomy.
Google culture, hierarchy, and AI startups
- Multiple commenters portray Google as comfortable but bureaucratic, where higher‑level decisions (e.g., around TensorFlow/Keras) can override project creators and dampen ambition.
- Others defend large hierarchies as necessary for 100k+‑employee firms and note that senior engineering levels typically have broad systems experience.
- There is broader discussion of AI startups: some claim top researchers can easily raise ~$100M; others argue that’s insufficient to sustain a competitive foundation‑model company, so future startups will likely focus on specialization and post‑training rather than training new general LLMs from scratch.