)]}'
{
  "commit": "83b229ada56b2d60ca46d6ed46172bf52a50081c",
  "tree": "816883cc81e7b9478e2a4214c5785aed7d4737d6",
  "parents": [
    "e8e23f7a307becfc9d7cc863e357a3fdfaa66198"
  ],
  "author": {
    "name": "Evan Brown",
    "email": "ezb@google.com",
    "time": "Fri Jan 09 14:17:15 2026 -0800"
  },
  "committer": {
    "name": "Copybara-Service",
    "email": "copybara-worker@google.com",
    "time": "Fri Jan 09 14:19:43 2026 -0800"
  },
  "message": "Refactor GroupSse2Impl to improve codegen for hashtable lookups on x86_64.\n\nCurrently, clang isn\u0027t using blsr to clear the least significant bit because we are using uint16_t as the bitmask integer type and there are only blsr instructions for 32 and 64 bit registers. When we change to use uint32_t, clang generates blsr, but it also adds an unnecessary movzx (zero-extending move) even though the high bits of the output of vpmovmskb are zeroed. In order to avoid this problem, we use inline asm.\n\nPiperOrigin-RevId: 854341062\nChange-Id: I38b7e8913bf76ffd8e94b081afb52128fbcb795d\n",
  "tree_diff": [
    {
      "type": "modify",
      "old_id": "f0fd354f1b5c148cc3927108e6171ad7446d8941",
      "old_mode": 33188,
      "old_path": "absl/container/internal/hashtable_control_bytes.h",
      "new_id": "70016e104ae08b0df0ae4dec2c39a98482e46af6",
      "new_mode": 33188,
      "new_path": "absl/container/internal/hashtable_control_bytes.h"
    }
  ]
}
